ansible-playbook [core 2.12.6] config file = /etc/ansible/ansible.cfg configured module search path = ['/root/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules'] ansible python module location = /usr/lib/python3.9/site-packages/ansible ansible collection location = /tmp/tmpp_vyhrdp executable location = /usr/bin/ansible-playbook python version = 3.9.13 (main, May 18 2022, 00:00:00) [GCC 11.3.1 20220421 (Red Hat 11.3.1-2)] jinja version = 2.11.3 libyaml = True Using /etc/ansible/ansible.cfg as config file Skipping callback 'debug', as we already have a stdout callback. Skipping callback 'default', as we already have a stdout callback. Skipping callback 'minimal', as we already have a stdout callback. Skipping callback 'oneline', as we already have a stdout callback. PLAYBOOK: tests_lvm_percent_size.yml ******************************************* 1 plays in /tmp/tmpafnwmbol/tests/tests_lvm_percent_size.yml PLAY [all] ********************************************************************* TASK [Gathering Facts] ********************************************************* task path: /tmp/tmpafnwmbol/tests/tests_lvm_percent_size.yml:2 Thursday 21 July 2022 15:22:31 +0000 (0:00:00.013) 0:00:00.013 ********* ok: [/cache/rhel-8-y.qcow2.snap] META: ran handlers TASK [include_role : linux-system-roles.storage] ******************************* task path: /tmp/tmpafnwmbol/tests/tests_lvm_percent_size.yml:17 Thursday 21 July 2022 15:22:32 +0000 (0:00:01.343) 0:00:01.356 ********* TASK [linux-system-roles.storage : set platform/version specific variables] **** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main.yml:2 Thursday 21 July 2022 15:22:32 +0000 (0:00:00.035) 0:00:01.392 ********* included: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for /cache/rhel-8-y.qcow2.snap TASK [linux-system-roles.storage : Ensure ansible_facts used by role] ********** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2 Thursday 21 July 2022 15:22:32 +0000 (0:00:00.031) 0:00:01.424 ********* ok: [/cache/rhel-8-y.qcow2.snap] TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8 Thursday 21 July 2022 15:22:33 +0000 (0:00:00.512) 0:00:01.936 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [/cache/rhel-8-y.qcow2.snap] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [/cache/rhel-8-y.qcow2.snap] => (item=RedHat_8.yml) => { "ansible_facts": { "blivet_package_list": [ "python3-blivet", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap", "vdo", "kmod-kvdo", "xfsprogs" ] }, "ansible_included_var_files": [ "/tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/vars/RedHat_8.yml" ], "ansible_loop_var": "item", "changed": false, "item": "RedHat_8.yml" } skipping: [/cache/rhel-8-y.qcow2.snap] => (item=RedHat_8.7.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat_8.7.yml", "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : define an empty list of pools to be used in testing] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main.yml:5 Thursday 21 July 2022 15:22:33 +0000 (0:00:00.056) 0:00:01.992 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : define an empty list of volumes to be used in testing] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main.yml:9 Thursday 21 July 2022 15:22:33 +0000 (0:00:00.028) 0:00:02.021 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : include the appropriate provider tasks] ***** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main.yml:13 Thursday 21 July 2022 15:22:33 +0000 (0:00:00.028) 0:00:02.049 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount included: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for /cache/rhel-8-y.qcow2.snap TASK [linux-system-roles.storage : get a list of rpm packages installed on host machine] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2 Thursday 21 July 2022 15:22:33 +0000 (0:00:00.045) 0:00:02.095 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : make sure blivet is available] ************** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:7 Thursday 21 July 2022 15:22:33 +0000 (0:00:00.018) 0:00:02.114 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [linux-system-roles.storage : show storage_pools] ************************* task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:14 Thursday 21 July 2022 15:22:35 +0000 (0:00:01.808) 0:00:03.923 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_pools": "VARIABLE IS NOT DEFINED!: 'storage_pools' is undefined" } TASK [linux-system-roles.storage : show storage_volumes] *********************** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:19 Thursday 21 July 2022 15:22:35 +0000 (0:00:00.033) 0:00:03.956 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [linux-system-roles.storage : get required packages] ********************** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:24 Thursday 21 July 2022 15:22:35 +0000 (0:00:00.033) 0:00:03.989 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [], "pools": [], "volumes": [] } TASK [linux-system-roles.storage : enable copr repositories if needed] ********* task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:37 Thursday 21 July 2022 15:22:36 +0000 (0:00:00.728) 0:00:04.717 ********* included: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml for /cache/rhel-8-y.qcow2.snap TASK [linux-system-roles.storage : check if the COPR support packages should be installed] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:2 Thursday 21 July 2022 15:22:36 +0000 (0:00:00.040) 0:00:04.758 ********* TASK [linux-system-roles.storage : make sure COPR support packages are present] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:13 Thursday 21 July 2022 15:22:36 +0000 (0:00:00.030) 0:00:04.789 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : enable COPRs] ******************************* task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:18 Thursday 21 July 2022 15:22:36 +0000 (0:00:00.033) 0:00:04.822 ********* TASK [linux-system-roles.storage : make sure required packages are installed] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:44 Thursday 21 July 2022 15:22:36 +0000 (0:00:00.032) 0:00:04.854 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [linux-system-roles.storage : get service facts] ************************** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:51 Thursday 21 July 2022 15:22:37 +0000 (0:00:01.397) 0:00:06.252 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "services": { "NetworkManager-dispatcher.service": { "name": "NetworkManager-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "NetworkManager-wait-online.service": { "name": "NetworkManager-wait-online.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "NetworkManager.service": { "name": "NetworkManager.service", "source": "systemd", "state": "running", "status": "enabled" }, "arp-ethers.service": { "name": "arp-ethers.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "auditd.service": { "name": "auditd.service", "source": "systemd", "state": "running", "status": "enabled" }, "auth-rpcgss-module.service": { "name": "auth-rpcgss-module.service", "source": "systemd", "state": "stopped", "status": "static" }, "autovt@.service": { "name": "autovt@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "blivet.service": { "name": "blivet.service", "source": "systemd", "state": "inactive", "status": "static" }, "blk-availability.service": { "name": "blk-availability.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "chrony-dnssrv@.service": { "name": "chrony-dnssrv@.service", "source": "systemd", "state": "unknown", "status": "static" }, "chrony-wait.service": { "name": "chrony-wait.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd.service": { "name": "chronyd.service", "source": "systemd", "state": "running", "status": "enabled" }, "cloud-config.service": { "name": "cloud-config.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-final.service": { "name": "cloud-final.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init-hotplugd.service": { "name": "cloud-init-hotplugd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cloud-init-local.service": { "name": "cloud-init-local.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init.service": { "name": "cloud-init.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cockpit-motd.service": { "name": "cockpit-motd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cockpit-wsinstance-http.service": { "name": "cockpit-wsinstance-http.service", "source": "systemd", "state": "inactive", "status": "static" }, "cockpit-wsinstance-https-factory@.service": { "name": "cockpit-wsinstance-https-factory@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cockpit-wsinstance-https@.service": { "name": "cockpit-wsinstance-https@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cockpit.service": { "name": "cockpit.service", "source": "systemd", "state": "inactive", "status": "static" }, "console-getty.service": { "name": "console-getty.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "container-getty@.service": { "name": "container-getty@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cpupower.service": { "name": "cpupower.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "crond.service": { "name": "crond.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-org.freedesktop.hostname1.service": { "name": "dbus-org.freedesktop.hostname1.service", "source": "systemd", "state": "active", "status": "static" }, "dbus-org.freedesktop.locale1.service": { "name": "dbus-org.freedesktop.locale1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.login1.service": { "name": "dbus-org.freedesktop.login1.service", "source": "systemd", "state": "active", "status": "static" }, "dbus-org.freedesktop.nm-dispatcher.service": { "name": "dbus-org.freedesktop.nm-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "dbus-org.freedesktop.portable1.service": { "name": "dbus-org.freedesktop.portable1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.timedate1.service": { "name": "dbus-org.freedesktop.timedate1.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "dbus.service": { "name": "dbus.service", "source": "systemd", "state": "running", "status": "static" }, "dbxtool.service": { "name": "dbxtool.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "debug-shell.service": { "name": "debug-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "display-manager.service": { "name": "display-manager.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "dm-event.service": { "name": "dm-event.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-makecache.service": { "name": "dnf-makecache.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-cmdline.service": { "name": "dracut-cmdline.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-initqueue.service": { "name": "dracut-initqueue.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-mount.service": { "name": "dracut-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-mount.service": { "name": "dracut-pre-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-pivot.service": { "name": "dracut-pre-pivot.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-trigger.service": { "name": "dracut-pre-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-udev.service": { "name": "dracut-pre-udev.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown.service": { "name": "dracut-shutdown.service", "source": "systemd", "state": "stopped", "status": "static" }, "emergency.service": { "name": "emergency.service", "source": "systemd", "state": "stopped", "status": "static" }, "fcoe.service": { "name": "fcoe.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "fstrim.service": { "name": "fstrim.service", "source": "systemd", "state": "inactive", "status": "static" }, "getty@.service": { "name": "getty@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "getty@tty1.service": { "name": "getty@tty1.service", "source": "systemd", "state": "running", "status": "active" }, "grub-boot-indeterminate.service": { "name": "grub-boot-indeterminate.service", "source": "systemd", "state": "inactive", "status": "static" }, "gssproxy.service": { "name": "gssproxy.service", "source": "systemd", "state": "running", "status": "disabled" }, "halt-local.service": { "name": "halt-local.service", "source": "systemd", "state": "inactive", "status": "static" }, "hv_kvp_daemon.service": { "name": "hv_kvp_daemon.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "import-state.service": { "name": "import-state.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "initrd-cleanup.service": { "name": "initrd-cleanup.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-parse-etc.service": { "name": "initrd-parse-etc.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-switch-root.service": { "name": "initrd-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-udevadm-cleanup-db.service": { "name": "initrd-udevadm-cleanup-db.service", "source": "systemd", "state": "stopped", "status": "static" }, "insights-client-boot.service": { "name": "insights-client-boot.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "insights-client-results.service": { "name": "insights-client-results.service", "source": "systemd", "state": "inactive", "status": "static" }, "insights-client.service": { "name": "insights-client.service", "source": "systemd", "state": "inactive", "status": "static" }, "irqbalance.service": { "name": "irqbalance.service", "source": "systemd", "state": "running", "status": "enabled" }, "iscsi-shutdown.service": { "name": "iscsi-shutdown.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "iscsi.service": { "name": "iscsi.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "iscsid.service": { "name": "iscsid.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "kdump.service": { "name": "kdump.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "kmod-static-nodes.service": { "name": "kmod-static-nodes.service", "source": "systemd", "state": "stopped", "status": "static" }, "kvm_stat.service": { "name": "kvm_stat.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "ldconfig.service": { "name": "ldconfig.service", "source": "systemd", "state": "stopped", "status": "static" }, "loadmodules.service": { "name": "loadmodules.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "lvm2-activation-early.service": { "name": "lvm2-activation-early.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "lvm2-activation.service": { "name": "lvm2-activation.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "lvm2-lvmpolld.service": { "name": "lvm2-lvmpolld.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-monitor.service": { "name": "lvm2-monitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "lvm2-pvscan@.service": { "name": "lvm2-pvscan@.service", "source": "systemd", "state": "unknown", "status": "static" }, "man-db-cache-update.service": { "name": "man-db-cache-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "man-db-restart-cache-update.service": { "name": "man-db-restart-cache-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "mdadm-grow-continue@.service": { "name": "mdadm-grow-continue@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdadm-last-resort@.service": { "name": "mdadm-last-resort@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdcheck_continue.service": { "name": "mdcheck_continue.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdcheck_start.service": { "name": "mdcheck_start.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmon@.service": { "name": "mdmon@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdmonitor-oneshot.service": { "name": "mdmonitor-oneshot.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmonitor.service": { "name": "mdmonitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "messagebus.service": { "name": "messagebus.service", "source": "systemd", "state": "active", "status": "static" }, "microcode.service": { "name": "microcode.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "multipathd.service": { "name": "multipathd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "ndctl-monitor.service": { "name": "ndctl-monitor.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "network.service": { "name": "network.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "nfs-blkmap.service": { "name": "nfs-blkmap.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-convert.service": { "name": "nfs-convert.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-idmapd.service": { "name": "nfs-idmapd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-mountd.service": { "name": "nfs-mountd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-server.service": { "name": "nfs-server.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-utils.service": { "name": "nfs-utils.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfsdcld.service": { "name": "nfsdcld.service", "source": "systemd", "state": "stopped", "status": "static" }, "nis-domainname.service": { "name": "nis-domainname.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "ntpd.service": { "name": "ntpd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ntpdate.service": { "name": "ntpdate.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "oddjobd.service": { "name": "oddjobd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "packagekit-offline-update.service": { "name": "packagekit-offline-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "packagekit.service": { "name": "packagekit.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-quit-wait.service": { "name": "plymouth-quit-wait.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "plymouth-start.service": { "name": "plymouth-start.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "polkit.service": { "name": "polkit.service", "source": "systemd", "state": "running", "status": "static" }, "qemu-guest-agent.service": { "name": "qemu-guest-agent.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "quotaon.service": { "name": "quotaon.service", "source": "systemd", "state": "inactive", "status": "static" }, "rbdmap.service": { "name": "rbdmap.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "rc-local.service": { "name": "rc-local.service", "source": "systemd", "state": "stopped", "status": "static" }, "rdisc.service": { "name": "rdisc.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rescue.service": { "name": "rescue.service", "source": "systemd", "state": "stopped", "status": "static" }, "rhcd.service": { "name": "rhcd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhsm-facts.service": { "name": "rhsm-facts.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhsm.service": { "name": "rhsm.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhsmcertd.service": { "name": "rhsmcertd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpc-gssd.service": { "name": "rpc-gssd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd-notify.service": { "name": "rpc-statd-notify.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd.service": { "name": "rpc-statd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpcbind.service": { "name": "rpcbind.service", "source": "systemd", "state": "running", "status": "enabled" }, "rsyslog.service": { "name": "rsyslog.service", "source": "systemd", "state": "running", "status": "enabled" }, "selinux-autorelabel-mark.service": { "name": "selinux-autorelabel-mark.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "selinux-autorelabel.service": { "name": "selinux-autorelabel.service", "source": "systemd", "state": "inactive", "status": "static" }, "serial-getty@.service": { "name": "serial-getty@.service", "source": "systemd", "state": "unknown", "status": "indirect" }, "serial-getty@ttyS0.service": { "name": "serial-getty@ttyS0.service", "source": "systemd", "state": "running", "status": "active" }, "snapd.seeded.service": { "name": "snapd.seeded.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sntp.service": { "name": "sntp.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sshd-keygen.service": { "name": "sshd-keygen.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sshd-keygen@.service": { "name": "sshd-keygen@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@ecdsa.service": { "name": "sshd-keygen@ecdsa.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd-keygen@ed25519.service": { "name": "sshd-keygen@ed25519.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd-keygen@rsa.service": { "name": "sshd-keygen@rsa.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd.service": { "name": "sshd.service", "source": "systemd", "state": "running", "status": "enabled" }, "sshd@.service": { "name": "sshd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "sssd-autofs.service": { "name": "sssd-autofs.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-kcm.service": { "name": "sssd-kcm.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "sssd-nss.service": { "name": "sssd-nss.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pac.service": { "name": "sssd-pac.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pam.service": { "name": "sssd-pam.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-ssh.service": { "name": "sssd-ssh.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-sudo.service": { "name": "sssd-sudo.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd.service": { "name": "sssd.service", "source": "systemd", "state": "running", "status": "enabled" }, "syslog.service": { "name": "syslog.service", "source": "systemd", "state": "active", "status": "enabled" }, "system-update-cleanup.service": { "name": "system-update-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-ask-password-console.service": { "name": "systemd-ask-password-console.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-wall.service": { "name": "systemd-ask-password-wall.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-backlight@.service": { "name": "systemd-backlight@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-binfmt.service": { "name": "systemd-binfmt.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-coredump@.service": { "name": "systemd-coredump@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-exit.service": { "name": "systemd-exit.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-firstboot.service": { "name": "systemd-firstboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck-root.service": { "name": "systemd-fsck-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck@.service": { "name": "systemd-fsck@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-fsck@dev-disk-by\\x2duuid-7B77\\x2d95E7.service": { "name": "systemd-fsck@dev-disk-by\\x2duuid-7B77\\x2d95E7.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-fsck@dev-vdb1.service": { "name": "systemd-fsck@dev-vdb1.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-fsck@dev-vdc1.service": { "name": "systemd-fsck@dev-vdc1.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-halt.service": { "name": "systemd-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hibernate-resume@.service": { "name": "systemd-hibernate-resume@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-hibernate.service": { "name": "systemd-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hostnamed.service": { "name": "systemd-hostnamed.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-hwdb-update.service": { "name": "systemd-hwdb-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-hybrid-sleep.service": { "name": "systemd-hybrid-sleep.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-initctl.service": { "name": "systemd-initctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-catalog-update.service": { "name": "systemd-journal-catalog-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-flush.service": { "name": "systemd-journal-flush.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journald.service": { "name": "systemd-journald.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-kexec.service": { "name": "systemd-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-localed.service": { "name": "systemd-localed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-logind.service": { "name": "systemd-logind.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-machine-id-commit.service": { "name": "systemd-machine-id-commit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-modules-load.service": { "name": "systemd-modules-load.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-networkd-wait-online.service": { "name": "systemd-networkd-wait-online.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "systemd-portabled.service": { "name": "systemd-portabled.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-poweroff.service": { "name": "systemd-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-quotacheck.service": { "name": "systemd-quotacheck.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-random-seed.service": { "name": "systemd-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-reboot.service": { "name": "systemd-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-remount-fs.service": { "name": "systemd-remount-fs.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-resolved.service": { "name": "systemd-resolved.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-rfkill.service": { "name": "systemd-rfkill.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend-then-hibernate.service": { "name": "systemd-suspend-then-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend.service": { "name": "systemd-suspend.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-sysctl.service": { "name": "systemd-sysctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-sysusers.service": { "name": "systemd-sysusers.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-timedated.service": { "name": "systemd-timedated.service", "source": "systemd", "state": "inactive", "status": "masked" }, "systemd-timesyncd.service": { "name": "systemd-timesyncd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "systemd-tmpfiles-clean.service": { "name": "systemd-tmpfiles-clean.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev.service": { "name": "systemd-tmpfiles-setup-dev.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup.service": { "name": "systemd-tmpfiles-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-settle.service": { "name": "systemd-udev-settle.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-trigger.service": { "name": "systemd-udev-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udevd.service": { "name": "systemd-udevd.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-update-done.service": { "name": "systemd-update-done.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp-runlevel.service": { "name": "systemd-update-utmp-runlevel.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp.service": { "name": "systemd-update-utmp.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-user-sessions.service": { "name": "systemd-user-sessions.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-vconsole-setup.service": { "name": "systemd-vconsole-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-volatile-root.service": { "name": "systemd-volatile-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "tcsd.service": { "name": "tcsd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "teamd@.service": { "name": "teamd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "timedatex.service": { "name": "timedatex.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "tuned.service": { "name": "tuned.service", "source": "systemd", "state": "running", "status": "enabled" }, "unbound-anchor.service": { "name": "unbound-anchor.service", "source": "systemd", "state": "stopped", "status": "static" }, "user-runtime-dir@.service": { "name": "user-runtime-dir@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user-runtime-dir@0.service": { "name": "user-runtime-dir@0.service", "source": "systemd", "state": "stopped", "status": "active" }, "user@.service": { "name": "user@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user@0.service": { "name": "user@0.service", "source": "systemd", "state": "running", "status": "active" }, "vdo-start-by-dev@.service": { "name": "vdo-start-by-dev@.service", "source": "systemd", "state": "unknown", "status": "static" }, "vdo.service": { "name": "vdo.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "ypbind.service": { "name": "ypbind.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "yppasswdd.service": { "name": "yppasswdd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ypserv.service": { "name": "ypserv.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ypxfrd.service": { "name": "ypxfrd.service", "source": "systemd", "state": "stopped", "status": "not-found" } } }, "changed": false } TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************ task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:58 Thursday 21 July 2022 15:22:39 +0000 (0:00:01.966) 0:00:08.218 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] ******* task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:71 Thursday 21 July 2022 15:22:39 +0000 (0:00:00.126) 0:00:08.345 ********* TASK [linux-system-roles.storage : manage the pools and volumes to match the specified state] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77 Thursday 21 July 2022 15:22:39 +0000 (0:00:00.022) 0:00:08.368 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [], "pools": [], "volumes": [] } TASK [linux-system-roles.storage : Workaround for udev issue on some platforms] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:91 Thursday 21 July 2022 15:22:40 +0000 (0:00:00.567) 0:00:08.936 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] ***** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:103 Thursday 21 July 2022 15:22:40 +0000 (0:00:00.035) 0:00:08.971 ********* TASK [linux-system-roles.storage : show blivet_output] ************************* task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:109 Thursday 21 July 2022 15:22:40 +0000 (0:00:00.021) 0:00:08.993 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "blivet_output": { "actions": [], "changed": false, "crypts": [], "failed": false, "leaves": [], "mounts": [], "packages": [], "pools": [], "volumes": [] } } TASK [linux-system-roles.storage : set the list of pools for test verification] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:114 Thursday 21 July 2022 15:22:40 +0000 (0:00:00.035) 0:00:09.028 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : set the list of volumes for test verification] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:118 Thursday 21 July 2022 15:22:40 +0000 (0:00:00.034) 0:00:09.063 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : remove obsolete mounts] ********************* task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:134 Thursday 21 July 2022 15:22:40 +0000 (0:00:00.038) 0:00:09.101 ********* TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:146 Thursday 21 July 2022 15:22:40 +0000 (0:00:00.036) 0:00:09.138 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : set up new/current mounts] ****************** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:151 Thursday 21 July 2022 15:22:40 +0000 (0:00:00.023) 0:00:09.162 ********* TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:163 Thursday 21 July 2022 15:22:40 +0000 (0:00:00.035) 0:00:09.198 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : retrieve facts for the /etc/crypttab file] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:171 Thursday 21 July 2022 15:22:40 +0000 (0:00:00.023) 0:00:09.221 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "stat": { "exists": false } } TASK [linux-system-roles.storage : manage /etc/crypttab to account for changes we just made] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:176 Thursday 21 July 2022 15:22:41 +0000 (0:00:00.509) 0:00:09.731 ********* TASK [linux-system-roles.storage : Update facts] ******************************* task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:198 Thursday 21 July 2022 15:22:41 +0000 (0:00:00.022) 0:00:09.754 ********* ok: [/cache/rhel-8-y.qcow2.snap] META: role_complete for /cache/rhel-8-y.qcow2.snap TASK [include_tasks] *********************************************************** task path: /tmp/tmpafnwmbol/tests/tests_lvm_percent_size.yml:20 Thursday 21 July 2022 15:22:42 +0000 (0:00:01.000) 0:00:10.755 ********* included: /tmp/tmpafnwmbol/tests/get_unused_disk.yml for /cache/rhel-8-y.qcow2.snap TASK [Find unused disks in the system] ***************************************** task path: /tmp/tmpafnwmbol/tests/get_unused_disk.yml:2 Thursday 21 July 2022 15:22:42 +0000 (0:00:00.035) 0:00:10.790 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "disks": [ "sda" ] } TASK [Set unused_disks if necessary] ******************************************* task path: /tmp/tmpafnwmbol/tests/get_unused_disk.yml:9 Thursday 21 July 2022 15:22:42 +0000 (0:00:00.513) 0:00:11.304 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "unused_disks": [ "sda" ] }, "changed": false } TASK [Exit playbook when there's not enough unused disks in the system] ******** task path: /tmp/tmpafnwmbol/tests/get_unused_disk.yml:14 Thursday 21 July 2022 15:22:42 +0000 (0:00:00.039) 0:00:11.343 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Print unused disks] ****************************************************** task path: /tmp/tmpafnwmbol/tests/get_unused_disk.yml:19 Thursday 21 July 2022 15:22:42 +0000 (0:00:00.039) 0:00:11.383 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "unused_disks": [ "sda" ] } TASK [Try to create LVM with an invalid size specification.] ******************* task path: /tmp/tmpafnwmbol/tests/tests_lvm_percent_size.yml:27 Thursday 21 July 2022 15:22:42 +0000 (0:00:00.034) 0:00:11.417 ********* TASK [linux-system-roles.storage : set platform/version specific variables] **** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main.yml:2 Thursday 21 July 2022 15:22:42 +0000 (0:00:00.037) 0:00:11.454 ********* included: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for /cache/rhel-8-y.qcow2.snap TASK [linux-system-roles.storage : Ensure ansible_facts used by role] ********** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2 Thursday 21 July 2022 15:22:42 +0000 (0:00:00.033) 0:00:11.488 ********* ok: [/cache/rhel-8-y.qcow2.snap] TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8 Thursday 21 July 2022 15:22:43 +0000 (0:00:00.550) 0:00:12.039 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [/cache/rhel-8-y.qcow2.snap] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [/cache/rhel-8-y.qcow2.snap] => (item=RedHat_8.yml) => { "ansible_facts": { "blivet_package_list": [ "python3-blivet", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap", "vdo", "kmod-kvdo", "xfsprogs" ] }, "ansible_included_var_files": [ "/tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/vars/RedHat_8.yml" ], "ansible_loop_var": "item", "changed": false, "item": "RedHat_8.yml" } skipping: [/cache/rhel-8-y.qcow2.snap] => (item=RedHat_8.7.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat_8.7.yml", "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : define an empty list of pools to be used in testing] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main.yml:5 Thursday 21 July 2022 15:22:43 +0000 (0:00:00.061) 0:00:12.100 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : define an empty list of volumes to be used in testing] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main.yml:9 Thursday 21 July 2022 15:22:43 +0000 (0:00:00.035) 0:00:12.136 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : include the appropriate provider tasks] ***** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main.yml:13 Thursday 21 July 2022 15:22:43 +0000 (0:00:00.033) 0:00:12.170 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount included: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for /cache/rhel-8-y.qcow2.snap TASK [linux-system-roles.storage : get a list of rpm packages installed on host machine] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2 Thursday 21 July 2022 15:22:43 +0000 (0:00:00.048) 0:00:12.218 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : make sure blivet is available] ************** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:7 Thursday 21 July 2022 15:22:43 +0000 (0:00:00.021) 0:00:12.239 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [linux-system-roles.storage : show storage_pools] ************************* task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:14 Thursday 21 July 2022 15:22:45 +0000 (0:00:01.408) 0:00:13.648 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_pools": [ { "disks": [ "sda" ], "name": "foo", "volumes": [ { "mount_point": "/opt/test1", "name": "test1", "size": "2x%" } ] } ] } TASK [linux-system-roles.storage : show storage_volumes] *********************** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:19 Thursday 21 July 2022 15:22:45 +0000 (0:00:00.038) 0:00:13.686 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [linux-system-roles.storage : get required packages] ********************** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:24 Thursday 21 July 2022 15:22:45 +0000 (0:00:00.037) 0:00:13.724 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [ "lvm2" ], "pools": [], "volumes": [] } TASK [linux-system-roles.storage : enable copr repositories if needed] ********* task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:37 Thursday 21 July 2022 15:22:46 +0000 (0:00:01.335) 0:00:15.059 ********* included: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml for /cache/rhel-8-y.qcow2.snap TASK [linux-system-roles.storage : check if the COPR support packages should be installed] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:2 Thursday 21 July 2022 15:22:46 +0000 (0:00:00.046) 0:00:15.105 ********* TASK [linux-system-roles.storage : make sure COPR support packages are present] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:13 Thursday 21 July 2022 15:22:46 +0000 (0:00:00.035) 0:00:15.141 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : enable COPRs] ******************************* task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:18 Thursday 21 July 2022 15:22:46 +0000 (0:00:00.040) 0:00:15.181 ********* TASK [linux-system-roles.storage : make sure required packages are installed] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:44 Thursday 21 July 2022 15:22:46 +0000 (0:00:00.034) 0:00:15.216 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [linux-system-roles.storage : get service facts] ************************** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:51 Thursday 21 July 2022 15:22:48 +0000 (0:00:01.359) 0:00:16.575 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "services": { "NetworkManager-dispatcher.service": { "name": "NetworkManager-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "NetworkManager-wait-online.service": { "name": "NetworkManager-wait-online.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "NetworkManager.service": { "name": "NetworkManager.service", "source": "systemd", "state": "running", "status": "enabled" }, "arp-ethers.service": { "name": "arp-ethers.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "auditd.service": { "name": "auditd.service", "source": "systemd", "state": "running", "status": "enabled" }, "auth-rpcgss-module.service": { "name": "auth-rpcgss-module.service", "source": "systemd", "state": "stopped", "status": "static" }, "autovt@.service": { "name": "autovt@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "blivet.service": { "name": "blivet.service", "source": "systemd", "state": "inactive", "status": "static" }, "blk-availability.service": { "name": "blk-availability.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "chrony-dnssrv@.service": { "name": "chrony-dnssrv@.service", "source": "systemd", "state": "unknown", "status": "static" }, "chrony-wait.service": { "name": "chrony-wait.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd.service": { "name": "chronyd.service", "source": "systemd", "state": "running", "status": "enabled" }, "cloud-config.service": { "name": "cloud-config.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-final.service": { "name": "cloud-final.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init-hotplugd.service": { "name": "cloud-init-hotplugd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cloud-init-local.service": { "name": "cloud-init-local.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init.service": { "name": "cloud-init.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cockpit-motd.service": { "name": "cockpit-motd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cockpit-wsinstance-http.service": { "name": "cockpit-wsinstance-http.service", "source": "systemd", "state": "inactive", "status": "static" }, "cockpit-wsinstance-https-factory@.service": { "name": "cockpit-wsinstance-https-factory@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cockpit-wsinstance-https@.service": { "name": "cockpit-wsinstance-https@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cockpit.service": { "name": "cockpit.service", "source": "systemd", "state": "inactive", "status": "static" }, "console-getty.service": { "name": "console-getty.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "container-getty@.service": { "name": "container-getty@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cpupower.service": { "name": "cpupower.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "crond.service": { "name": "crond.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-org.freedesktop.hostname1.service": { "name": "dbus-org.freedesktop.hostname1.service", "source": "systemd", "state": "active", "status": "static" }, "dbus-org.freedesktop.locale1.service": { "name": "dbus-org.freedesktop.locale1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.login1.service": { "name": "dbus-org.freedesktop.login1.service", "source": "systemd", "state": "active", "status": "static" }, "dbus-org.freedesktop.nm-dispatcher.service": { "name": "dbus-org.freedesktop.nm-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "dbus-org.freedesktop.portable1.service": { "name": "dbus-org.freedesktop.portable1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.timedate1.service": { "name": "dbus-org.freedesktop.timedate1.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "dbus.service": { "name": "dbus.service", "source": "systemd", "state": "running", "status": "static" }, "dbxtool.service": { "name": "dbxtool.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "debug-shell.service": { "name": "debug-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "display-manager.service": { "name": "display-manager.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "dm-event.service": { "name": "dm-event.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-makecache.service": { "name": "dnf-makecache.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-cmdline.service": { "name": "dracut-cmdline.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-initqueue.service": { "name": "dracut-initqueue.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-mount.service": { "name": "dracut-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-mount.service": { "name": "dracut-pre-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-pivot.service": { "name": "dracut-pre-pivot.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-trigger.service": { "name": "dracut-pre-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-udev.service": { "name": "dracut-pre-udev.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown.service": { "name": "dracut-shutdown.service", "source": "systemd", "state": "stopped", "status": "static" }, "emergency.service": { "name": "emergency.service", "source": "systemd", "state": "stopped", "status": "static" }, "fcoe.service": { "name": "fcoe.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "fstrim.service": { "name": "fstrim.service", "source": "systemd", "state": "inactive", "status": "static" }, "getty@.service": { "name": "getty@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "getty@tty1.service": { "name": "getty@tty1.service", "source": "systemd", "state": "running", "status": "active" }, "grub-boot-indeterminate.service": { "name": "grub-boot-indeterminate.service", "source": "systemd", "state": "inactive", "status": "static" }, "gssproxy.service": { "name": "gssproxy.service", "source": "systemd", "state": "running", "status": "disabled" }, "halt-local.service": { "name": "halt-local.service", "source": "systemd", "state": "inactive", "status": "static" }, "hv_kvp_daemon.service": { "name": "hv_kvp_daemon.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "import-state.service": { "name": "import-state.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "initrd-cleanup.service": { "name": "initrd-cleanup.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-parse-etc.service": { "name": "initrd-parse-etc.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-switch-root.service": { "name": "initrd-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-udevadm-cleanup-db.service": { "name": "initrd-udevadm-cleanup-db.service", "source": "systemd", "state": "stopped", "status": "static" }, "insights-client-boot.service": { "name": "insights-client-boot.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "insights-client-results.service": { "name": "insights-client-results.service", "source": "systemd", "state": "inactive", "status": "static" }, "insights-client.service": { "name": "insights-client.service", "source": "systemd", "state": "inactive", "status": "static" }, "irqbalance.service": { "name": "irqbalance.service", "source": "systemd", "state": "running", "status": "enabled" }, "iscsi-shutdown.service": { "name": "iscsi-shutdown.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "iscsi.service": { "name": "iscsi.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "iscsid.service": { "name": "iscsid.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "kdump.service": { "name": "kdump.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "kmod-static-nodes.service": { "name": "kmod-static-nodes.service", "source": "systemd", "state": "stopped", "status": "static" }, "kvm_stat.service": { "name": "kvm_stat.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "ldconfig.service": { "name": "ldconfig.service", "source": "systemd", "state": "stopped", "status": "static" }, "loadmodules.service": { "name": "loadmodules.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "lvm2-activation-early.service": { "name": "lvm2-activation-early.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "lvm2-activation.service": { "name": "lvm2-activation.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "lvm2-lvmpolld.service": { "name": "lvm2-lvmpolld.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-monitor.service": { "name": "lvm2-monitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "lvm2-pvscan@.service": { "name": "lvm2-pvscan@.service", "source": "systemd", "state": "unknown", "status": "static" }, "man-db-cache-update.service": { "name": "man-db-cache-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "man-db-restart-cache-update.service": { "name": "man-db-restart-cache-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "mdadm-grow-continue@.service": { "name": "mdadm-grow-continue@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdadm-last-resort@.service": { "name": "mdadm-last-resort@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdcheck_continue.service": { "name": "mdcheck_continue.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdcheck_start.service": { "name": "mdcheck_start.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmon@.service": { "name": "mdmon@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdmonitor-oneshot.service": { "name": "mdmonitor-oneshot.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmonitor.service": { "name": "mdmonitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "messagebus.service": { "name": "messagebus.service", "source": "systemd", "state": "active", "status": "static" }, "microcode.service": { "name": "microcode.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "multipathd.service": { "name": "multipathd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "ndctl-monitor.service": { "name": "ndctl-monitor.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "network.service": { "name": "network.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "nfs-blkmap.service": { "name": "nfs-blkmap.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-convert.service": { "name": "nfs-convert.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-idmapd.service": { "name": "nfs-idmapd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-mountd.service": { "name": "nfs-mountd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-server.service": { "name": "nfs-server.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-utils.service": { "name": "nfs-utils.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfsdcld.service": { "name": "nfsdcld.service", "source": "systemd", "state": "stopped", "status": "static" }, "nis-domainname.service": { "name": "nis-domainname.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "ntpd.service": { "name": "ntpd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ntpdate.service": { "name": "ntpdate.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "oddjobd.service": { "name": "oddjobd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "packagekit-offline-update.service": { "name": "packagekit-offline-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "packagekit.service": { "name": "packagekit.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-quit-wait.service": { "name": "plymouth-quit-wait.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "plymouth-start.service": { "name": "plymouth-start.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "polkit.service": { "name": "polkit.service", "source": "systemd", "state": "running", "status": "static" }, "qemu-guest-agent.service": { "name": "qemu-guest-agent.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "quotaon.service": { "name": "quotaon.service", "source": "systemd", "state": "inactive", "status": "static" }, "rbdmap.service": { "name": "rbdmap.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "rc-local.service": { "name": "rc-local.service", "source": "systemd", "state": "stopped", "status": "static" }, "rdisc.service": { "name": "rdisc.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rescue.service": { "name": "rescue.service", "source": "systemd", "state": "stopped", "status": "static" }, "rhcd.service": { "name": "rhcd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhsm-facts.service": { "name": "rhsm-facts.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhsm.service": { "name": "rhsm.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhsmcertd.service": { "name": "rhsmcertd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpc-gssd.service": { "name": "rpc-gssd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd-notify.service": { "name": "rpc-statd-notify.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd.service": { "name": "rpc-statd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpcbind.service": { "name": "rpcbind.service", "source": "systemd", "state": "running", "status": "enabled" }, "rsyslog.service": { "name": "rsyslog.service", "source": "systemd", "state": "running", "status": "enabled" }, "selinux-autorelabel-mark.service": { "name": "selinux-autorelabel-mark.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "selinux-autorelabel.service": { "name": "selinux-autorelabel.service", "source": "systemd", "state": "inactive", "status": "static" }, "serial-getty@.service": { "name": "serial-getty@.service", "source": "systemd", "state": "unknown", "status": "indirect" }, "serial-getty@ttyS0.service": { "name": "serial-getty@ttyS0.service", "source": "systemd", "state": "running", "status": "active" }, "snapd.seeded.service": { "name": "snapd.seeded.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sntp.service": { "name": "sntp.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sshd-keygen.service": { "name": "sshd-keygen.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sshd-keygen@.service": { "name": "sshd-keygen@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@ecdsa.service": { "name": "sshd-keygen@ecdsa.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd-keygen@ed25519.service": { "name": "sshd-keygen@ed25519.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd-keygen@rsa.service": { "name": "sshd-keygen@rsa.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd.service": { "name": "sshd.service", "source": "systemd", "state": "running", "status": "enabled" }, "sshd@.service": { "name": "sshd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "sssd-autofs.service": { "name": "sssd-autofs.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-kcm.service": { "name": "sssd-kcm.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "sssd-nss.service": { "name": "sssd-nss.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pac.service": { "name": "sssd-pac.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pam.service": { "name": "sssd-pam.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-ssh.service": { "name": "sssd-ssh.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-sudo.service": { "name": "sssd-sudo.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd.service": { "name": "sssd.service", "source": "systemd", "state": "running", "status": "enabled" }, "syslog.service": { "name": "syslog.service", "source": "systemd", "state": "active", "status": "enabled" }, "system-update-cleanup.service": { "name": "system-update-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-ask-password-console.service": { "name": "systemd-ask-password-console.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-wall.service": { "name": "systemd-ask-password-wall.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-backlight@.service": { "name": "systemd-backlight@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-binfmt.service": { "name": "systemd-binfmt.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-coredump@.service": { "name": "systemd-coredump@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-exit.service": { "name": "systemd-exit.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-firstboot.service": { "name": "systemd-firstboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck-root.service": { "name": "systemd-fsck-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck@.service": { "name": "systemd-fsck@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-fsck@dev-disk-by\\x2duuid-7B77\\x2d95E7.service": { "name": "systemd-fsck@dev-disk-by\\x2duuid-7B77\\x2d95E7.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-fsck@dev-vdb1.service": { "name": "systemd-fsck@dev-vdb1.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-fsck@dev-vdc1.service": { "name": "systemd-fsck@dev-vdc1.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-halt.service": { "name": "systemd-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hibernate-resume@.service": { "name": "systemd-hibernate-resume@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-hibernate.service": { "name": "systemd-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hostnamed.service": { "name": "systemd-hostnamed.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-hwdb-update.service": { "name": "systemd-hwdb-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-hybrid-sleep.service": { "name": "systemd-hybrid-sleep.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-initctl.service": { "name": "systemd-initctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-catalog-update.service": { "name": "systemd-journal-catalog-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-flush.service": { "name": "systemd-journal-flush.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journald.service": { "name": "systemd-journald.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-kexec.service": { "name": "systemd-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-localed.service": { "name": "systemd-localed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-logind.service": { "name": "systemd-logind.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-machine-id-commit.service": { "name": "systemd-machine-id-commit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-modules-load.service": { "name": "systemd-modules-load.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-networkd-wait-online.service": { "name": "systemd-networkd-wait-online.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "systemd-portabled.service": { "name": "systemd-portabled.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-poweroff.service": { "name": "systemd-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-quotacheck.service": { "name": "systemd-quotacheck.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-random-seed.service": { "name": "systemd-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-reboot.service": { "name": "systemd-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-remount-fs.service": { "name": "systemd-remount-fs.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-resolved.service": { "name": "systemd-resolved.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-rfkill.service": { "name": "systemd-rfkill.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend-then-hibernate.service": { "name": "systemd-suspend-then-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend.service": { "name": "systemd-suspend.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-sysctl.service": { "name": "systemd-sysctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-sysusers.service": { "name": "systemd-sysusers.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-timedated.service": { "name": "systemd-timedated.service", "source": "systemd", "state": "inactive", "status": "masked" }, "systemd-timesyncd.service": { "name": "systemd-timesyncd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "systemd-tmpfiles-clean.service": { "name": "systemd-tmpfiles-clean.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev.service": { "name": "systemd-tmpfiles-setup-dev.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup.service": { "name": "systemd-tmpfiles-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-settle.service": { "name": "systemd-udev-settle.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-trigger.service": { "name": "systemd-udev-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udevd.service": { "name": "systemd-udevd.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-update-done.service": { "name": "systemd-update-done.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp-runlevel.service": { "name": "systemd-update-utmp-runlevel.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp.service": { "name": "systemd-update-utmp.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-user-sessions.service": { "name": "systemd-user-sessions.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-vconsole-setup.service": { "name": "systemd-vconsole-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-volatile-root.service": { "name": "systemd-volatile-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "tcsd.service": { "name": "tcsd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "teamd@.service": { "name": "teamd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "timedatex.service": { "name": "timedatex.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "tuned.service": { "name": "tuned.service", "source": "systemd", "state": "running", "status": "enabled" }, "unbound-anchor.service": { "name": "unbound-anchor.service", "source": "systemd", "state": "stopped", "status": "static" }, "user-runtime-dir@.service": { "name": "user-runtime-dir@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user-runtime-dir@0.service": { "name": "user-runtime-dir@0.service", "source": "systemd", "state": "stopped", "status": "active" }, "user@.service": { "name": "user@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user@0.service": { "name": "user@0.service", "source": "systemd", "state": "running", "status": "active" }, "vdo-start-by-dev@.service": { "name": "vdo-start-by-dev@.service", "source": "systemd", "state": "unknown", "status": "static" }, "vdo.service": { "name": "vdo.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "ypbind.service": { "name": "ypbind.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "yppasswdd.service": { "name": "yppasswdd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ypserv.service": { "name": "ypserv.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ypxfrd.service": { "name": "ypxfrd.service", "source": "systemd", "state": "stopped", "status": "not-found" } } }, "changed": false } TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************ task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:58 Thursday 21 July 2022 15:22:49 +0000 (0:00:01.715) 0:00:18.291 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] ******* task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:71 Thursday 21 July 2022 15:22:49 +0000 (0:00:00.057) 0:00:18.348 ********* TASK [linux-system-roles.storage : manage the pools and volumes to match the specified state] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77 Thursday 21 July 2022 15:22:49 +0000 (0:00:00.020) 0:00:18.369 ********* fatal: [/cache/rhel-8-y.qcow2.snap]: FAILED! => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [], "pools": [], "volumes": [] } MSG: invalid percentage '2x%' size specified for volume in pool 'foo' TASK [linux-system-roles.storage : failed message] ***************************** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:99 Thursday 21 July 2022 15:22:51 +0000 (0:00:01.304) 0:00:19.674 ********* fatal: [/cache/rhel-8-y.qcow2.snap]: FAILED! => { "changed": false } MSG: {'changed': False, 'actions': [], 'leaves': [], 'mounts': [], 'crypts': [], 'pools': [], 'volumes': [], 'packages': [], 'failed': True, 'msg': "invalid percentage '2x%' size specified for volume in pool 'foo'", 'invocation': {'module_args': {'pools': [{'disks': ['sda'], 'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'name': 'foo', 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_metadata_version': None, 'state': 'present', 'type': 'lvm', 'volumes': [{'encryption': None, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': None, 'fs_label': '', 'fs_type': None, 'mount_options': None, 'mount_point': '/opt/test1', 'name': 'test1', 'raid_level': None, 'size': '2x%', 'state': 'present', 'type': None, 'cached': None, 'cache_devices': [], 'cache_mode': None, 'cache_size': None, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None}], 'raid_chunk_size': None}], 'volumes': [], 'use_partitions': None, 'disklabel_type': None, 'pool_defaults': {'state': 'present', 'type': 'lvm', 'disks': [], 'volumes': [], 'encryption': False, 'encryption_password': None, 'encryption_key': None, 'encryption_cipher': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None}, 'volume_defaults': {'state': 'present', 'type': 'lvm', 'size': 0, 'disks': [], 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_overwrite_existing': True, 'mount_point': '', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, 'encryption': False, 'encryption_password': None, 'encryption_key': None, 'encryption_cipher': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'compression': None, 'deduplication': None, 'vdo_pool_size': None, 'thin': None, 'thin_pool_name': None, 'thin_pool_size': None, 'cached': False, 'cache_size': 0, 'cache_mode': None, 'cache_devices': []}, 'safe_mode': False, 'packages_only': False, 'diskvolume_mkfs_option_map': {}}}, '_ansible_no_log': False} TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] ***** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:103 Thursday 21 July 2022 15:22:51 +0000 (0:00:00.039) 0:00:19.714 ********* TASK [Check that we failed in the role] **************************************** task path: /tmp/tmpafnwmbol/tests/tests_lvm_percent_size.yml:44 Thursday 21 July 2022 15:22:51 +0000 (0:00:00.022) 0:00:19.736 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Check for the expected error message] ************************************ task path: /tmp/tmpafnwmbol/tests/tests_lvm_percent_size.yml:50 Thursday 21 July 2022 15:22:51 +0000 (0:00:00.034) 0:00:19.771 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Create two LVM logical volumes under volume group 'foo' using percentage sizes] *** task path: /tmp/tmpafnwmbol/tests/tests_lvm_percent_size.yml:63 Thursday 21 July 2022 15:22:51 +0000 (0:00:00.033) 0:00:19.804 ********* TASK [linux-system-roles.storage : set platform/version specific variables] **** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main.yml:2 Thursday 21 July 2022 15:22:51 +0000 (0:00:00.038) 0:00:19.843 ********* included: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for /cache/rhel-8-y.qcow2.snap TASK [linux-system-roles.storage : Ensure ansible_facts used by role] ********** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2 Thursday 21 July 2022 15:22:51 +0000 (0:00:00.035) 0:00:19.878 ********* ok: [/cache/rhel-8-y.qcow2.snap] TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8 Thursday 21 July 2022 15:22:51 +0000 (0:00:00.528) 0:00:20.407 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [/cache/rhel-8-y.qcow2.snap] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [/cache/rhel-8-y.qcow2.snap] => (item=RedHat_8.yml) => { "ansible_facts": { "blivet_package_list": [ "python3-blivet", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap", "vdo", "kmod-kvdo", "xfsprogs" ] }, "ansible_included_var_files": [ "/tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/vars/RedHat_8.yml" ], "ansible_loop_var": "item", "changed": false, "item": "RedHat_8.yml" } skipping: [/cache/rhel-8-y.qcow2.snap] => (item=RedHat_8.7.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat_8.7.yml", "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : define an empty list of pools to be used in testing] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main.yml:5 Thursday 21 July 2022 15:22:51 +0000 (0:00:00.058) 0:00:20.465 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : define an empty list of volumes to be used in testing] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main.yml:9 Thursday 21 July 2022 15:22:51 +0000 (0:00:00.031) 0:00:20.497 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : include the appropriate provider tasks] ***** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main.yml:13 Thursday 21 July 2022 15:22:52 +0000 (0:00:00.065) 0:00:20.563 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount included: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for /cache/rhel-8-y.qcow2.snap TASK [linux-system-roles.storage : get a list of rpm packages installed on host machine] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2 Thursday 21 July 2022 15:22:52 +0000 (0:00:00.047) 0:00:20.610 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : make sure blivet is available] ************** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:7 Thursday 21 July 2022 15:22:52 +0000 (0:00:00.021) 0:00:20.632 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [linux-system-roles.storage : show storage_pools] ************************* task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:14 Thursday 21 July 2022 15:22:53 +0000 (0:00:01.365) 0:00:21.997 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_pools": [ { "disks": [ "sda" ], "name": "foo", "volumes": [ { "mount_point": "/opt/test1", "name": "test1", "size": "60%" }, { "fs_type": "ext4", "mount_point": "/opt/test2", "name": "test2", "size": "40%" } ] } ] } TASK [linux-system-roles.storage : show storage_volumes] *********************** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:19 Thursday 21 July 2022 15:22:53 +0000 (0:00:00.039) 0:00:22.037 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [linux-system-roles.storage : get required packages] ********************** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:24 Thursday 21 July 2022 15:22:53 +0000 (0:00:00.036) 0:00:22.073 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [ "e2fsprogs", "lvm2" ], "pools": [], "volumes": [] } TASK [linux-system-roles.storage : enable copr repositories if needed] ********* task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:37 Thursday 21 July 2022 15:22:54 +0000 (0:00:01.327) 0:00:23.401 ********* included: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml for /cache/rhel-8-y.qcow2.snap TASK [linux-system-roles.storage : check if the COPR support packages should be installed] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:2 Thursday 21 July 2022 15:22:54 +0000 (0:00:00.045) 0:00:23.447 ********* TASK [linux-system-roles.storage : make sure COPR support packages are present] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:13 Thursday 21 July 2022 15:22:54 +0000 (0:00:00.032) 0:00:23.479 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : enable COPRs] ******************************* task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:18 Thursday 21 July 2022 15:22:54 +0000 (0:00:00.038) 0:00:23.518 ********* TASK [linux-system-roles.storage : make sure required packages are installed] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:44 Thursday 21 July 2022 15:22:55 +0000 (0:00:00.033) 0:00:23.551 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [linux-system-roles.storage : get service facts] ************************** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:51 Thursday 21 July 2022 15:22:56 +0000 (0:00:01.332) 0:00:24.884 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "services": { "NetworkManager-dispatcher.service": { "name": "NetworkManager-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "NetworkManager-wait-online.service": { "name": "NetworkManager-wait-online.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "NetworkManager.service": { "name": "NetworkManager.service", "source": "systemd", "state": "running", "status": "enabled" }, "arp-ethers.service": { "name": "arp-ethers.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "auditd.service": { "name": "auditd.service", "source": "systemd", "state": "running", "status": "enabled" }, "auth-rpcgss-module.service": { "name": "auth-rpcgss-module.service", "source": "systemd", "state": "stopped", "status": "static" }, "autovt@.service": { "name": "autovt@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "blivet.service": { "name": "blivet.service", "source": "systemd", "state": "inactive", "status": "static" }, "blk-availability.service": { "name": "blk-availability.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "chrony-dnssrv@.service": { "name": "chrony-dnssrv@.service", "source": "systemd", "state": "unknown", "status": "static" }, "chrony-wait.service": { "name": "chrony-wait.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd.service": { "name": "chronyd.service", "source": "systemd", "state": "running", "status": "enabled" }, "cloud-config.service": { "name": "cloud-config.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-final.service": { "name": "cloud-final.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init-hotplugd.service": { "name": "cloud-init-hotplugd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cloud-init-local.service": { "name": "cloud-init-local.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init.service": { "name": "cloud-init.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cockpit-motd.service": { "name": "cockpit-motd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cockpit-wsinstance-http.service": { "name": "cockpit-wsinstance-http.service", "source": "systemd", "state": "inactive", "status": "static" }, "cockpit-wsinstance-https-factory@.service": { "name": "cockpit-wsinstance-https-factory@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cockpit-wsinstance-https@.service": { "name": "cockpit-wsinstance-https@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cockpit.service": { "name": "cockpit.service", "source": "systemd", "state": "inactive", "status": "static" }, "console-getty.service": { "name": "console-getty.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "container-getty@.service": { "name": "container-getty@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cpupower.service": { "name": "cpupower.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "crond.service": { "name": "crond.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-org.freedesktop.hostname1.service": { "name": "dbus-org.freedesktop.hostname1.service", "source": "systemd", "state": "active", "status": "static" }, "dbus-org.freedesktop.locale1.service": { "name": "dbus-org.freedesktop.locale1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.login1.service": { "name": "dbus-org.freedesktop.login1.service", "source": "systemd", "state": "active", "status": "static" }, "dbus-org.freedesktop.nm-dispatcher.service": { "name": "dbus-org.freedesktop.nm-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "dbus-org.freedesktop.portable1.service": { "name": "dbus-org.freedesktop.portable1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.timedate1.service": { "name": "dbus-org.freedesktop.timedate1.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "dbus.service": { "name": "dbus.service", "source": "systemd", "state": "running", "status": "static" }, "dbxtool.service": { "name": "dbxtool.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "debug-shell.service": { "name": "debug-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "display-manager.service": { "name": "display-manager.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "dm-event.service": { "name": "dm-event.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-makecache.service": { "name": "dnf-makecache.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-cmdline.service": { "name": "dracut-cmdline.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-initqueue.service": { "name": "dracut-initqueue.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-mount.service": { "name": "dracut-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-mount.service": { "name": "dracut-pre-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-pivot.service": { "name": "dracut-pre-pivot.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-trigger.service": { "name": "dracut-pre-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-udev.service": { "name": "dracut-pre-udev.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown.service": { "name": "dracut-shutdown.service", "source": "systemd", "state": "stopped", "status": "static" }, "emergency.service": { "name": "emergency.service", "source": "systemd", "state": "stopped", "status": "static" }, "fcoe.service": { "name": "fcoe.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "fstrim.service": { "name": "fstrim.service", "source": "systemd", "state": "inactive", "status": "static" }, "getty@.service": { "name": "getty@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "getty@tty1.service": { "name": "getty@tty1.service", "source": "systemd", "state": "running", "status": "active" }, "grub-boot-indeterminate.service": { "name": "grub-boot-indeterminate.service", "source": "systemd", "state": "inactive", "status": "static" }, "gssproxy.service": { "name": "gssproxy.service", "source": "systemd", "state": "running", "status": "disabled" }, "halt-local.service": { "name": "halt-local.service", "source": "systemd", "state": "inactive", "status": "static" }, "hv_kvp_daemon.service": { "name": "hv_kvp_daemon.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "import-state.service": { "name": "import-state.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "initrd-cleanup.service": { "name": "initrd-cleanup.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-parse-etc.service": { "name": "initrd-parse-etc.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-switch-root.service": { "name": "initrd-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-udevadm-cleanup-db.service": { "name": "initrd-udevadm-cleanup-db.service", "source": "systemd", "state": "stopped", "status": "static" }, "insights-client-boot.service": { "name": "insights-client-boot.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "insights-client-results.service": { "name": "insights-client-results.service", "source": "systemd", "state": "inactive", "status": "static" }, "insights-client.service": { "name": "insights-client.service", "source": "systemd", "state": "inactive", "status": "static" }, "irqbalance.service": { "name": "irqbalance.service", "source": "systemd", "state": "running", "status": "enabled" }, "iscsi-shutdown.service": { "name": "iscsi-shutdown.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "iscsi.service": { "name": "iscsi.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "iscsid.service": { "name": "iscsid.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "kdump.service": { "name": "kdump.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "kmod-static-nodes.service": { "name": "kmod-static-nodes.service", "source": "systemd", "state": "stopped", "status": "static" }, "kvm_stat.service": { "name": "kvm_stat.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "ldconfig.service": { "name": "ldconfig.service", "source": "systemd", "state": "stopped", "status": "static" }, "loadmodules.service": { "name": "loadmodules.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "lvm2-activation-early.service": { "name": "lvm2-activation-early.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "lvm2-activation.service": { "name": "lvm2-activation.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "lvm2-lvmpolld.service": { "name": "lvm2-lvmpolld.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-monitor.service": { "name": "lvm2-monitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "lvm2-pvscan@.service": { "name": "lvm2-pvscan@.service", "source": "systemd", "state": "unknown", "status": "static" }, "man-db-cache-update.service": { "name": "man-db-cache-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "man-db-restart-cache-update.service": { "name": "man-db-restart-cache-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "mdadm-grow-continue@.service": { "name": "mdadm-grow-continue@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdadm-last-resort@.service": { "name": "mdadm-last-resort@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdcheck_continue.service": { "name": "mdcheck_continue.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdcheck_start.service": { "name": "mdcheck_start.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmon@.service": { "name": "mdmon@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdmonitor-oneshot.service": { "name": "mdmonitor-oneshot.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmonitor.service": { "name": "mdmonitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "messagebus.service": { "name": "messagebus.service", "source": "systemd", "state": "active", "status": "static" }, "microcode.service": { "name": "microcode.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "multipathd.service": { "name": "multipathd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "ndctl-monitor.service": { "name": "ndctl-monitor.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "network.service": { "name": "network.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "nfs-blkmap.service": { "name": "nfs-blkmap.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-convert.service": { "name": "nfs-convert.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-idmapd.service": { "name": "nfs-idmapd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-mountd.service": { "name": "nfs-mountd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-server.service": { "name": "nfs-server.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-utils.service": { "name": "nfs-utils.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfsdcld.service": { "name": "nfsdcld.service", "source": "systemd", "state": "stopped", "status": "static" }, "nis-domainname.service": { "name": "nis-domainname.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "ntpd.service": { "name": "ntpd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ntpdate.service": { "name": "ntpdate.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "oddjobd.service": { "name": "oddjobd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "packagekit-offline-update.service": { "name": "packagekit-offline-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "packagekit.service": { "name": "packagekit.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-quit-wait.service": { "name": "plymouth-quit-wait.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "plymouth-start.service": { "name": "plymouth-start.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "polkit.service": { "name": "polkit.service", "source": "systemd", "state": "running", "status": "static" }, "qemu-guest-agent.service": { "name": "qemu-guest-agent.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "quotaon.service": { "name": "quotaon.service", "source": "systemd", "state": "inactive", "status": "static" }, "rbdmap.service": { "name": "rbdmap.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "rc-local.service": { "name": "rc-local.service", "source": "systemd", "state": "stopped", "status": "static" }, "rdisc.service": { "name": "rdisc.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rescue.service": { "name": "rescue.service", "source": "systemd", "state": "stopped", "status": "static" }, "rhcd.service": { "name": "rhcd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhsm-facts.service": { "name": "rhsm-facts.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhsm.service": { "name": "rhsm.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhsmcertd.service": { "name": "rhsmcertd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpc-gssd.service": { "name": "rpc-gssd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd-notify.service": { "name": "rpc-statd-notify.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd.service": { "name": "rpc-statd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpcbind.service": { "name": "rpcbind.service", "source": "systemd", "state": "running", "status": "enabled" }, "rsyslog.service": { "name": "rsyslog.service", "source": "systemd", "state": "running", "status": "enabled" }, "selinux-autorelabel-mark.service": { "name": "selinux-autorelabel-mark.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "selinux-autorelabel.service": { "name": "selinux-autorelabel.service", "source": "systemd", "state": "inactive", "status": "static" }, "serial-getty@.service": { "name": "serial-getty@.service", "source": "systemd", "state": "unknown", "status": "indirect" }, "serial-getty@ttyS0.service": { "name": "serial-getty@ttyS0.service", "source": "systemd", "state": "running", "status": "active" }, "snapd.seeded.service": { "name": "snapd.seeded.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sntp.service": { "name": "sntp.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sshd-keygen.service": { "name": "sshd-keygen.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sshd-keygen@.service": { "name": "sshd-keygen@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@ecdsa.service": { "name": "sshd-keygen@ecdsa.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd-keygen@ed25519.service": { "name": "sshd-keygen@ed25519.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd-keygen@rsa.service": { "name": "sshd-keygen@rsa.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd.service": { "name": "sshd.service", "source": "systemd", "state": "running", "status": "enabled" }, "sshd@.service": { "name": "sshd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "sssd-autofs.service": { "name": "sssd-autofs.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-kcm.service": { "name": "sssd-kcm.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "sssd-nss.service": { "name": "sssd-nss.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pac.service": { "name": "sssd-pac.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pam.service": { "name": "sssd-pam.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-ssh.service": { "name": "sssd-ssh.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-sudo.service": { "name": "sssd-sudo.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd.service": { "name": "sssd.service", "source": "systemd", "state": "running", "status": "enabled" }, "syslog.service": { "name": "syslog.service", "source": "systemd", "state": "active", "status": "enabled" }, "system-update-cleanup.service": { "name": "system-update-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-ask-password-console.service": { "name": "systemd-ask-password-console.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-wall.service": { "name": "systemd-ask-password-wall.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-backlight@.service": { "name": "systemd-backlight@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-binfmt.service": { "name": "systemd-binfmt.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-coredump@.service": { "name": "systemd-coredump@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-exit.service": { "name": "systemd-exit.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-firstboot.service": { "name": "systemd-firstboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck-root.service": { "name": "systemd-fsck-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck@.service": { "name": "systemd-fsck@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-fsck@dev-disk-by\\x2duuid-7B77\\x2d95E7.service": { "name": "systemd-fsck@dev-disk-by\\x2duuid-7B77\\x2d95E7.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-fsck@dev-vdb1.service": { "name": "systemd-fsck@dev-vdb1.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-fsck@dev-vdc1.service": { "name": "systemd-fsck@dev-vdc1.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-halt.service": { "name": "systemd-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hibernate-resume@.service": { "name": "systemd-hibernate-resume@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-hibernate.service": { "name": "systemd-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hostnamed.service": { "name": "systemd-hostnamed.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-hwdb-update.service": { "name": "systemd-hwdb-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-hybrid-sleep.service": { "name": "systemd-hybrid-sleep.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-initctl.service": { "name": "systemd-initctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-catalog-update.service": { "name": "systemd-journal-catalog-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-flush.service": { "name": "systemd-journal-flush.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journald.service": { "name": "systemd-journald.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-kexec.service": { "name": "systemd-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-localed.service": { "name": "systemd-localed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-logind.service": { "name": "systemd-logind.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-machine-id-commit.service": { "name": "systemd-machine-id-commit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-modules-load.service": { "name": "systemd-modules-load.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-networkd-wait-online.service": { "name": "systemd-networkd-wait-online.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "systemd-portabled.service": { "name": "systemd-portabled.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-poweroff.service": { "name": "systemd-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-quotacheck.service": { "name": "systemd-quotacheck.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-random-seed.service": { "name": "systemd-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-reboot.service": { "name": "systemd-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-remount-fs.service": { "name": "systemd-remount-fs.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-resolved.service": { "name": "systemd-resolved.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-rfkill.service": { "name": "systemd-rfkill.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend-then-hibernate.service": { "name": "systemd-suspend-then-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend.service": { "name": "systemd-suspend.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-sysctl.service": { "name": "systemd-sysctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-sysusers.service": { "name": "systemd-sysusers.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-timedated.service": { "name": "systemd-timedated.service", "source": "systemd", "state": "inactive", "status": "masked" }, "systemd-timesyncd.service": { "name": "systemd-timesyncd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "systemd-tmpfiles-clean.service": { "name": "systemd-tmpfiles-clean.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev.service": { "name": "systemd-tmpfiles-setup-dev.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup.service": { "name": "systemd-tmpfiles-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-settle.service": { "name": "systemd-udev-settle.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-trigger.service": { "name": "systemd-udev-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udevd.service": { "name": "systemd-udevd.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-update-done.service": { "name": "systemd-update-done.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp-runlevel.service": { "name": "systemd-update-utmp-runlevel.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp.service": { "name": "systemd-update-utmp.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-user-sessions.service": { "name": "systemd-user-sessions.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-vconsole-setup.service": { "name": "systemd-vconsole-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-volatile-root.service": { "name": "systemd-volatile-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "tcsd.service": { "name": "tcsd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "teamd@.service": { "name": "teamd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "timedatex.service": { "name": "timedatex.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "tuned.service": { "name": "tuned.service", "source": "systemd", "state": "running", "status": "enabled" }, "unbound-anchor.service": { "name": "unbound-anchor.service", "source": "systemd", "state": "stopped", "status": "static" }, "user-runtime-dir@.service": { "name": "user-runtime-dir@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user-runtime-dir@0.service": { "name": "user-runtime-dir@0.service", "source": "systemd", "state": "stopped", "status": "active" }, "user@.service": { "name": "user@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user@0.service": { "name": "user@0.service", "source": "systemd", "state": "running", "status": "active" }, "vdo-start-by-dev@.service": { "name": "vdo-start-by-dev@.service", "source": "systemd", "state": "unknown", "status": "static" }, "vdo.service": { "name": "vdo.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "ypbind.service": { "name": "ypbind.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "yppasswdd.service": { "name": "yppasswdd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ypserv.service": { "name": "ypserv.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ypxfrd.service": { "name": "ypxfrd.service", "source": "systemd", "state": "stopped", "status": "not-found" } } }, "changed": false } TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************ task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:58 Thursday 21 July 2022 15:22:58 +0000 (0:00:01.702) 0:00:26.586 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] ******* task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:71 Thursday 21 July 2022 15:22:58 +0000 (0:00:00.063) 0:00:26.650 ********* TASK [linux-system-roles.storage : manage the pools and volumes to match the specified state] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77 Thursday 21 July 2022 15:22:58 +0000 (0:00:00.023) 0:00:26.674 ********* changed: [/cache/rhel-8-y.qcow2.snap] => { "actions": [ { "action": "create format", "device": "/dev/sda", "fs_type": "lvmpv" }, { "action": "create device", "device": "/dev/foo", "fs_type": null }, { "action": "create device", "device": "/dev/mapper/foo-test2", "fs_type": null }, { "action": "create format", "device": "/dev/mapper/foo-test2", "fs_type": "ext4" }, { "action": "create device", "device": "/dev/mapper/foo-test1", "fs_type": null }, { "action": "create format", "device": "/dev/mapper/foo-test1", "fs_type": "xfs" } ], "changed": true, "crypts": [], "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/vda2", "/dev/vda3", "/dev/sdb", "/dev/sdc", "/dev/nvme0n1", "/dev/nvme1n1", "/dev/nvme2n1", "/dev/vdb1", "/dev/vdc1", "/dev/vdd", "/dev/vde", "/dev/vdf", "/dev/mapper/foo-test1", "/dev/mapper/foo-test2" ], "mounts": [ { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" }, { "dump": 0, "fstype": "ext4", "opts": "defaults", "passno": 0, "path": "/opt/test2", "src": "/dev/mapper/foo-test2", "state": "mounted" } ], "packages": [ "xfsprogs", "dosfstools", "lvm2", "e2fsprogs" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "60%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test2", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test2", "_raw_device": "/dev/mapper/foo-test2", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "40%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } TASK [linux-system-roles.storage : Workaround for udev issue on some platforms] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:91 Thursday 21 July 2022 15:23:00 +0000 (0:00:02.267) 0:00:28.941 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] ***** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:103 Thursday 21 July 2022 15:23:00 +0000 (0:00:00.038) 0:00:28.980 ********* TASK [linux-system-roles.storage : show blivet_output] ************************* task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:109 Thursday 21 July 2022 15:23:00 +0000 (0:00:00.021) 0:00:29.001 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "blivet_output": { "actions": [ { "action": "create format", "device": "/dev/sda", "fs_type": "lvmpv" }, { "action": "create device", "device": "/dev/foo", "fs_type": null }, { "action": "create device", "device": "/dev/mapper/foo-test2", "fs_type": null }, { "action": "create format", "device": "/dev/mapper/foo-test2", "fs_type": "ext4" }, { "action": "create device", "device": "/dev/mapper/foo-test1", "fs_type": null }, { "action": "create format", "device": "/dev/mapper/foo-test1", "fs_type": "xfs" } ], "changed": true, "crypts": [], "failed": false, "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/vda2", "/dev/vda3", "/dev/sdb", "/dev/sdc", "/dev/nvme0n1", "/dev/nvme1n1", "/dev/nvme2n1", "/dev/vdb1", "/dev/vdc1", "/dev/vdd", "/dev/vde", "/dev/vdf", "/dev/mapper/foo-test1", "/dev/mapper/foo-test2" ], "mounts": [ { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" }, { "dump": 0, "fstype": "ext4", "opts": "defaults", "passno": 0, "path": "/opt/test2", "src": "/dev/mapper/foo-test2", "state": "mounted" } ], "packages": [ "xfsprogs", "dosfstools", "lvm2", "e2fsprogs" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "60%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test2", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test2", "_raw_device": "/dev/mapper/foo-test2", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "40%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } } TASK [linux-system-roles.storage : set the list of pools for test verification] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:114 Thursday 21 July 2022 15:23:00 +0000 (0:00:00.073) 0:00:29.075 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "60%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test2", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test2", "_raw_device": "/dev/mapper/foo-test2", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "40%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] }, "changed": false } TASK [linux-system-roles.storage : set the list of volumes for test verification] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:118 Thursday 21 July 2022 15:23:00 +0000 (0:00:00.073) 0:00:29.149 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : remove obsolete mounts] ********************* task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:134 Thursday 21 July 2022 15:23:00 +0000 (0:00:00.034) 0:00:29.184 ********* TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:146 Thursday 21 July 2022 15:23:00 +0000 (0:00:00.039) 0:00:29.223 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : set up new/current mounts] ****************** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:151 Thursday 21 July 2022 15:23:01 +0000 (0:00:00.872) 0:00:30.095 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount changed: [/cache/rhel-8-y.qcow2.snap] => (item={'src': '/dev/mapper/foo-test1', 'path': '/opt/test1', 'fstype': 'xfs', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted'}) => { "ansible_loop_var": "mount_info", "backup_file": "", "boot": "yes", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" }, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test1" } redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount changed: [/cache/rhel-8-y.qcow2.snap] => (item={'src': '/dev/mapper/foo-test2', 'path': '/opt/test2', 'fstype': 'ext4', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted'}) => { "ansible_loop_var": "mount_info", "backup_file": "", "boot": "yes", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "ext4", "mount_info": { "dump": 0, "fstype": "ext4", "opts": "defaults", "passno": 0, "path": "/opt/test2", "src": "/dev/mapper/foo-test2", "state": "mounted" }, "name": "/opt/test2", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test2" } TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:163 Thursday 21 July 2022 15:23:02 +0000 (0:00:00.928) 0:00:31.023 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : retrieve facts for the /etc/crypttab file] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:171 Thursday 21 July 2022 15:23:03 +0000 (0:00:00.652) 0:00:31.676 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "stat": { "exists": false } } TASK [linux-system-roles.storage : manage /etc/crypttab to account for changes we just made] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:176 Thursday 21 July 2022 15:23:03 +0000 (0:00:00.389) 0:00:32.066 ********* TASK [linux-system-roles.storage : Update facts] ******************************* task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:198 Thursday 21 July 2022 15:23:03 +0000 (0:00:00.023) 0:00:32.089 ********* ok: [/cache/rhel-8-y.qcow2.snap] META: role_complete for /cache/rhel-8-y.qcow2.snap TASK [include_tasks] *********************************************************** task path: /tmp/tmpafnwmbol/tests/tests_lvm_percent_size.yml:79 Thursday 21 July 2022 15:23:04 +0000 (0:00:01.023) 0:00:33.113 ********* included: /tmp/tmpafnwmbol/tests/verify-role-results.yml for /cache/rhel-8-y.qcow2.snap TASK [Print out pool information] ********************************************** task path: /tmp/tmpafnwmbol/tests/verify-role-results.yml:1 Thursday 21 July 2022 15:23:04 +0000 (0:00:00.041) 0:00:33.155 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "60%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test2", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test2", "_raw_device": "/dev/mapper/foo-test2", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "40%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] } TASK [Print out volume information] ******************************************** task path: /tmp/tmpafnwmbol/tests/verify-role-results.yml:6 Thursday 21 July 2022 15:23:04 +0000 (0:00:00.052) 0:00:33.207 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => {} TASK [Collect info about the volumes.] ***************************************** task path: /tmp/tmpafnwmbol/tests/verify-role-results.yml:14 Thursday 21 July 2022 15:23:04 +0000 (0:00:00.036) 0:00:33.244 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "info": { "/dev/mapper/foo-test1": { "fstype": "xfs", "label": "", "name": "/dev/mapper/foo-test1", "size": "6G", "type": "lvm", "uuid": "3db40ee5-790d-4909-8628-933255bd71de" }, "/dev/mapper/foo-test2": { "fstype": "ext4", "label": "", "name": "/dev/mapper/foo-test2", "size": "4G", "type": "lvm", "uuid": "ee7d1310-5176-4ec6-80b8-28181ffffed3" }, "/dev/nvme0n1": { "fstype": "", "label": "", "name": "/dev/nvme0n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme1n1": { "fstype": "", "label": "", "name": "/dev/nvme1n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme2n1": { "fstype": "", "label": "", "name": "/dev/nvme2n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sda": { "fstype": "LVM2_member", "label": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "irzZ3c-YIeb-eoxj-fRdj-JSvm-ls7D-iZNahz" }, "/dev/sdb": { "fstype": "", "label": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc": { "fstype": "", "label": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sr0": { "fstype": "iso9660", "label": "cidata", "name": "/dev/sr0", "size": "364K", "type": "rom", "uuid": "2022-07-21-15-22-18-00" }, "/dev/vda": { "fstype": "", "label": "", "name": "/dev/vda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vda1": { "fstype": "", "label": "", "name": "/dev/vda1", "size": "1M", "type": "partition", "uuid": "" }, "/dev/vda2": { "fstype": "vfat", "label": "", "name": "/dev/vda2", "size": "100M", "type": "partition", "uuid": "7B77-95E7" }, "/dev/vda3": { "fstype": "xfs", "label": "root", "name": "/dev/vda3", "size": "9.9G", "type": "partition", "uuid": "58dca2db-4ccd-4171-a017-dd53ad8c1c54" }, "/dev/vdb": { "fstype": "", "label": "", "name": "/dev/vdb", "size": "2G", "type": "disk", "uuid": "" }, "/dev/vdb1": { "fstype": "ext4", "label": "yumcache", "name": "/dev/vdb1", "size": "2G", "type": "partition", "uuid": "4d1ed9a1-5608-4ff8-ad89-21d95caa66fa" }, "/dev/vdc": { "fstype": "", "label": "", "name": "/dev/vdc", "size": "2G", "type": "disk", "uuid": "" }, "/dev/vdc1": { "fstype": "ext4", "label": "yumvarlib", "name": "/dev/vdc1", "size": "2G", "type": "partition", "uuid": "a7abcc97-edc5-476c-8ccd-694b4621fd28" }, "/dev/vdd": { "fstype": "", "label": "", "name": "/dev/vdd", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vde": { "fstype": "", "label": "", "name": "/dev/vde", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vdf": { "fstype": "", "label": "", "name": "/dev/vdf", "size": "10G", "type": "disk", "uuid": "" } } } TASK [Read the /etc/fstab file for volume existence] *************************** task path: /tmp/tmpafnwmbol/tests/verify-role-results.yml:19 Thursday 21 July 2022 15:23:05 +0000 (0:00:00.494) 0:00:33.738 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "cmd": [ "cat", "/etc/fstab" ], "delta": "0:00:00.003842", "end": "2022-07-21 11:23:05.637624", "rc": 0, "start": "2022-07-21 11:23:05.633782" } STDOUT: UUID=58dca2db-4ccd-4171-a017-dd53ad8c1c54 / xfs defaults 0 0 UUID=7B77-95E7 /boot/efi vfat defaults,uid=0,gid=0,umask=077,shortname=winnt 0 2 /dev/vdb1 /var/cache/dnf auto defaults,nofail,x-systemd.requires=cloud-init.service,_netdev,comment=cloudconfig 0 2 /dev/vdc1 /var/lib/dnf auto defaults,nofail,x-systemd.requires=cloud-init.service,_netdev,comment=cloudconfig 0 2 /dev/mapper/foo-test1 /opt/test1 xfs defaults 0 0 /dev/mapper/foo-test2 /opt/test2 ext4 defaults 0 0 TASK [Read the /etc/crypttab file] ********************************************* task path: /tmp/tmpafnwmbol/tests/verify-role-results.yml:24 Thursday 21 July 2022 15:23:05 +0000 (0:00:00.505) 0:00:34.244 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "cmd": [ "cat", "/etc/crypttab" ], "delta": "0:00:00.002761", "end": "2022-07-21 11:23:06.034332", "failed_when_result": false, "rc": 1, "start": "2022-07-21 11:23:06.031571" } STDERR: cat: /etc/crypttab: No such file or directory MSG: non-zero return code TASK [Verify the volumes listed in storage_pools were correctly managed] ******* task path: /tmp/tmpafnwmbol/tests/verify-role-results.yml:33 Thursday 21 July 2022 15:23:06 +0000 (0:00:00.396) 0:00:34.640 ********* included: /tmp/tmpafnwmbol/tests/test-verify-pool.yml for /cache/rhel-8-y.qcow2.snap => (item={'disks': ['sda'], 'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'name': 'foo', 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_metadata_version': None, 'state': 'present', 'type': 'lvm', 'volumes': [{'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'name': 'test1', 'raid_level': None, 'size': '60%', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': [], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test1', '_raw_device': '/dev/mapper/foo-test1', '_mount_id': '/dev/mapper/foo-test1', '_kernel_device': '/dev/dm-1', '_raw_kernel_device': '/dev/dm-1'}, {'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'ext4', 'mount_options': 'defaults', 'mount_point': '/opt/test2', 'name': 'test2', 'raid_level': None, 'size': '40%', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': [], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test2', '_raw_device': '/dev/mapper/foo-test2', '_mount_id': '/dev/mapper/foo-test2', '_kernel_device': '/dev/dm-0', '_raw_kernel_device': '/dev/dm-0'}], 'raid_chunk_size': None}) TASK [Set _storage_pool_tests] ************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-pool.yml:5 Thursday 21 July 2022 15:23:06 +0000 (0:00:00.060) 0:00:34.701 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_pool_tests": [ "members", "volumes" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool.yml:18 Thursday 21 July 2022 15:23:06 +0000 (0:00:00.064) 0:00:34.765 ********* included: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml for /cache/rhel-8-y.qcow2.snap => (item=members) included: /tmp/tmpafnwmbol/tests/test-verify-pool-volumes.yml for /cache/rhel-8-y.qcow2.snap => (item=volumes) TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:1 Thursday 21 July 2022 15:23:06 +0000 (0:00:00.087) 0:00:34.853 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_pv_count": "1", "_storage_test_pool_pvs_lvm": [ "/dev/sda" ] }, "changed": false } TASK [Get the canonical device path for each member device] ******************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:10 Thursday 21 July 2022 15:23:06 +0000 (0:00:00.058) 0:00:34.911 ********* ok: [/cache/rhel-8-y.qcow2.snap] => (item=/dev/sda) => { "ansible_loop_var": "pv", "changed": false, "device": "/dev/sda", "pv": "/dev/sda" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:19 Thursday 21 July 2022 15:23:06 +0000 (0:00:00.470) 0:00:35.382 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "__pvs_lvm_len": "1" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:23 Thursday 21 July 2022 15:23:06 +0000 (0:00:00.050) 0:00:35.433 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_pool_pvs": [ "/dev/sda" ] }, "changed": false } TASK [Verify PV count] ********************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:27 Thursday 21 July 2022 15:23:06 +0000 (0:00:00.049) 0:00:35.483 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:34 Thursday 21 July 2022 15:23:07 +0000 (0:00:00.048) 0:00:35.531 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:38 Thursday 21 July 2022 15:23:07 +0000 (0:00:00.039) 0:00:35.571 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:42 Thursday 21 July 2022 15:23:07 +0000 (0:00:00.065) 0:00:35.637 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check the type of each PV] *********************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:46 Thursday 21 July 2022 15:23:07 +0000 (0:00:00.027) 0:00:35.664 ********* ok: [/cache/rhel-8-y.qcow2.snap] => (item=/dev/sda) => { "ansible_loop_var": "pv", "changed": false, "pv": "/dev/sda" } MSG: All assertions passed TASK [Check MD RAID] *********************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:56 Thursday 21 July 2022 15:23:07 +0000 (0:00:00.043) 0:00:35.707 ********* included: /tmp/tmpafnwmbol/tests/verify-pool-md.yml for /cache/rhel-8-y.qcow2.snap TASK [get information about RAID] ********************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-md.yml:6 Thursday 21 July 2022 15:23:07 +0000 (0:00:00.042) 0:00:35.750 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-md.yml:12 Thursday 21 July 2022 15:23:07 +0000 (0:00:00.026) 0:00:35.776 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-md.yml:16 Thursday 21 July 2022 15:23:07 +0000 (0:00:00.024) 0:00:35.800 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-md.yml:20 Thursday 21 July 2022 15:23:07 +0000 (0:00:00.023) 0:00:35.824 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-md.yml:24 Thursday 21 July 2022 15:23:07 +0000 (0:00:00.023) 0:00:35.848 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-md.yml:30 Thursday 21 July 2022 15:23:07 +0000 (0:00:00.022) 0:00:35.871 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmpafnwmbol/tests/verify-pool-md.yml:36 Thursday 21 July 2022 15:23:07 +0000 (0:00:00.025) 0:00:35.896 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-md.yml:44 Thursday 21 July 2022 15:23:07 +0000 (0:00:00.025) 0:00:35.921 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_md_active_devices_re": null, "storage_test_md_metadata_version_re": null, "storage_test_md_spare_devices_re": null }, "changed": false } TASK [Check LVM RAID] ********************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:59 Thursday 21 July 2022 15:23:07 +0000 (0:00:00.037) 0:00:35.959 ********* included: /tmp/tmpafnwmbol/tests/verify-pool-members-lvmraid.yml for /cache/rhel-8-y.qcow2.snap TASK [Validate pool member LVM RAID settings] ********************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-members-lvmraid.yml:1 Thursday 21 July 2022 15:23:07 +0000 (0:00:00.045) 0:00:36.004 ********* included: /tmp/tmpafnwmbol/tests/verify-pool-member-lvmraid.yml for /cache/rhel-8-y.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'name': 'test1', 'raid_level': None, 'size': '60%', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': [], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test1', '_raw_device': '/dev/mapper/foo-test1', '_mount_id': '/dev/mapper/foo-test1', '_kernel_device': '/dev/dm-1', '_raw_kernel_device': '/dev/dm-1'}) included: /tmp/tmpafnwmbol/tests/verify-pool-member-lvmraid.yml for /cache/rhel-8-y.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'ext4', 'mount_options': 'defaults', 'mount_point': '/opt/test2', 'name': 'test2', 'raid_level': None, 'size': '40%', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': [], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test2', '_raw_device': '/dev/mapper/foo-test2', '_mount_id': '/dev/mapper/foo-test2', '_kernel_device': '/dev/dm-0', '_raw_kernel_device': '/dev/dm-0'}) TASK [Get information about LVM RAID] ****************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-lvmraid.yml:3 Thursday 21 July 2022 15:23:07 +0000 (0:00:00.049) 0:00:36.054 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is LVM RAID] ******************************************* task path: /tmp/tmpafnwmbol/tests/verify-pool-member-lvmraid.yml:8 Thursday 21 July 2022 15:23:07 +0000 (0:00:00.029) 0:00:36.083 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-lvmraid.yml:12 Thursday 21 July 2022 15:23:07 +0000 (0:00:00.028) 0:00:36.111 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about LVM RAID] ****************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-lvmraid.yml:3 Thursday 21 July 2022 15:23:07 +0000 (0:00:00.027) 0:00:36.139 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is LVM RAID] ******************************************* task path: /tmp/tmpafnwmbol/tests/verify-pool-member-lvmraid.yml:8 Thursday 21 July 2022 15:23:07 +0000 (0:00:00.026) 0:00:36.165 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-lvmraid.yml:12 Thursday 21 July 2022 15:23:07 +0000 (0:00:00.025) 0:00:36.190 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check Thin Pools] ******************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:62 Thursday 21 July 2022 15:23:07 +0000 (0:00:00.025) 0:00:36.216 ********* included: /tmp/tmpafnwmbol/tests/verify-pool-members-thin.yml for /cache/rhel-8-y.qcow2.snap TASK [Validate pool member thinpool settings] ********************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-members-thin.yml:1 Thursday 21 July 2022 15:23:07 +0000 (0:00:00.044) 0:00:36.260 ********* included: /tmp/tmpafnwmbol/tests/verify-pool-member-thin.yml for /cache/rhel-8-y.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'name': 'test1', 'raid_level': None, 'size': '60%', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': [], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test1', '_raw_device': '/dev/mapper/foo-test1', '_mount_id': '/dev/mapper/foo-test1', '_kernel_device': '/dev/dm-1', '_raw_kernel_device': '/dev/dm-1'}) included: /tmp/tmpafnwmbol/tests/verify-pool-member-thin.yml for /cache/rhel-8-y.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'ext4', 'mount_options': 'defaults', 'mount_point': '/opt/test2', 'name': 'test2', 'raid_level': None, 'size': '40%', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': [], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test2', '_raw_device': '/dev/mapper/foo-test2', '_mount_id': '/dev/mapper/foo-test2', '_kernel_device': '/dev/dm-0', '_raw_kernel_device': '/dev/dm-0'}) TASK [Get information about thinpool] ****************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-thin.yml:3 Thursday 21 July 2022 15:23:07 +0000 (0:00:00.048) 0:00:36.308 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-thin.yml:8 Thursday 21 July 2022 15:23:07 +0000 (0:00:00.053) 0:00:36.362 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-thin.yml:13 Thursday 21 July 2022 15:23:07 +0000 (0:00:00.023) 0:00:36.386 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-thin.yml:17 Thursday 21 July 2022 15:23:07 +0000 (0:00:00.024) 0:00:36.410 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about thinpool] ****************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-thin.yml:3 Thursday 21 July 2022 15:23:07 +0000 (0:00:00.023) 0:00:36.434 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-thin.yml:8 Thursday 21 July 2022 15:23:07 +0000 (0:00:00.023) 0:00:36.457 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-thin.yml:13 Thursday 21 July 2022 15:23:07 +0000 (0:00:00.021) 0:00:36.479 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-thin.yml:17 Thursday 21 July 2022 15:23:07 +0000 (0:00:00.021) 0:00:36.501 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check member encryption] ************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:65 Thursday 21 July 2022 15:23:08 +0000 (0:00:00.021) 0:00:36.523 ********* included: /tmp/tmpafnwmbol/tests/verify-pool-members-encryption.yml for /cache/rhel-8-y.qcow2.snap TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-members-encryption.yml:4 Thursday 21 July 2022 15:23:08 +0000 (0:00:00.045) 0:00:36.568 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Validate pool member LUKS settings] ************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-members-encryption.yml:8 Thursday 21 July 2022 15:23:08 +0000 (0:00:00.048) 0:00:36.616 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => (item=/dev/sda) => { "_storage_test_pool_member_path": "/dev/sda", "ansible_loop_var": "_storage_test_pool_member_path", "changed": false, "skip_reason": "Conditional result was False" } TASK [Validate pool member crypttab entries] *********************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-members-encryption.yml:15 Thursday 21 July 2022 15:23:08 +0000 (0:00:00.028) 0:00:36.644 ********* included: /tmp/tmpafnwmbol/tests/verify-pool-member-crypttab.yml for /cache/rhel-8-y.qcow2.snap => (item=/dev/sda) TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-crypttab.yml:1 Thursday 21 July 2022 15:23:08 +0000 (0:00:00.041) 0:00:36.686 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": [] }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpafnwmbol/tests/verify-pool-member-crypttab.yml:6 Thursday 21 July 2022 15:23:08 +0000 (0:00:00.046) 0:00:36.732 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpafnwmbol/tests/verify-pool-member-crypttab.yml:11 Thursday 21 July 2022 15:23:08 +0000 (0:00:00.048) 0:00:36.781 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-crypttab.yml:17 Thursday 21 July 2022 15:23:08 +0000 (0:00:00.036) 0:00:36.818 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-crypttab.yml:23 Thursday 21 July 2022 15:23:08 +0000 (0:00:00.038) 0:00:36.857 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-crypttab.yml:29 Thursday 21 July 2022 15:23:08 +0000 (0:00:00.039) 0:00:36.896 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-members-encryption.yml:22 Thursday 21 July 2022 15:23:08 +0000 (0:00:00.034) 0:00:36.930 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_crypttab_key_file": null }, "changed": false } TASK [Check VDO] *************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:68 Thursday 21 July 2022 15:23:08 +0000 (0:00:00.032) 0:00:36.963 ********* included: /tmp/tmpafnwmbol/tests/verify-pool-members-vdo.yml for /cache/rhel-8-y.qcow2.snap TASK [Validate pool member VDO settings] *************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-members-vdo.yml:1 Thursday 21 July 2022 15:23:08 +0000 (0:00:00.048) 0:00:37.012 ********* included: /tmp/tmpafnwmbol/tests/verify-pool-member-vdo.yml for /cache/rhel-8-y.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'name': 'test1', 'raid_level': None, 'size': '60%', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': [], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test1', '_raw_device': '/dev/mapper/foo-test1', '_mount_id': '/dev/mapper/foo-test1', '_kernel_device': '/dev/dm-1', '_raw_kernel_device': '/dev/dm-1'}) included: /tmp/tmpafnwmbol/tests/verify-pool-member-vdo.yml for /cache/rhel-8-y.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'ext4', 'mount_options': 'defaults', 'mount_point': '/opt/test2', 'name': 'test2', 'raid_level': None, 'size': '40%', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': [], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test2', '_raw_device': '/dev/mapper/foo-test2', '_mount_id': '/dev/mapper/foo-test2', '_kernel_device': '/dev/dm-0', '_raw_kernel_device': '/dev/dm-0'}) TASK [get information about VDO deduplication] ********************************* task path: /tmp/tmpafnwmbol/tests/verify-pool-member-vdo.yml:3 Thursday 21 July 2022 15:23:08 +0000 (0:00:00.051) 0:00:37.064 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-vdo.yml:8 Thursday 21 July 2022 15:23:08 +0000 (0:00:00.024) 0:00:37.088 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-vdo.yml:11 Thursday 21 July 2022 15:23:08 +0000 (0:00:00.023) 0:00:37.111 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-vdo.yml:16 Thursday 21 July 2022 15:23:08 +0000 (0:00:00.022) 0:00:37.134 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-vdo.yml:21 Thursday 21 July 2022 15:23:08 +0000 (0:00:00.024) 0:00:37.159 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-vdo.yml:24 Thursday 21 July 2022 15:23:08 +0000 (0:00:00.026) 0:00:37.186 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-vdo.yml:29 Thursday 21 July 2022 15:23:08 +0000 (0:00:00.024) 0:00:37.210 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-vdo.yml:39 Thursday 21 July 2022 15:23:08 +0000 (0:00:00.024) 0:00:37.235 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [get information about VDO deduplication] ********************************* task path: /tmp/tmpafnwmbol/tests/verify-pool-member-vdo.yml:3 Thursday 21 July 2022 15:23:08 +0000 (0:00:00.034) 0:00:37.269 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-vdo.yml:8 Thursday 21 July 2022 15:23:08 +0000 (0:00:00.026) 0:00:37.296 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-vdo.yml:11 Thursday 21 July 2022 15:23:08 +0000 (0:00:00.024) 0:00:37.320 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-vdo.yml:16 Thursday 21 July 2022 15:23:08 +0000 (0:00:00.023) 0:00:37.343 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-vdo.yml:21 Thursday 21 July 2022 15:23:08 +0000 (0:00:00.023) 0:00:37.367 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-vdo.yml:24 Thursday 21 July 2022 15:23:08 +0000 (0:00:00.027) 0:00:37.394 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-vdo.yml:29 Thursday 21 July 2022 15:23:08 +0000 (0:00:00.024) 0:00:37.418 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-vdo.yml:39 Thursday 21 July 2022 15:23:08 +0000 (0:00:00.023) 0:00:37.442 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:71 Thursday 21 July 2022 15:23:09 +0000 (0:00:00.079) 0:00:37.521 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "__pvs_lvm_len": null, "_storage_test_expected_pv_count": null, "_storage_test_expected_pv_type": null, "_storage_test_pool_pvs": [], "_storage_test_pool_pvs_lvm": [] }, "changed": false } TASK [verify the volumes] ****************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool-volumes.yml:3 Thursday 21 July 2022 15:23:09 +0000 (0:00:00.078) 0:00:37.600 ********* included: /tmp/tmpafnwmbol/tests/test-verify-volume.yml for /cache/rhel-8-y.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'name': 'test1', 'raid_level': None, 'size': '60%', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': [], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test1', '_raw_device': '/dev/mapper/foo-test1', '_mount_id': '/dev/mapper/foo-test1', '_kernel_device': '/dev/dm-1', '_raw_kernel_device': '/dev/dm-1'}) included: /tmp/tmpafnwmbol/tests/test-verify-volume.yml for /cache/rhel-8-y.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'ext4', 'mount_options': 'defaults', 'mount_point': '/opt/test2', 'name': 'test2', 'raid_level': None, 'size': '40%', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': [], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test2', '_raw_device': '/dev/mapper/foo-test2', '_mount_id': '/dev/mapper/foo-test2', '_kernel_device': '/dev/dm-0', '_raw_kernel_device': '/dev/dm-0'}) TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume.yml:2 Thursday 21 July 2022 15:23:09 +0000 (0:00:00.048) 0:00:37.648 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_volume_present": true, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume.yml:10 Thursday 21 July 2022 15:23:09 +0000 (0:00:00.052) 0:00:37.701 ********* included: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml for /cache/rhel-8-y.qcow2.snap => (item=mount) included: /tmp/tmpafnwmbol/tests/test-verify-volume-fstab.yml for /cache/rhel-8-y.qcow2.snap => (item=fstab) included: /tmp/tmpafnwmbol/tests/test-verify-volume-fs.yml for /cache/rhel-8-y.qcow2.snap => (item=fs) included: /tmp/tmpafnwmbol/tests/test-verify-volume-device.yml for /cache/rhel-8-y.qcow2.snap => (item=device) included: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml for /cache/rhel-8-y.qcow2.snap => (item=encryption) included: /tmp/tmpafnwmbol/tests/test-verify-volume-md.yml for /cache/rhel-8-y.qcow2.snap => (item=md) included: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml for /cache/rhel-8-y.qcow2.snap => (item=size) included: /tmp/tmpafnwmbol/tests/test-verify-volume-cache.yml for /cache/rhel-8-y.qcow2.snap => (item=cache) TASK [Get expected mount device based on device type] ************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml:6 Thursday 21 July 2022 15:23:09 +0000 (0:00:00.085) 0:00:37.786 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/foo-test1" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml:14 Thursday 21 July 2022 15:23:09 +0000 (0:00:00.040) 0:00:37.827 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_mount_device_matches": [ { "block_available": 1551090, "block_size": 4096, "block_total": 1570304, "block_used": 19214, "device": "/dev/mapper/foo-test1", "fstype": "xfs", "inode_available": 3145725, "inode_total": 3145728, "inode_used": 3, "mount": "/opt/test1", "options": "rw,seclabel,relatime,attr2,inode64,logbufs=8,logbsize=32k,noquota", "size_available": 6353264640, "size_total": 6431965184, "uuid": "3db40ee5-790d-4909-8628-933255bd71de" } ], "storage_test_mount_expected_match_count": "1", "storage_test_mount_point_matches": [ { "block_available": 1551090, "block_size": 4096, "block_total": 1570304, "block_used": 19214, "device": "/dev/mapper/foo-test1", "fstype": "xfs", "inode_available": 3145725, "inode_total": 3145728, "inode_used": 3, "mount": "/opt/test1", "options": "rw,seclabel,relatime,attr2,inode64,logbufs=8,logbsize=32k,noquota", "size_available": 6353264640, "size_total": 6431965184, "uuid": "3db40ee5-790d-4909-8628-933255bd71de" } ], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Verify the current mount state by device] ******************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml:28 Thursday 21 July 2022 15:23:09 +0000 (0:00:00.055) 0:00:37.883 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify the current mount state by mount point] *************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml:37 Thursday 21 July 2022 15:23:09 +0000 (0:00:00.052) 0:00:37.935 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify the mount fs type] ************************************************ task path: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml:45 Thursday 21 July 2022 15:23:09 +0000 (0:00:00.048) 0:00:37.984 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [command] ***************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml:54 Thursday 21 July 2022 15:23:09 +0000 (0:00:00.049) 0:00:38.033 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml:58 Thursday 21 July 2022 15:23:09 +0000 (0:00:00.023) 0:00:38.056 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml:63 Thursday 21 July 2022 15:23:09 +0000 (0:00:00.022) 0:00:38.079 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml:75 Thursday 21 July 2022 15:23:09 +0000 (0:00:00.021) 0:00:38.101 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-fstab.yml:2 Thursday 21 July 2022 15:23:09 +0000 (0:00:00.033) 0:00:38.135 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "1", "storage_test_fstab_expected_mount_options_matches": "1", "storage_test_fstab_expected_mount_point_matches": "1", "storage_test_fstab_id_matches": [ "/dev/mapper/foo-test1 " ], "storage_test_fstab_mount_options_matches": [ " /opt/test1 xfs defaults " ], "storage_test_fstab_mount_point_matches": [ " /opt/test1 " ] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-fstab.yml:25 Thursday 21 July 2022 15:23:09 +0000 (0:00:00.060) 0:00:38.195 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify the fstab mount point] ******************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-fstab.yml:32 Thursday 21 July 2022 15:23:09 +0000 (0:00:00.049) 0:00:38.245 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-fstab.yml:39 Thursday 21 July 2022 15:23:09 +0000 (0:00:00.049) 0:00:38.294 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-fstab.yml:49 Thursday 21 July 2022 15:23:09 +0000 (0:00:00.035) 0:00:38.330 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-fs.yml:4 Thursday 21 July 2022 15:23:09 +0000 (0:00:00.034) 0:00:38.365 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify fs label] ********************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-fs.yml:10 Thursday 21 July 2022 15:23:09 +0000 (0:00:00.039) 0:00:38.404 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [See whether the device node is present] ********************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-device.yml:4 Thursday 21 July 2022 15:23:09 +0000 (0:00:00.039) 0:00:38.444 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "stat": { "atime": 1658416980.2722602, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1658416980.2722602, "dev": 6, "device_type": 64769, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 62659, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/symlink", "mode": "0660", "mtime": 1658416980.2722602, "nlink": 1, "path": "/dev/mapper/foo-test1", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-device.yml:10 Thursday 21 July 2022 15:23:10 +0000 (0:00:00.386) 0:00:38.830 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about this volume] ********************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-device.yml:18 Thursday 21 July 2022 15:23:10 +0000 (0:00:00.039) 0:00:38.869 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [(1/2) Process volume type (set initial value)] *************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-device.yml:24 Thursday 21 July 2022 15:23:10 +0000 (0:00:00.037) 0:00:38.907 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [(2/2) Process volume type (get RAID value)] ****************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-device.yml:28 Thursday 21 July 2022 15:23:10 +0000 (0:00:00.037) 0:00:38.944 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-device.yml:33 Thursday 21 July 2022 15:23:10 +0000 (0:00:00.024) 0:00:38.969 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Stat the LUKS device, if encrypted] ************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:3 Thursday 21 July 2022 15:23:10 +0000 (0:00:00.040) 0:00:39.009 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:10 Thursday 21 July 2022 15:23:10 +0000 (0:00:00.025) 0:00:39.034 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [Collect LUKS info for this volume] *************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:15 Thursday 21 July 2022 15:23:11 +0000 (0:00:01.334) 0:00:40.369 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:21 Thursday 21 July 2022 15:23:11 +0000 (0:00:00.026) 0:00:40.395 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:30 Thursday 21 July 2022 15:23:11 +0000 (0:00:00.024) 0:00:40.420 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:38 Thursday 21 July 2022 15:23:11 +0000 (0:00:00.052) 0:00:40.473 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:44 Thursday 21 July 2022 15:23:11 +0000 (0:00:00.025) 0:00:40.498 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:49 Thursday 21 July 2022 15:23:12 +0000 (0:00:00.025) 0:00:40.524 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:55 Thursday 21 July 2022 15:23:12 +0000 (0:00:00.025) 0:00:40.549 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:61 Thursday 21 July 2022 15:23:12 +0000 (0:00:00.027) 0:00:40.577 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:67 Thursday 21 July 2022 15:23:12 +0000 (0:00:00.031) 0:00:40.608 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:74 Thursday 21 July 2022 15:23:12 +0000 (0:00:00.057) 0:00:40.666 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:79 Thursday 21 July 2022 15:23:12 +0000 (0:00:00.056) 0:00:40.723 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:85 Thursday 21 July 2022 15:23:12 +0000 (0:00:00.039) 0:00:40.762 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:91 Thursday 21 July 2022 15:23:12 +0000 (0:00:00.039) 0:00:40.801 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:97 Thursday 21 July 2022 15:23:12 +0000 (0:00:00.036) 0:00:40.838 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [get information about RAID] ********************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-md.yml:7 Thursday 21 July 2022 15:23:12 +0000 (0:00:00.039) 0:00:40.878 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-md.yml:13 Thursday 21 July 2022 15:23:12 +0000 (0:00:00.036) 0:00:40.914 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-md.yml:17 Thursday 21 July 2022 15:23:12 +0000 (0:00:00.035) 0:00:40.949 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-md.yml:21 Thursday 21 July 2022 15:23:12 +0000 (0:00:00.035) 0:00:40.984 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-md.yml:25 Thursday 21 July 2022 15:23:12 +0000 (0:00:00.041) 0:00:41.026 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-md.yml:31 Thursday 21 July 2022 15:23:12 +0000 (0:00:00.039) 0:00:41.065 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-md.yml:37 Thursday 21 July 2022 15:23:12 +0000 (0:00:00.035) 0:00:41.101 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the actual size of the volume] ************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:3 Thursday 21 July 2022 15:23:12 +0000 (0:00:00.034) 0:00:41.135 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "bytes": 6442450944, "changed": false, "lvm": "6g", "parted": "6GiB", "size": "6 GiB" } TASK [parse the requested size of the volume] ********************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:9 Thursday 21 July 2022 15:23:13 +0000 (0:00:00.518) 0:00:41.654 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Establish base value for expected size] ********************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:15 Thursday 21 July 2022 15:23:13 +0000 (0:00:00.041) 0:00:41.695 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:20 Thursday 21 July 2022 15:23:13 +0000 (0:00:00.038) 0:00:41.734 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_expected_size": "VARIABLE IS NOT DEFINED!: 'storage_test_expected_size' is undefined" } TASK [debug] ******************************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:25 Thursday 21 July 2022 15:23:13 +0000 (0:00:00.038) 0:00:41.772 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_pool": { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "60%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test2", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test2", "_raw_device": "/dev/mapper/foo-test2", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "40%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } } TASK [debug] ******************************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:28 Thursday 21 July 2022 15:23:13 +0000 (0:00:00.054) 0:00:41.827 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_blkinfo": { "changed": false, "failed": false, "info": { "/dev/mapper/foo-test1": { "fstype": "xfs", "label": "", "name": "/dev/mapper/foo-test1", "size": "6G", "type": "lvm", "uuid": "3db40ee5-790d-4909-8628-933255bd71de" }, "/dev/mapper/foo-test2": { "fstype": "ext4", "label": "", "name": "/dev/mapper/foo-test2", "size": "4G", "type": "lvm", "uuid": "ee7d1310-5176-4ec6-80b8-28181ffffed3" }, "/dev/nvme0n1": { "fstype": "", "label": "", "name": "/dev/nvme0n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme1n1": { "fstype": "", "label": "", "name": "/dev/nvme1n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme2n1": { "fstype": "", "label": "", "name": "/dev/nvme2n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sda": { "fstype": "LVM2_member", "label": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "irzZ3c-YIeb-eoxj-fRdj-JSvm-ls7D-iZNahz" }, "/dev/sdb": { "fstype": "", "label": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc": { "fstype": "", "label": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sr0": { "fstype": "iso9660", "label": "cidata", "name": "/dev/sr0", "size": "364K", "type": "rom", "uuid": "2022-07-21-15-22-18-00" }, "/dev/vda": { "fstype": "", "label": "", "name": "/dev/vda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vda1": { "fstype": "", "label": "", "name": "/dev/vda1", "size": "1M", "type": "partition", "uuid": "" }, "/dev/vda2": { "fstype": "vfat", "label": "", "name": "/dev/vda2", "size": "100M", "type": "partition", "uuid": "7B77-95E7" }, "/dev/vda3": { "fstype": "xfs", "label": "root", "name": "/dev/vda3", "size": "9.9G", "type": "partition", "uuid": "58dca2db-4ccd-4171-a017-dd53ad8c1c54" }, "/dev/vdb": { "fstype": "", "label": "", "name": "/dev/vdb", "size": "2G", "type": "disk", "uuid": "" }, "/dev/vdb1": { "fstype": "ext4", "label": "yumcache", "name": "/dev/vdb1", "size": "2G", "type": "partition", "uuid": "4d1ed9a1-5608-4ff8-ad89-21d95caa66fa" }, "/dev/vdc": { "fstype": "", "label": "", "name": "/dev/vdc", "size": "2G", "type": "disk", "uuid": "" }, "/dev/vdc1": { "fstype": "ext4", "label": "yumvarlib", "name": "/dev/vdc1", "size": "2G", "type": "partition", "uuid": "a7abcc97-edc5-476c-8ccd-694b4621fd28" }, "/dev/vdd": { "fstype": "", "label": "", "name": "/dev/vdd", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vde": { "fstype": "", "label": "", "name": "/dev/vde", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vdf": { "fstype": "", "label": "", "name": "/dev/vdf", "size": "10G", "type": "disk", "uuid": "" } } } } TASK [Get the size of parent/pool device] ************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:31 Thursday 21 July 2022 15:23:13 +0000 (0:00:00.054) 0:00:41.882 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "bytes": 10737418240, "changed": false, "lvm": "10g", "parted": "10GiB", "size": "10 GiB" } TASK [debug] ******************************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:36 Thursday 21 July 2022 15:23:13 +0000 (0:00:00.396) 0:00:42.279 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_pool_size": { "bytes": 10737418240, "changed": false, "failed": false, "lvm": "10g", "parted": "10GiB", "size": "10 GiB" } } TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:39 Thursday 21 July 2022 15:23:13 +0000 (0:00:00.087) 0:00:42.366 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_expected_size": "6442450944.0" }, "changed": false } TASK [debug] ******************************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:44 Thursday 21 July 2022 15:23:13 +0000 (0:00:00.049) 0:00:42.416 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_actual_size": { "bytes": 6442450944, "changed": false, "failed": false, "lvm": "6g", "parted": "6GiB", "size": "6 GiB" } } TASK [debug] ******************************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:47 Thursday 21 July 2022 15:23:13 +0000 (0:00:00.073) 0:00:42.490 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_expected_size": "6442450944.0" } TASK [assert] ****************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:50 Thursday 21 July 2022 15:23:14 +0000 (0:00:00.110) 0:00:42.601 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Get information about the LV] ******************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-cache.yml:6 Thursday 21 July 2022 15:23:14 +0000 (0:00:00.053) 0:00:42.654 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "cmd": [ "lvs", "--noheadings", "--nameprefixes", "--units=b", "--nosuffix", "--unquoted", "-o", "name,attr,cache_total_blocks,chunk_size,segtype", "foo/test1" ], "delta": "0:00:00.034110", "end": "2022-07-21 11:23:14.464048", "rc": 0, "start": "2022-07-21 11:23:14.429938" } STDOUT: LVM2_LV_NAME=test1 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-cache.yml:14 Thursday 21 July 2022 15:23:14 +0000 (0:00:00.415) 0:00:43.070 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_lv_segtype": [ "linear" ] }, "changed": false } TASK [check segment type] ****************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-cache.yml:17 Thursday 21 July 2022 15:23:14 +0000 (0:00:00.049) 0:00:43.119 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-cache.yml:22 Thursday 21 July 2022 15:23:14 +0000 (0:00:00.050) 0:00:43.169 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the requested cache size] ****************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-cache.yml:26 Thursday 21 July 2022 15:23:14 +0000 (0:00:00.036) 0:00:43.206 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-cache.yml:32 Thursday 21 July 2022 15:23:14 +0000 (0:00:00.039) 0:00:43.246 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-cache.yml:36 Thursday 21 July 2022 15:23:14 +0000 (0:00:00.039) 0:00:43.285 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume.yml:16 Thursday 21 July 2022 15:23:14 +0000 (0:00:00.038) 0:00:43.324 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume.yml:2 Thursday 21 July 2022 15:23:14 +0000 (0:00:00.037) 0:00:43.362 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_volume_present": true, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume.yml:10 Thursday 21 July 2022 15:23:14 +0000 (0:00:00.054) 0:00:43.416 ********* included: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml for /cache/rhel-8-y.qcow2.snap => (item=mount) included: /tmp/tmpafnwmbol/tests/test-verify-volume-fstab.yml for /cache/rhel-8-y.qcow2.snap => (item=fstab) included: /tmp/tmpafnwmbol/tests/test-verify-volume-fs.yml for /cache/rhel-8-y.qcow2.snap => (item=fs) included: /tmp/tmpafnwmbol/tests/test-verify-volume-device.yml for /cache/rhel-8-y.qcow2.snap => (item=device) included: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml for /cache/rhel-8-y.qcow2.snap => (item=encryption) included: /tmp/tmpafnwmbol/tests/test-verify-volume-md.yml for /cache/rhel-8-y.qcow2.snap => (item=md) included: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml for /cache/rhel-8-y.qcow2.snap => (item=size) included: /tmp/tmpafnwmbol/tests/test-verify-volume-cache.yml for /cache/rhel-8-y.qcow2.snap => (item=cache) TASK [Get expected mount device based on device type] ************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml:6 Thursday 21 July 2022 15:23:14 +0000 (0:00:00.078) 0:00:43.494 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/foo-test2" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml:14 Thursday 21 July 2022 15:23:15 +0000 (0:00:00.040) 0:00:43.535 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_mount_device_matches": [ { "block_available": 954137, "block_size": 4096, "block_total": 1010616, "block_used": 56479, "device": "/dev/mapper/foo-test2", "fstype": "ext4", "inode_available": 262133, "inode_total": 262144, "inode_used": 11, "mount": "/opt/test2", "options": "rw,seclabel,relatime", "size_available": 3908145152, "size_total": 4139483136, "uuid": "ee7d1310-5176-4ec6-80b8-28181ffffed3" } ], "storage_test_mount_expected_match_count": "1", "storage_test_mount_point_matches": [ { "block_available": 954137, "block_size": 4096, "block_total": 1010616, "block_used": 56479, "device": "/dev/mapper/foo-test2", "fstype": "ext4", "inode_available": 262133, "inode_total": 262144, "inode_used": 11, "mount": "/opt/test2", "options": "rw,seclabel,relatime", "size_available": 3908145152, "size_total": 4139483136, "uuid": "ee7d1310-5176-4ec6-80b8-28181ffffed3" } ], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Verify the current mount state by device] ******************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml:28 Thursday 21 July 2022 15:23:15 +0000 (0:00:00.054) 0:00:43.589 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify the current mount state by mount point] *************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml:37 Thursday 21 July 2022 15:23:15 +0000 (0:00:00.055) 0:00:43.645 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify the mount fs type] ************************************************ task path: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml:45 Thursday 21 July 2022 15:23:15 +0000 (0:00:00.049) 0:00:43.695 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [command] ***************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml:54 Thursday 21 July 2022 15:23:15 +0000 (0:00:00.053) 0:00:43.748 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml:58 Thursday 21 July 2022 15:23:15 +0000 (0:00:00.025) 0:00:43.774 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml:63 Thursday 21 July 2022 15:23:15 +0000 (0:00:00.024) 0:00:43.798 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml:75 Thursday 21 July 2022 15:23:15 +0000 (0:00:00.026) 0:00:43.824 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-fstab.yml:2 Thursday 21 July 2022 15:23:15 +0000 (0:00:00.037) 0:00:43.862 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "1", "storage_test_fstab_expected_mount_options_matches": "1", "storage_test_fstab_expected_mount_point_matches": "1", "storage_test_fstab_id_matches": [ "/dev/mapper/foo-test2 " ], "storage_test_fstab_mount_options_matches": [ " /opt/test2 ext4 defaults " ], "storage_test_fstab_mount_point_matches": [ " /opt/test2 " ] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-fstab.yml:25 Thursday 21 July 2022 15:23:15 +0000 (0:00:00.065) 0:00:43.927 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify the fstab mount point] ******************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-fstab.yml:32 Thursday 21 July 2022 15:23:15 +0000 (0:00:00.051) 0:00:43.978 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-fstab.yml:39 Thursday 21 July 2022 15:23:15 +0000 (0:00:00.050) 0:00:44.029 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-fstab.yml:49 Thursday 21 July 2022 15:23:15 +0000 (0:00:00.041) 0:00:44.070 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-fs.yml:4 Thursday 21 July 2022 15:23:15 +0000 (0:00:00.034) 0:00:44.104 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify fs label] ********************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-fs.yml:10 Thursday 21 July 2022 15:23:15 +0000 (0:00:00.039) 0:00:44.144 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [See whether the device node is present] ********************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-device.yml:4 Thursday 21 July 2022 15:23:15 +0000 (0:00:00.042) 0:00:44.186 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "stat": { "atime": 1658416980.0722601, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1658416980.0722601, "dev": 6, "device_type": 64768, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 62525, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/symlink", "mode": "0660", "mtime": 1658416980.0722601, "nlink": 1, "path": "/dev/mapper/foo-test2", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-device.yml:10 Thursday 21 July 2022 15:23:16 +0000 (0:00:00.377) 0:00:44.564 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about this volume] ********************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-device.yml:18 Thursday 21 July 2022 15:23:16 +0000 (0:00:00.074) 0:00:44.638 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [(1/2) Process volume type (set initial value)] *************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-device.yml:24 Thursday 21 July 2022 15:23:16 +0000 (0:00:00.111) 0:00:44.750 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [(2/2) Process volume type (get RAID value)] ****************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-device.yml:28 Thursday 21 July 2022 15:23:16 +0000 (0:00:00.035) 0:00:44.785 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-device.yml:33 Thursday 21 July 2022 15:23:16 +0000 (0:00:00.025) 0:00:44.810 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Stat the LUKS device, if encrypted] ************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:3 Thursday 21 July 2022 15:23:16 +0000 (0:00:00.050) 0:00:44.861 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:10 Thursday 21 July 2022 15:23:16 +0000 (0:00:00.025) 0:00:44.886 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [Collect LUKS info for this volume] *************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:15 Thursday 21 July 2022 15:23:17 +0000 (0:00:01.294) 0:00:46.180 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:21 Thursday 21 July 2022 15:23:17 +0000 (0:00:00.025) 0:00:46.206 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:30 Thursday 21 July 2022 15:23:17 +0000 (0:00:00.024) 0:00:46.230 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:38 Thursday 21 July 2022 15:23:17 +0000 (0:00:00.052) 0:00:46.283 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:44 Thursday 21 July 2022 15:23:17 +0000 (0:00:00.025) 0:00:46.308 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:49 Thursday 21 July 2022 15:23:17 +0000 (0:00:00.024) 0:00:46.333 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:55 Thursday 21 July 2022 15:23:17 +0000 (0:00:00.025) 0:00:46.358 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:61 Thursday 21 July 2022 15:23:17 +0000 (0:00:00.024) 0:00:46.382 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:67 Thursday 21 July 2022 15:23:17 +0000 (0:00:00.024) 0:00:46.407 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:74 Thursday 21 July 2022 15:23:17 +0000 (0:00:00.049) 0:00:46.456 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:79 Thursday 21 July 2022 15:23:17 +0000 (0:00:00.053) 0:00:46.510 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:85 Thursday 21 July 2022 15:23:18 +0000 (0:00:00.035) 0:00:46.545 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:91 Thursday 21 July 2022 15:23:18 +0000 (0:00:00.034) 0:00:46.579 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:97 Thursday 21 July 2022 15:23:18 +0000 (0:00:00.033) 0:00:46.613 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [get information about RAID] ********************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-md.yml:7 Thursday 21 July 2022 15:23:18 +0000 (0:00:00.033) 0:00:46.647 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-md.yml:13 Thursday 21 July 2022 15:23:18 +0000 (0:00:00.034) 0:00:46.682 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-md.yml:17 Thursday 21 July 2022 15:23:18 +0000 (0:00:00.033) 0:00:46.715 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-md.yml:21 Thursday 21 July 2022 15:23:18 +0000 (0:00:00.033) 0:00:46.749 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-md.yml:25 Thursday 21 July 2022 15:23:18 +0000 (0:00:00.049) 0:00:46.799 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-md.yml:31 Thursday 21 July 2022 15:23:18 +0000 (0:00:00.040) 0:00:46.839 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-md.yml:37 Thursday 21 July 2022 15:23:18 +0000 (0:00:00.044) 0:00:46.884 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the actual size of the volume] ************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:3 Thursday 21 July 2022 15:23:18 +0000 (0:00:00.042) 0:00:46.927 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "bytes": 4294967296, "changed": false, "lvm": "4g", "parted": "4GiB", "size": "4 GiB" } TASK [parse the requested size of the volume] ********************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:9 Thursday 21 July 2022 15:23:18 +0000 (0:00:00.377) 0:00:47.304 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Establish base value for expected size] ********************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:15 Thursday 21 July 2022 15:23:18 +0000 (0:00:00.040) 0:00:47.344 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:20 Thursday 21 July 2022 15:23:18 +0000 (0:00:00.037) 0:00:47.382 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_expected_size": "6442450944.0" } TASK [debug] ******************************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:25 Thursday 21 July 2022 15:23:18 +0000 (0:00:00.039) 0:00:47.421 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_pool": { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "60%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test2", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test2", "_raw_device": "/dev/mapper/foo-test2", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "40%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } } TASK [debug] ******************************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:28 Thursday 21 July 2022 15:23:18 +0000 (0:00:00.054) 0:00:47.476 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_blkinfo": { "changed": false, "failed": false, "info": { "/dev/mapper/foo-test1": { "fstype": "xfs", "label": "", "name": "/dev/mapper/foo-test1", "size": "6G", "type": "lvm", "uuid": "3db40ee5-790d-4909-8628-933255bd71de" }, "/dev/mapper/foo-test2": { "fstype": "ext4", "label": "", "name": "/dev/mapper/foo-test2", "size": "4G", "type": "lvm", "uuid": "ee7d1310-5176-4ec6-80b8-28181ffffed3" }, "/dev/nvme0n1": { "fstype": "", "label": "", "name": "/dev/nvme0n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme1n1": { "fstype": "", "label": "", "name": "/dev/nvme1n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme2n1": { "fstype": "", "label": "", "name": "/dev/nvme2n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sda": { "fstype": "LVM2_member", "label": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "irzZ3c-YIeb-eoxj-fRdj-JSvm-ls7D-iZNahz" }, "/dev/sdb": { "fstype": "", "label": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc": { "fstype": "", "label": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sr0": { "fstype": "iso9660", "label": "cidata", "name": "/dev/sr0", "size": "364K", "type": "rom", "uuid": "2022-07-21-15-22-18-00" }, "/dev/vda": { "fstype": "", "label": "", "name": "/dev/vda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vda1": { "fstype": "", "label": "", "name": "/dev/vda1", "size": "1M", "type": "partition", "uuid": "" }, "/dev/vda2": { "fstype": "vfat", "label": "", "name": "/dev/vda2", "size": "100M", "type": "partition", "uuid": "7B77-95E7" }, "/dev/vda3": { "fstype": "xfs", "label": "root", "name": "/dev/vda3", "size": "9.9G", "type": "partition", "uuid": "58dca2db-4ccd-4171-a017-dd53ad8c1c54" }, "/dev/vdb": { "fstype": "", "label": "", "name": "/dev/vdb", "size": "2G", "type": "disk", "uuid": "" }, "/dev/vdb1": { "fstype": "ext4", "label": "yumcache", "name": "/dev/vdb1", "size": "2G", "type": "partition", "uuid": "4d1ed9a1-5608-4ff8-ad89-21d95caa66fa" }, "/dev/vdc": { "fstype": "", "label": "", "name": "/dev/vdc", "size": "2G", "type": "disk", "uuid": "" }, "/dev/vdc1": { "fstype": "ext4", "label": "yumvarlib", "name": "/dev/vdc1", "size": "2G", "type": "partition", "uuid": "a7abcc97-edc5-476c-8ccd-694b4621fd28" }, "/dev/vdd": { "fstype": "", "label": "", "name": "/dev/vdd", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vde": { "fstype": "", "label": "", "name": "/dev/vde", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vdf": { "fstype": "", "label": "", "name": "/dev/vdf", "size": "10G", "type": "disk", "uuid": "" } } } } TASK [Get the size of parent/pool device] ************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:31 Thursday 21 July 2022 15:23:19 +0000 (0:00:00.091) 0:00:47.568 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "bytes": 10737418240, "changed": false, "lvm": "10g", "parted": "10GiB", "size": "10 GiB" } TASK [debug] ******************************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:36 Thursday 21 July 2022 15:23:19 +0000 (0:00:00.392) 0:00:47.960 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_pool_size": { "bytes": 10737418240, "changed": false, "failed": false, "lvm": "10g", "parted": "10GiB", "size": "10 GiB" } } TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:39 Thursday 21 July 2022 15:23:19 +0000 (0:00:00.084) 0:00:48.045 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_expected_size": "4294967296.0" }, "changed": false } TASK [debug] ******************************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:44 Thursday 21 July 2022 15:23:19 +0000 (0:00:00.052) 0:00:48.097 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_actual_size": { "bytes": 4294967296, "changed": false, "failed": false, "lvm": "4g", "parted": "4GiB", "size": "4 GiB" } } TASK [debug] ******************************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:47 Thursday 21 July 2022 15:23:19 +0000 (0:00:00.036) 0:00:48.134 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_expected_size": "4294967296.0" } TASK [assert] ****************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:50 Thursday 21 July 2022 15:23:19 +0000 (0:00:00.074) 0:00:48.208 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Get information about the LV] ******************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-cache.yml:6 Thursday 21 July 2022 15:23:19 +0000 (0:00:00.129) 0:00:48.338 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "cmd": [ "lvs", "--noheadings", "--nameprefixes", "--units=b", "--nosuffix", "--unquoted", "-o", "name,attr,cache_total_blocks,chunk_size,segtype", "foo/test2" ], "delta": "0:00:00.033923", "end": "2022-07-21 11:23:20.149649", "rc": 0, "start": "2022-07-21 11:23:20.115726" } STDOUT: LVM2_LV_NAME=test2 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-cache.yml:14 Thursday 21 July 2022 15:23:20 +0000 (0:00:00.422) 0:00:48.760 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_lv_segtype": [ "linear" ] }, "changed": false } TASK [check segment type] ****************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-cache.yml:17 Thursday 21 July 2022 15:23:20 +0000 (0:00:00.049) 0:00:48.810 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-cache.yml:22 Thursday 21 July 2022 15:23:20 +0000 (0:00:00.049) 0:00:48.859 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the requested cache size] ****************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-cache.yml:26 Thursday 21 July 2022 15:23:20 +0000 (0:00:00.037) 0:00:48.897 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-cache.yml:32 Thursday 21 July 2022 15:23:20 +0000 (0:00:00.036) 0:00:48.933 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-cache.yml:36 Thursday 21 July 2022 15:23:20 +0000 (0:00:00.045) 0:00:48.978 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume.yml:16 Thursday 21 July 2022 15:23:20 +0000 (0:00:00.037) 0:00:49.016 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [Verify the volumes with no pool were correctly managed] ****************** task path: /tmp/tmpafnwmbol/tests/verify-role-results.yml:43 Thursday 21 July 2022 15:23:20 +0000 (0:00:00.037) 0:00:49.053 ********* TASK [Clean up variable namespace] ********************************************* task path: /tmp/tmpafnwmbol/tests/verify-role-results.yml:53 Thursday 21 July 2022 15:23:20 +0000 (0:00:00.023) 0:00:49.076 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_blkinfo": null, "storage_test_crypttab": null, "storage_test_fstab": null }, "changed": false } TASK [Repeat the previous invocation to verify idempotence] ******************** task path: /tmp/tmpafnwmbol/tests/tests_lvm_percent_size.yml:81 Thursday 21 July 2022 15:23:20 +0000 (0:00:00.037) 0:00:49.114 ********* TASK [linux-system-roles.storage : set platform/version specific variables] **** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main.yml:2 Thursday 21 July 2022 15:23:20 +0000 (0:00:00.044) 0:00:49.159 ********* included: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for /cache/rhel-8-y.qcow2.snap TASK [linux-system-roles.storage : Ensure ansible_facts used by role] ********** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2 Thursday 21 July 2022 15:23:20 +0000 (0:00:00.036) 0:00:49.195 ********* ok: [/cache/rhel-8-y.qcow2.snap] TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8 Thursday 21 July 2022 15:23:21 +0000 (0:00:00.541) 0:00:49.737 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [/cache/rhel-8-y.qcow2.snap] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [/cache/rhel-8-y.qcow2.snap] => (item=RedHat_8.yml) => { "ansible_facts": { "blivet_package_list": [ "python3-blivet", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap", "vdo", "kmod-kvdo", "xfsprogs" ] }, "ansible_included_var_files": [ "/tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/vars/RedHat_8.yml" ], "ansible_loop_var": "item", "changed": false, "item": "RedHat_8.yml" } skipping: [/cache/rhel-8-y.qcow2.snap] => (item=RedHat_8.7.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat_8.7.yml", "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : define an empty list of pools to be used in testing] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main.yml:5 Thursday 21 July 2022 15:23:21 +0000 (0:00:00.063) 0:00:49.800 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : define an empty list of volumes to be used in testing] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main.yml:9 Thursday 21 July 2022 15:23:21 +0000 (0:00:00.033) 0:00:49.834 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : include the appropriate provider tasks] ***** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main.yml:13 Thursday 21 July 2022 15:23:21 +0000 (0:00:00.033) 0:00:49.868 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount included: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for /cache/rhel-8-y.qcow2.snap TASK [linux-system-roles.storage : get a list of rpm packages installed on host machine] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2 Thursday 21 July 2022 15:23:21 +0000 (0:00:00.045) 0:00:49.913 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : make sure blivet is available] ************** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:7 Thursday 21 July 2022 15:23:21 +0000 (0:00:00.020) 0:00:49.934 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [linux-system-roles.storage : show storage_pools] ************************* task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:14 Thursday 21 July 2022 15:23:22 +0000 (0:00:01.279) 0:00:51.213 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_pools": [ { "disks": [ "sda" ], "name": "foo", "volumes": [ { "mount_point": "/opt/test1", "name": "test1", "size": "60%" }, { "mount_point": "/opt/test2", "name": "test2", "size": "40%" } ] } ] } TASK [linux-system-roles.storage : show storage_volumes] *********************** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:19 Thursday 21 July 2022 15:23:22 +0000 (0:00:00.076) 0:00:51.290 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [linux-system-roles.storage : get required packages] ********************** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:24 Thursday 21 July 2022 15:23:22 +0000 (0:00:00.108) 0:00:51.399 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [ "lvm2" ], "pools": [], "volumes": [] } TASK [linux-system-roles.storage : enable copr repositories if needed] ********* task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:37 Thursday 21 July 2022 15:23:24 +0000 (0:00:01.572) 0:00:52.971 ********* included: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml for /cache/rhel-8-y.qcow2.snap TASK [linux-system-roles.storage : check if the COPR support packages should be installed] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:2 Thursday 21 July 2022 15:23:24 +0000 (0:00:00.048) 0:00:53.019 ********* TASK [linux-system-roles.storage : make sure COPR support packages are present] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:13 Thursday 21 July 2022 15:23:24 +0000 (0:00:00.039) 0:00:53.059 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : enable COPRs] ******************************* task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:18 Thursday 21 July 2022 15:23:24 +0000 (0:00:00.039) 0:00:53.099 ********* TASK [linux-system-roles.storage : make sure required packages are installed] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:44 Thursday 21 July 2022 15:23:24 +0000 (0:00:00.037) 0:00:53.136 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [linux-system-roles.storage : get service facts] ************************** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:51 Thursday 21 July 2022 15:23:25 +0000 (0:00:01.316) 0:00:54.453 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "services": { "NetworkManager-dispatcher.service": { "name": "NetworkManager-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "NetworkManager-wait-online.service": { "name": "NetworkManager-wait-online.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "NetworkManager.service": { "name": "NetworkManager.service", "source": "systemd", "state": "running", "status": "enabled" }, "arp-ethers.service": { "name": "arp-ethers.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "auditd.service": { "name": "auditd.service", "source": "systemd", "state": "running", "status": "enabled" }, "auth-rpcgss-module.service": { "name": "auth-rpcgss-module.service", "source": "systemd", "state": "stopped", "status": "static" }, "autovt@.service": { "name": "autovt@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "blivet.service": { "name": "blivet.service", "source": "systemd", "state": "inactive", "status": "static" }, "blk-availability.service": { "name": "blk-availability.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "chrony-dnssrv@.service": { "name": "chrony-dnssrv@.service", "source": "systemd", "state": "unknown", "status": "static" }, "chrony-wait.service": { "name": "chrony-wait.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd.service": { "name": "chronyd.service", "source": "systemd", "state": "running", "status": "enabled" }, "cloud-config.service": { "name": "cloud-config.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-final.service": { "name": "cloud-final.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init-hotplugd.service": { "name": "cloud-init-hotplugd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cloud-init-local.service": { "name": "cloud-init-local.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init.service": { "name": "cloud-init.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cockpit-motd.service": { "name": "cockpit-motd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cockpit-wsinstance-http.service": { "name": "cockpit-wsinstance-http.service", "source": "systemd", "state": "inactive", "status": "static" }, "cockpit-wsinstance-https-factory@.service": { "name": "cockpit-wsinstance-https-factory@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cockpit-wsinstance-https@.service": { "name": "cockpit-wsinstance-https@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cockpit.service": { "name": "cockpit.service", "source": "systemd", "state": "inactive", "status": "static" }, "console-getty.service": { "name": "console-getty.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "container-getty@.service": { "name": "container-getty@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cpupower.service": { "name": "cpupower.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "crond.service": { "name": "crond.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-org.freedesktop.hostname1.service": { "name": "dbus-org.freedesktop.hostname1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.locale1.service": { "name": "dbus-org.freedesktop.locale1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.login1.service": { "name": "dbus-org.freedesktop.login1.service", "source": "systemd", "state": "active", "status": "static" }, "dbus-org.freedesktop.nm-dispatcher.service": { "name": "dbus-org.freedesktop.nm-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "dbus-org.freedesktop.portable1.service": { "name": "dbus-org.freedesktop.portable1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.timedate1.service": { "name": "dbus-org.freedesktop.timedate1.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "dbus.service": { "name": "dbus.service", "source": "systemd", "state": "running", "status": "static" }, "dbxtool.service": { "name": "dbxtool.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "debug-shell.service": { "name": "debug-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "display-manager.service": { "name": "display-manager.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "dm-event.service": { "name": "dm-event.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-makecache.service": { "name": "dnf-makecache.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-cmdline.service": { "name": "dracut-cmdline.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-initqueue.service": { "name": "dracut-initqueue.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-mount.service": { "name": "dracut-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-mount.service": { "name": "dracut-pre-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-pivot.service": { "name": "dracut-pre-pivot.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-trigger.service": { "name": "dracut-pre-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-udev.service": { "name": "dracut-pre-udev.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown.service": { "name": "dracut-shutdown.service", "source": "systemd", "state": "stopped", "status": "static" }, "emergency.service": { "name": "emergency.service", "source": "systemd", "state": "stopped", "status": "static" }, "fcoe.service": { "name": "fcoe.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "fstrim.service": { "name": "fstrim.service", "source": "systemd", "state": "inactive", "status": "static" }, "getty@.service": { "name": "getty@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "getty@tty1.service": { "name": "getty@tty1.service", "source": "systemd", "state": "running", "status": "active" }, "grub-boot-indeterminate.service": { "name": "grub-boot-indeterminate.service", "source": "systemd", "state": "inactive", "status": "static" }, "gssproxy.service": { "name": "gssproxy.service", "source": "systemd", "state": "running", "status": "disabled" }, "halt-local.service": { "name": "halt-local.service", "source": "systemd", "state": "inactive", "status": "static" }, "hv_kvp_daemon.service": { "name": "hv_kvp_daemon.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "import-state.service": { "name": "import-state.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "initrd-cleanup.service": { "name": "initrd-cleanup.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-parse-etc.service": { "name": "initrd-parse-etc.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-switch-root.service": { "name": "initrd-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-udevadm-cleanup-db.service": { "name": "initrd-udevadm-cleanup-db.service", "source": "systemd", "state": "stopped", "status": "static" }, "insights-client-boot.service": { "name": "insights-client-boot.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "insights-client-results.service": { "name": "insights-client-results.service", "source": "systemd", "state": "inactive", "status": "static" }, "insights-client.service": { "name": "insights-client.service", "source": "systemd", "state": "inactive", "status": "static" }, "irqbalance.service": { "name": "irqbalance.service", "source": "systemd", "state": "running", "status": "enabled" }, "iscsi-shutdown.service": { "name": "iscsi-shutdown.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "iscsi.service": { "name": "iscsi.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "iscsid.service": { "name": "iscsid.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "kdump.service": { "name": "kdump.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "kmod-static-nodes.service": { "name": "kmod-static-nodes.service", "source": "systemd", "state": "stopped", "status": "static" }, "kvm_stat.service": { "name": "kvm_stat.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "ldconfig.service": { "name": "ldconfig.service", "source": "systemd", "state": "stopped", "status": "static" }, "loadmodules.service": { "name": "loadmodules.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "lvm2-activation-early.service": { "name": "lvm2-activation-early.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "lvm2-activation.service": { "name": "lvm2-activation.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "lvm2-lvmpolld.service": { "name": "lvm2-lvmpolld.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-monitor.service": { "name": "lvm2-monitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "lvm2-pvscan@.service": { "name": "lvm2-pvscan@.service", "source": "systemd", "state": "unknown", "status": "static" }, "lvm2-pvscan@8:0.service": { "name": "lvm2-pvscan@8:0.service", "source": "systemd", "state": "stopped", "status": "active" }, "man-db-cache-update.service": { "name": "man-db-cache-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "man-db-restart-cache-update.service": { "name": "man-db-restart-cache-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "mdadm-grow-continue@.service": { "name": "mdadm-grow-continue@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdadm-last-resort@.service": { "name": "mdadm-last-resort@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdcheck_continue.service": { "name": "mdcheck_continue.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdcheck_start.service": { "name": "mdcheck_start.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmon@.service": { "name": "mdmon@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdmonitor-oneshot.service": { "name": "mdmonitor-oneshot.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmonitor.service": { "name": "mdmonitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "messagebus.service": { "name": "messagebus.service", "source": "systemd", "state": "active", "status": "static" }, "microcode.service": { "name": "microcode.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "multipathd.service": { "name": "multipathd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "ndctl-monitor.service": { "name": "ndctl-monitor.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "network.service": { "name": "network.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "nfs-blkmap.service": { "name": "nfs-blkmap.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-convert.service": { "name": "nfs-convert.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-idmapd.service": { "name": "nfs-idmapd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-mountd.service": { "name": "nfs-mountd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-server.service": { "name": "nfs-server.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-utils.service": { "name": "nfs-utils.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfsdcld.service": { "name": "nfsdcld.service", "source": "systemd", "state": "stopped", "status": "static" }, "nis-domainname.service": { "name": "nis-domainname.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "ntpd.service": { "name": "ntpd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ntpdate.service": { "name": "ntpdate.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "oddjobd.service": { "name": "oddjobd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "packagekit-offline-update.service": { "name": "packagekit-offline-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "packagekit.service": { "name": "packagekit.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-quit-wait.service": { "name": "plymouth-quit-wait.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "plymouth-start.service": { "name": "plymouth-start.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "polkit.service": { "name": "polkit.service", "source": "systemd", "state": "running", "status": "static" }, "qemu-guest-agent.service": { "name": "qemu-guest-agent.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "quotaon.service": { "name": "quotaon.service", "source": "systemd", "state": "inactive", "status": "static" }, "rbdmap.service": { "name": "rbdmap.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "rc-local.service": { "name": "rc-local.service", "source": "systemd", "state": "stopped", "status": "static" }, "rdisc.service": { "name": "rdisc.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rescue.service": { "name": "rescue.service", "source": "systemd", "state": "stopped", "status": "static" }, "rhcd.service": { "name": "rhcd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhsm-facts.service": { "name": "rhsm-facts.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhsm.service": { "name": "rhsm.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhsmcertd.service": { "name": "rhsmcertd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpc-gssd.service": { "name": "rpc-gssd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd-notify.service": { "name": "rpc-statd-notify.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd.service": { "name": "rpc-statd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpcbind.service": { "name": "rpcbind.service", "source": "systemd", "state": "running", "status": "enabled" }, "rsyslog.service": { "name": "rsyslog.service", "source": "systemd", "state": "running", "status": "enabled" }, "selinux-autorelabel-mark.service": { "name": "selinux-autorelabel-mark.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "selinux-autorelabel.service": { "name": "selinux-autorelabel.service", "source": "systemd", "state": "inactive", "status": "static" }, "serial-getty@.service": { "name": "serial-getty@.service", "source": "systemd", "state": "unknown", "status": "indirect" }, "serial-getty@ttyS0.service": { "name": "serial-getty@ttyS0.service", "source": "systemd", "state": "running", "status": "active" }, "snapd.seeded.service": { "name": "snapd.seeded.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sntp.service": { "name": "sntp.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sshd-keygen.service": { "name": "sshd-keygen.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sshd-keygen@.service": { "name": "sshd-keygen@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@ecdsa.service": { "name": "sshd-keygen@ecdsa.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd-keygen@ed25519.service": { "name": "sshd-keygen@ed25519.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd-keygen@rsa.service": { "name": "sshd-keygen@rsa.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd.service": { "name": "sshd.service", "source": "systemd", "state": "running", "status": "enabled" }, "sshd@.service": { "name": "sshd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "sssd-autofs.service": { "name": "sssd-autofs.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-kcm.service": { "name": "sssd-kcm.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "sssd-nss.service": { "name": "sssd-nss.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pac.service": { "name": "sssd-pac.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pam.service": { "name": "sssd-pam.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-ssh.service": { "name": "sssd-ssh.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-sudo.service": { "name": "sssd-sudo.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd.service": { "name": "sssd.service", "source": "systemd", "state": "running", "status": "enabled" }, "syslog.service": { "name": "syslog.service", "source": "systemd", "state": "active", "status": "enabled" }, "system-update-cleanup.service": { "name": "system-update-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-ask-password-console.service": { "name": "systemd-ask-password-console.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-wall.service": { "name": "systemd-ask-password-wall.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-backlight@.service": { "name": "systemd-backlight@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-binfmt.service": { "name": "systemd-binfmt.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-coredump@.service": { "name": "systemd-coredump@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-exit.service": { "name": "systemd-exit.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-firstboot.service": { "name": "systemd-firstboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck-root.service": { "name": "systemd-fsck-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck@.service": { "name": "systemd-fsck@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-fsck@dev-disk-by\\x2duuid-7B77\\x2d95E7.service": { "name": "systemd-fsck@dev-disk-by\\x2duuid-7B77\\x2d95E7.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-fsck@dev-vdb1.service": { "name": "systemd-fsck@dev-vdb1.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-fsck@dev-vdc1.service": { "name": "systemd-fsck@dev-vdc1.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-halt.service": { "name": "systemd-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hibernate-resume@.service": { "name": "systemd-hibernate-resume@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-hibernate.service": { "name": "systemd-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hostnamed.service": { "name": "systemd-hostnamed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hwdb-update.service": { "name": "systemd-hwdb-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-hybrid-sleep.service": { "name": "systemd-hybrid-sleep.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-initctl.service": { "name": "systemd-initctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-catalog-update.service": { "name": "systemd-journal-catalog-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-flush.service": { "name": "systemd-journal-flush.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journald.service": { "name": "systemd-journald.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-kexec.service": { "name": "systemd-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-localed.service": { "name": "systemd-localed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-logind.service": { "name": "systemd-logind.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-machine-id-commit.service": { "name": "systemd-machine-id-commit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-modules-load.service": { "name": "systemd-modules-load.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-networkd-wait-online.service": { "name": "systemd-networkd-wait-online.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "systemd-portabled.service": { "name": "systemd-portabled.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-poweroff.service": { "name": "systemd-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-quotacheck.service": { "name": "systemd-quotacheck.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-random-seed.service": { "name": "systemd-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-reboot.service": { "name": "systemd-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-remount-fs.service": { "name": "systemd-remount-fs.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-resolved.service": { "name": "systemd-resolved.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-rfkill.service": { "name": "systemd-rfkill.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend-then-hibernate.service": { "name": "systemd-suspend-then-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend.service": { "name": "systemd-suspend.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-sysctl.service": { "name": "systemd-sysctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-sysusers.service": { "name": "systemd-sysusers.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-timedated.service": { "name": "systemd-timedated.service", "source": "systemd", "state": "inactive", "status": "masked" }, "systemd-timesyncd.service": { "name": "systemd-timesyncd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "systemd-tmpfiles-clean.service": { "name": "systemd-tmpfiles-clean.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev.service": { "name": "systemd-tmpfiles-setup-dev.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup.service": { "name": "systemd-tmpfiles-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-settle.service": { "name": "systemd-udev-settle.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-trigger.service": { "name": "systemd-udev-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udevd.service": { "name": "systemd-udevd.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-update-done.service": { "name": "systemd-update-done.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp-runlevel.service": { "name": "systemd-update-utmp-runlevel.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp.service": { "name": "systemd-update-utmp.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-user-sessions.service": { "name": "systemd-user-sessions.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-vconsole-setup.service": { "name": "systemd-vconsole-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-volatile-root.service": { "name": "systemd-volatile-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "tcsd.service": { "name": "tcsd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "teamd@.service": { "name": "teamd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "timedatex.service": { "name": "timedatex.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "tuned.service": { "name": "tuned.service", "source": "systemd", "state": "running", "status": "enabled" }, "unbound-anchor.service": { "name": "unbound-anchor.service", "source": "systemd", "state": "stopped", "status": "static" }, "user-runtime-dir@.service": { "name": "user-runtime-dir@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user-runtime-dir@0.service": { "name": "user-runtime-dir@0.service", "source": "systemd", "state": "stopped", "status": "active" }, "user@.service": { "name": "user@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user@0.service": { "name": "user@0.service", "source": "systemd", "state": "running", "status": "active" }, "vdo-start-by-dev@.service": { "name": "vdo-start-by-dev@.service", "source": "systemd", "state": "unknown", "status": "static" }, "vdo.service": { "name": "vdo.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "ypbind.service": { "name": "ypbind.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "yppasswdd.service": { "name": "yppasswdd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ypserv.service": { "name": "ypserv.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ypxfrd.service": { "name": "ypxfrd.service", "source": "systemd", "state": "stopped", "status": "not-found" } } }, "changed": false } TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************ task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:58 Thursday 21 July 2022 15:23:27 +0000 (0:00:01.716) 0:00:56.169 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] ******* task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:71 Thursday 21 July 2022 15:23:27 +0000 (0:00:00.057) 0:00:56.227 ********* TASK [linux-system-roles.storage : manage the pools and volumes to match the specified state] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77 Thursday 21 July 2022 15:23:27 +0000 (0:00:00.021) 0:00:56.248 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "actions": [], "changed": false, "crypts": [], "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/vda2", "/dev/vda3", "/dev/mapper/foo-test1", "/dev/mapper/foo-test2", "/dev/sdb", "/dev/sdc", "/dev/nvme0n1", "/dev/nvme1n1", "/dev/nvme2n1", "/dev/vdb1", "/dev/vdc1", "/dev/vdd", "/dev/vde", "/dev/vdf" ], "mounts": [ { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" }, { "dump": 0, "fstype": "ext4", "opts": "defaults", "passno": 0, "path": "/opt/test2", "src": "/dev/mapper/foo-test2", "state": "mounted" } ], "packages": [ "dosfstools", "e2fsprogs", "lvm2", "xfsprogs" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "60%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test2", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test2", "_raw_device": "/dev/mapper/foo-test2", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "40%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } TASK [linux-system-roles.storage : Workaround for udev issue on some platforms] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:91 Thursday 21 July 2022 15:23:29 +0000 (0:00:01.748) 0:00:57.997 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] ***** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:103 Thursday 21 July 2022 15:23:29 +0000 (0:00:00.036) 0:00:58.034 ********* TASK [linux-system-roles.storage : show blivet_output] ************************* task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:109 Thursday 21 July 2022 15:23:29 +0000 (0:00:00.022) 0:00:58.056 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "blivet_output": { "actions": [], "changed": false, "crypts": [], "failed": false, "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/vda2", "/dev/vda3", "/dev/mapper/foo-test1", "/dev/mapper/foo-test2", "/dev/sdb", "/dev/sdc", "/dev/nvme0n1", "/dev/nvme1n1", "/dev/nvme2n1", "/dev/vdb1", "/dev/vdc1", "/dev/vdd", "/dev/vde", "/dev/vdf" ], "mounts": [ { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" }, { "dump": 0, "fstype": "ext4", "opts": "defaults", "passno": 0, "path": "/opt/test2", "src": "/dev/mapper/foo-test2", "state": "mounted" } ], "packages": [ "dosfstools", "e2fsprogs", "lvm2", "xfsprogs" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "60%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test2", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test2", "_raw_device": "/dev/mapper/foo-test2", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "40%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } } TASK [linux-system-roles.storage : set the list of pools for test verification] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:114 Thursday 21 July 2022 15:23:29 +0000 (0:00:00.040) 0:00:58.097 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "60%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test2", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test2", "_raw_device": "/dev/mapper/foo-test2", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "40%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] }, "changed": false } TASK [linux-system-roles.storage : set the list of volumes for test verification] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:118 Thursday 21 July 2022 15:23:29 +0000 (0:00:00.040) 0:00:58.138 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : remove obsolete mounts] ********************* task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:134 Thursday 21 July 2022 15:23:29 +0000 (0:00:00.040) 0:00:58.178 ********* TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:146 Thursday 21 July 2022 15:23:29 +0000 (0:00:00.037) 0:00:58.216 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : set up new/current mounts] ****************** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:151 Thursday 21 July 2022 15:23:30 +0000 (0:00:00.623) 0:00:58.839 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount ok: [/cache/rhel-8-y.qcow2.snap] => (item={'src': '/dev/mapper/foo-test1', 'path': '/opt/test1', 'fstype': 'xfs', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted'}) => { "ansible_loop_var": "mount_info", "backup_file": "", "boot": "yes", "changed": false, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" }, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test1" } redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount ok: [/cache/rhel-8-y.qcow2.snap] => (item={'src': '/dev/mapper/foo-test2', 'path': '/opt/test2', 'fstype': 'ext4', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted'}) => { "ansible_loop_var": "mount_info", "backup_file": "", "boot": "yes", "changed": false, "dump": "0", "fstab": "/etc/fstab", "fstype": "ext4", "mount_info": { "dump": 0, "fstype": "ext4", "opts": "defaults", "passno": 0, "path": "/opt/test2", "src": "/dev/mapper/foo-test2", "state": "mounted" }, "name": "/opt/test2", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test2" } TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:163 Thursday 21 July 2022 15:23:31 +0000 (0:00:00.806) 0:00:59.646 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : retrieve facts for the /etc/crypttab file] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:171 Thursday 21 July 2022 15:23:31 +0000 (0:00:00.665) 0:01:00.311 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "stat": { "exists": false } } TASK [linux-system-roles.storage : manage /etc/crypttab to account for changes we just made] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:176 Thursday 21 July 2022 15:23:32 +0000 (0:00:00.378) 0:01:00.689 ********* TASK [linux-system-roles.storage : Update facts] ******************************* task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:198 Thursday 21 July 2022 15:23:32 +0000 (0:00:00.025) 0:01:00.715 ********* ok: [/cache/rhel-8-y.qcow2.snap] META: role_complete for /cache/rhel-8-y.qcow2.snap TASK [include_tasks] *********************************************************** task path: /tmp/tmpafnwmbol/tests/tests_lvm_percent_size.yml:95 Thursday 21 July 2022 15:23:33 +0000 (0:00:00.995) 0:01:01.711 ********* included: /tmp/tmpafnwmbol/tests/verify-role-results.yml for /cache/rhel-8-y.qcow2.snap TASK [Print out pool information] ********************************************** task path: /tmp/tmpafnwmbol/tests/verify-role-results.yml:1 Thursday 21 July 2022 15:23:33 +0000 (0:00:00.040) 0:01:01.751 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "60%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test2", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test2", "_raw_device": "/dev/mapper/foo-test2", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "40%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] } TASK [Print out volume information] ******************************************** task path: /tmp/tmpafnwmbol/tests/verify-role-results.yml:6 Thursday 21 July 2022 15:23:33 +0000 (0:00:00.054) 0:01:01.805 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => {} TASK [Collect info about the volumes.] ***************************************** task path: /tmp/tmpafnwmbol/tests/verify-role-results.yml:14 Thursday 21 July 2022 15:23:33 +0000 (0:00:00.040) 0:01:01.846 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "info": { "/dev/mapper/foo-test1": { "fstype": "xfs", "label": "", "name": "/dev/mapper/foo-test1", "size": "6G", "type": "lvm", "uuid": "3db40ee5-790d-4909-8628-933255bd71de" }, "/dev/mapper/foo-test2": { "fstype": "ext4", "label": "", "name": "/dev/mapper/foo-test2", "size": "4G", "type": "lvm", "uuid": "ee7d1310-5176-4ec6-80b8-28181ffffed3" }, "/dev/nvme0n1": { "fstype": "", "label": "", "name": "/dev/nvme0n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme1n1": { "fstype": "", "label": "", "name": "/dev/nvme1n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme2n1": { "fstype": "", "label": "", "name": "/dev/nvme2n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sda": { "fstype": "LVM2_member", "label": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "irzZ3c-YIeb-eoxj-fRdj-JSvm-ls7D-iZNahz" }, "/dev/sdb": { "fstype": "", "label": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc": { "fstype": "", "label": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sr0": { "fstype": "iso9660", "label": "cidata", "name": "/dev/sr0", "size": "364K", "type": "rom", "uuid": "2022-07-21-15-22-18-00" }, "/dev/vda": { "fstype": "", "label": "", "name": "/dev/vda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vda1": { "fstype": "", "label": "", "name": "/dev/vda1", "size": "1M", "type": "partition", "uuid": "" }, "/dev/vda2": { "fstype": "vfat", "label": "", "name": "/dev/vda2", "size": "100M", "type": "partition", "uuid": "7B77-95E7" }, "/dev/vda3": { "fstype": "xfs", "label": "root", "name": "/dev/vda3", "size": "9.9G", "type": "partition", "uuid": "58dca2db-4ccd-4171-a017-dd53ad8c1c54" }, "/dev/vdb": { "fstype": "", "label": "", "name": "/dev/vdb", "size": "2G", "type": "disk", "uuid": "" }, "/dev/vdb1": { "fstype": "ext4", "label": "yumcache", "name": "/dev/vdb1", "size": "2G", "type": "partition", "uuid": "4d1ed9a1-5608-4ff8-ad89-21d95caa66fa" }, "/dev/vdc": { "fstype": "", "label": "", "name": "/dev/vdc", "size": "2G", "type": "disk", "uuid": "" }, "/dev/vdc1": { "fstype": "ext4", "label": "yumvarlib", "name": "/dev/vdc1", "size": "2G", "type": "partition", "uuid": "a7abcc97-edc5-476c-8ccd-694b4621fd28" }, "/dev/vdd": { "fstype": "", "label": "", "name": "/dev/vdd", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vde": { "fstype": "", "label": "", "name": "/dev/vde", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vdf": { "fstype": "", "label": "", "name": "/dev/vdf", "size": "10G", "type": "disk", "uuid": "" } } } TASK [Read the /etc/fstab file for volume existence] *************************** task path: /tmp/tmpafnwmbol/tests/verify-role-results.yml:19 Thursday 21 July 2022 15:23:33 +0000 (0:00:00.382) 0:01:02.228 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "cmd": [ "cat", "/etc/fstab" ], "delta": "0:00:00.003126", "end": "2022-07-21 11:23:34.002271", "rc": 0, "start": "2022-07-21 11:23:33.999145" } STDOUT: UUID=58dca2db-4ccd-4171-a017-dd53ad8c1c54 / xfs defaults 0 0 UUID=7B77-95E7 /boot/efi vfat defaults,uid=0,gid=0,umask=077,shortname=winnt 0 2 /dev/vdb1 /var/cache/dnf auto defaults,nofail,x-systemd.requires=cloud-init.service,_netdev,comment=cloudconfig 0 2 /dev/vdc1 /var/lib/dnf auto defaults,nofail,x-systemd.requires=cloud-init.service,_netdev,comment=cloudconfig 0 2 /dev/mapper/foo-test1 /opt/test1 xfs defaults 0 0 /dev/mapper/foo-test2 /opt/test2 ext4 defaults 0 0 TASK [Read the /etc/crypttab file] ********************************************* task path: /tmp/tmpafnwmbol/tests/verify-role-results.yml:24 Thursday 21 July 2022 15:23:34 +0000 (0:00:00.381) 0:01:02.610 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "cmd": [ "cat", "/etc/crypttab" ], "delta": "0:00:00.003154", "end": "2022-07-21 11:23:34.381110", "failed_when_result": false, "rc": 1, "start": "2022-07-21 11:23:34.377956" } STDERR: cat: /etc/crypttab: No such file or directory MSG: non-zero return code TASK [Verify the volumes listed in storage_pools were correctly managed] ******* task path: /tmp/tmpafnwmbol/tests/verify-role-results.yml:33 Thursday 21 July 2022 15:23:34 +0000 (0:00:00.380) 0:01:02.990 ********* included: /tmp/tmpafnwmbol/tests/test-verify-pool.yml for /cache/rhel-8-y.qcow2.snap => (item={'disks': ['sda'], 'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'name': 'foo', 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_metadata_version': None, 'state': 'present', 'type': 'lvm', 'volumes': [{'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'name': 'test1', 'raid_level': None, 'size': '60%', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test1', '_raw_device': '/dev/mapper/foo-test1', '_mount_id': '/dev/mapper/foo-test1', '_kernel_device': '/dev/dm-1', '_raw_kernel_device': '/dev/dm-1'}, {'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'ext4', 'mount_options': 'defaults', 'mount_point': '/opt/test2', 'name': 'test2', 'raid_level': None, 'size': '40%', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test2', '_raw_device': '/dev/mapper/foo-test2', '_mount_id': '/dev/mapper/foo-test2', '_kernel_device': '/dev/dm-0', '_raw_kernel_device': '/dev/dm-0'}], 'raid_chunk_size': None}) TASK [Set _storage_pool_tests] ************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-pool.yml:5 Thursday 21 July 2022 15:23:34 +0000 (0:00:00.062) 0:01:03.053 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_pool_tests": [ "members", "volumes" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool.yml:18 Thursday 21 July 2022 15:23:34 +0000 (0:00:00.038) 0:01:03.092 ********* included: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml for /cache/rhel-8-y.qcow2.snap => (item=members) included: /tmp/tmpafnwmbol/tests/test-verify-pool-volumes.yml for /cache/rhel-8-y.qcow2.snap => (item=volumes) TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:1 Thursday 21 July 2022 15:23:34 +0000 (0:00:00.045) 0:01:03.137 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_pv_count": "1", "_storage_test_pool_pvs_lvm": [ "/dev/sda" ] }, "changed": false } TASK [Get the canonical device path for each member device] ******************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:10 Thursday 21 July 2022 15:23:34 +0000 (0:00:00.114) 0:01:03.251 ********* ok: [/cache/rhel-8-y.qcow2.snap] => (item=/dev/sda) => { "ansible_loop_var": "pv", "changed": false, "device": "/dev/sda", "pv": "/dev/sda" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:19 Thursday 21 July 2022 15:23:35 +0000 (0:00:00.363) 0:01:03.615 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "__pvs_lvm_len": "1" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:23 Thursday 21 July 2022 15:23:35 +0000 (0:00:00.047) 0:01:03.663 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_pool_pvs": [ "/dev/sda" ] }, "changed": false } TASK [Verify PV count] ********************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:27 Thursday 21 July 2022 15:23:35 +0000 (0:00:00.048) 0:01:03.711 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:34 Thursday 21 July 2022 15:23:35 +0000 (0:00:00.048) 0:01:03.760 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:38 Thursday 21 July 2022 15:23:35 +0000 (0:00:00.041) 0:01:03.801 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:42 Thursday 21 July 2022 15:23:35 +0000 (0:00:00.049) 0:01:03.851 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check the type of each PV] *********************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:46 Thursday 21 July 2022 15:23:35 +0000 (0:00:00.025) 0:01:03.876 ********* ok: [/cache/rhel-8-y.qcow2.snap] => (item=/dev/sda) => { "ansible_loop_var": "pv", "changed": false, "pv": "/dev/sda" } MSG: All assertions passed TASK [Check MD RAID] *********************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:56 Thursday 21 July 2022 15:23:35 +0000 (0:00:00.045) 0:01:03.922 ********* included: /tmp/tmpafnwmbol/tests/verify-pool-md.yml for /cache/rhel-8-y.qcow2.snap TASK [get information about RAID] ********************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-md.yml:6 Thursday 21 July 2022 15:23:35 +0000 (0:00:00.042) 0:01:03.965 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-md.yml:12 Thursday 21 July 2022 15:23:35 +0000 (0:00:00.024) 0:01:03.989 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-md.yml:16 Thursday 21 July 2022 15:23:35 +0000 (0:00:00.024) 0:01:04.014 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-md.yml:20 Thursday 21 July 2022 15:23:35 +0000 (0:00:00.026) 0:01:04.040 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-md.yml:24 Thursday 21 July 2022 15:23:35 +0000 (0:00:00.027) 0:01:04.068 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-md.yml:30 Thursday 21 July 2022 15:23:35 +0000 (0:00:00.026) 0:01:04.094 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmpafnwmbol/tests/verify-pool-md.yml:36 Thursday 21 July 2022 15:23:35 +0000 (0:00:00.030) 0:01:04.124 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-md.yml:44 Thursday 21 July 2022 15:23:35 +0000 (0:00:00.029) 0:01:04.154 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_md_active_devices_re": null, "storage_test_md_metadata_version_re": null, "storage_test_md_spare_devices_re": null }, "changed": false } TASK [Check LVM RAID] ********************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:59 Thursday 21 July 2022 15:23:35 +0000 (0:00:00.040) 0:01:04.194 ********* included: /tmp/tmpafnwmbol/tests/verify-pool-members-lvmraid.yml for /cache/rhel-8-y.qcow2.snap TASK [Validate pool member LVM RAID settings] ********************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-members-lvmraid.yml:1 Thursday 21 July 2022 15:23:35 +0000 (0:00:00.049) 0:01:04.244 ********* included: /tmp/tmpafnwmbol/tests/verify-pool-member-lvmraid.yml for /cache/rhel-8-y.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'name': 'test1', 'raid_level': None, 'size': '60%', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test1', '_raw_device': '/dev/mapper/foo-test1', '_mount_id': '/dev/mapper/foo-test1', '_kernel_device': '/dev/dm-1', '_raw_kernel_device': '/dev/dm-1'}) included: /tmp/tmpafnwmbol/tests/verify-pool-member-lvmraid.yml for /cache/rhel-8-y.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'ext4', 'mount_options': 'defaults', 'mount_point': '/opt/test2', 'name': 'test2', 'raid_level': None, 'size': '40%', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test2', '_raw_device': '/dev/mapper/foo-test2', '_mount_id': '/dev/mapper/foo-test2', '_kernel_device': '/dev/dm-0', '_raw_kernel_device': '/dev/dm-0'}) TASK [Get information about LVM RAID] ****************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-lvmraid.yml:3 Thursday 21 July 2022 15:23:35 +0000 (0:00:00.054) 0:01:04.299 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is LVM RAID] ******************************************* task path: /tmp/tmpafnwmbol/tests/verify-pool-member-lvmraid.yml:8 Thursday 21 July 2022 15:23:35 +0000 (0:00:00.030) 0:01:04.330 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-lvmraid.yml:12 Thursday 21 July 2022 15:23:35 +0000 (0:00:00.027) 0:01:04.357 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about LVM RAID] ****************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-lvmraid.yml:3 Thursday 21 July 2022 15:23:35 +0000 (0:00:00.027) 0:01:04.385 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is LVM RAID] ******************************************* task path: /tmp/tmpafnwmbol/tests/verify-pool-member-lvmraid.yml:8 Thursday 21 July 2022 15:23:35 +0000 (0:00:00.026) 0:01:04.411 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-lvmraid.yml:12 Thursday 21 July 2022 15:23:35 +0000 (0:00:00.026) 0:01:04.438 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check Thin Pools] ******************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:62 Thursday 21 July 2022 15:23:35 +0000 (0:00:00.029) 0:01:04.468 ********* included: /tmp/tmpafnwmbol/tests/verify-pool-members-thin.yml for /cache/rhel-8-y.qcow2.snap TASK [Validate pool member thinpool settings] ********************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-members-thin.yml:1 Thursday 21 July 2022 15:23:35 +0000 (0:00:00.045) 0:01:04.513 ********* included: /tmp/tmpafnwmbol/tests/verify-pool-member-thin.yml for /cache/rhel-8-y.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'name': 'test1', 'raid_level': None, 'size': '60%', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test1', '_raw_device': '/dev/mapper/foo-test1', '_mount_id': '/dev/mapper/foo-test1', '_kernel_device': '/dev/dm-1', '_raw_kernel_device': '/dev/dm-1'}) included: /tmp/tmpafnwmbol/tests/verify-pool-member-thin.yml for /cache/rhel-8-y.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'ext4', 'mount_options': 'defaults', 'mount_point': '/opt/test2', 'name': 'test2', 'raid_level': None, 'size': '40%', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test2', '_raw_device': '/dev/mapper/foo-test2', '_mount_id': '/dev/mapper/foo-test2', '_kernel_device': '/dev/dm-0', '_raw_kernel_device': '/dev/dm-0'}) TASK [Get information about thinpool] ****************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-thin.yml:3 Thursday 21 July 2022 15:23:36 +0000 (0:00:00.055) 0:01:04.569 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-thin.yml:8 Thursday 21 July 2022 15:23:36 +0000 (0:00:00.026) 0:01:04.595 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-thin.yml:13 Thursday 21 July 2022 15:23:36 +0000 (0:00:00.026) 0:01:04.621 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-thin.yml:17 Thursday 21 July 2022 15:23:36 +0000 (0:00:00.026) 0:01:04.647 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about thinpool] ****************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-thin.yml:3 Thursday 21 July 2022 15:23:36 +0000 (0:00:00.072) 0:01:04.720 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-thin.yml:8 Thursday 21 July 2022 15:23:36 +0000 (0:00:00.026) 0:01:04.746 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-thin.yml:13 Thursday 21 July 2022 15:23:36 +0000 (0:00:00.024) 0:01:04.771 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-thin.yml:17 Thursday 21 July 2022 15:23:36 +0000 (0:00:00.023) 0:01:04.794 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check member encryption] ************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:65 Thursday 21 July 2022 15:23:36 +0000 (0:00:00.025) 0:01:04.819 ********* included: /tmp/tmpafnwmbol/tests/verify-pool-members-encryption.yml for /cache/rhel-8-y.qcow2.snap TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-members-encryption.yml:4 Thursday 21 July 2022 15:23:36 +0000 (0:00:00.045) 0:01:04.865 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Validate pool member LUKS settings] ************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-members-encryption.yml:8 Thursday 21 July 2022 15:23:36 +0000 (0:00:00.050) 0:01:04.915 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => (item=/dev/sda) => { "_storage_test_pool_member_path": "/dev/sda", "ansible_loop_var": "_storage_test_pool_member_path", "changed": false, "skip_reason": "Conditional result was False" } TASK [Validate pool member crypttab entries] *********************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-members-encryption.yml:15 Thursday 21 July 2022 15:23:36 +0000 (0:00:00.027) 0:01:04.943 ********* included: /tmp/tmpafnwmbol/tests/verify-pool-member-crypttab.yml for /cache/rhel-8-y.qcow2.snap => (item=/dev/sda) TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-crypttab.yml:1 Thursday 21 July 2022 15:23:36 +0000 (0:00:00.044) 0:01:04.987 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": [] }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpafnwmbol/tests/verify-pool-member-crypttab.yml:6 Thursday 21 July 2022 15:23:36 +0000 (0:00:00.048) 0:01:05.036 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpafnwmbol/tests/verify-pool-member-crypttab.yml:11 Thursday 21 July 2022 15:23:36 +0000 (0:00:00.046) 0:01:05.083 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-crypttab.yml:17 Thursday 21 July 2022 15:23:36 +0000 (0:00:00.035) 0:01:05.118 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-crypttab.yml:23 Thursday 21 July 2022 15:23:36 +0000 (0:00:00.036) 0:01:05.155 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-crypttab.yml:29 Thursday 21 July 2022 15:23:36 +0000 (0:00:00.041) 0:01:05.196 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-members-encryption.yml:22 Thursday 21 July 2022 15:23:36 +0000 (0:00:00.036) 0:01:05.233 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_crypttab_key_file": null }, "changed": false } TASK [Check VDO] *************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:68 Thursday 21 July 2022 15:23:36 +0000 (0:00:00.039) 0:01:05.272 ********* included: /tmp/tmpafnwmbol/tests/verify-pool-members-vdo.yml for /cache/rhel-8-y.qcow2.snap TASK [Validate pool member VDO settings] *************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-members-vdo.yml:1 Thursday 21 July 2022 15:23:36 +0000 (0:00:00.051) 0:01:05.324 ********* included: /tmp/tmpafnwmbol/tests/verify-pool-member-vdo.yml for /cache/rhel-8-y.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'name': 'test1', 'raid_level': None, 'size': '60%', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test1', '_raw_device': '/dev/mapper/foo-test1', '_mount_id': '/dev/mapper/foo-test1', '_kernel_device': '/dev/dm-1', '_raw_kernel_device': '/dev/dm-1'}) included: /tmp/tmpafnwmbol/tests/verify-pool-member-vdo.yml for /cache/rhel-8-y.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'ext4', 'mount_options': 'defaults', 'mount_point': '/opt/test2', 'name': 'test2', 'raid_level': None, 'size': '40%', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test2', '_raw_device': '/dev/mapper/foo-test2', '_mount_id': '/dev/mapper/foo-test2', '_kernel_device': '/dev/dm-0', '_raw_kernel_device': '/dev/dm-0'}) TASK [get information about VDO deduplication] ********************************* task path: /tmp/tmpafnwmbol/tests/verify-pool-member-vdo.yml:3 Thursday 21 July 2022 15:23:36 +0000 (0:00:00.053) 0:01:05.377 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-vdo.yml:8 Thursday 21 July 2022 15:23:36 +0000 (0:00:00.024) 0:01:05.401 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-vdo.yml:11 Thursday 21 July 2022 15:23:36 +0000 (0:00:00.023) 0:01:05.424 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-vdo.yml:16 Thursday 21 July 2022 15:23:36 +0000 (0:00:00.026) 0:01:05.450 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-vdo.yml:21 Thursday 21 July 2022 15:23:36 +0000 (0:00:00.023) 0:01:05.474 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-vdo.yml:24 Thursday 21 July 2022 15:23:36 +0000 (0:00:00.023) 0:01:05.498 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-vdo.yml:29 Thursday 21 July 2022 15:23:37 +0000 (0:00:00.023) 0:01:05.522 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-vdo.yml:39 Thursday 21 July 2022 15:23:37 +0000 (0:00:00.023) 0:01:05.545 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [get information about VDO deduplication] ********************************* task path: /tmp/tmpafnwmbol/tests/verify-pool-member-vdo.yml:3 Thursday 21 July 2022 15:23:37 +0000 (0:00:00.035) 0:01:05.580 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-vdo.yml:8 Thursday 21 July 2022 15:23:37 +0000 (0:00:00.026) 0:01:05.607 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-vdo.yml:11 Thursday 21 July 2022 15:23:37 +0000 (0:00:00.024) 0:01:05.631 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-vdo.yml:16 Thursday 21 July 2022 15:23:37 +0000 (0:00:00.025) 0:01:05.656 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-vdo.yml:21 Thursday 21 July 2022 15:23:37 +0000 (0:00:00.024) 0:01:05.681 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-vdo.yml:24 Thursday 21 July 2022 15:23:37 +0000 (0:00:00.024) 0:01:05.705 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-vdo.yml:29 Thursday 21 July 2022 15:23:37 +0000 (0:00:00.023) 0:01:05.728 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-vdo.yml:39 Thursday 21 July 2022 15:23:37 +0000 (0:00:00.023) 0:01:05.752 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:71 Thursday 21 July 2022 15:23:37 +0000 (0:00:00.038) 0:01:05.790 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "__pvs_lvm_len": null, "_storage_test_expected_pv_count": null, "_storage_test_expected_pv_type": null, "_storage_test_pool_pvs": [], "_storage_test_pool_pvs_lvm": [] }, "changed": false } TASK [verify the volumes] ****************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool-volumes.yml:3 Thursday 21 July 2022 15:23:37 +0000 (0:00:00.036) 0:01:05.827 ********* included: /tmp/tmpafnwmbol/tests/test-verify-volume.yml for /cache/rhel-8-y.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'name': 'test1', 'raid_level': None, 'size': '60%', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test1', '_raw_device': '/dev/mapper/foo-test1', '_mount_id': '/dev/mapper/foo-test1', '_kernel_device': '/dev/dm-1', '_raw_kernel_device': '/dev/dm-1'}) included: /tmp/tmpafnwmbol/tests/test-verify-volume.yml for /cache/rhel-8-y.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'ext4', 'mount_options': 'defaults', 'mount_point': '/opt/test2', 'name': 'test2', 'raid_level': None, 'size': '40%', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test2', '_raw_device': '/dev/mapper/foo-test2', '_mount_id': '/dev/mapper/foo-test2', '_kernel_device': '/dev/dm-0', '_raw_kernel_device': '/dev/dm-0'}) TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume.yml:2 Thursday 21 July 2022 15:23:37 +0000 (0:00:00.047) 0:01:05.874 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_volume_present": true, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume.yml:10 Thursday 21 July 2022 15:23:37 +0000 (0:00:00.133) 0:01:06.008 ********* included: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml for /cache/rhel-8-y.qcow2.snap => (item=mount) included: /tmp/tmpafnwmbol/tests/test-verify-volume-fstab.yml for /cache/rhel-8-y.qcow2.snap => (item=fstab) included: /tmp/tmpafnwmbol/tests/test-verify-volume-fs.yml for /cache/rhel-8-y.qcow2.snap => (item=fs) included: /tmp/tmpafnwmbol/tests/test-verify-volume-device.yml for /cache/rhel-8-y.qcow2.snap => (item=device) included: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml for /cache/rhel-8-y.qcow2.snap => (item=encryption) included: /tmp/tmpafnwmbol/tests/test-verify-volume-md.yml for /cache/rhel-8-y.qcow2.snap => (item=md) included: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml for /cache/rhel-8-y.qcow2.snap => (item=size) included: /tmp/tmpafnwmbol/tests/test-verify-volume-cache.yml for /cache/rhel-8-y.qcow2.snap => (item=cache) TASK [Get expected mount device based on device type] ************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml:6 Thursday 21 July 2022 15:23:37 +0000 (0:00:00.077) 0:01:06.086 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/foo-test1" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml:14 Thursday 21 July 2022 15:23:37 +0000 (0:00:00.042) 0:01:06.129 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_mount_device_matches": [ { "block_available": 1551090, "block_size": 4096, "block_total": 1570304, "block_used": 19214, "device": "/dev/mapper/foo-test1", "fstype": "xfs", "inode_available": 3145725, "inode_total": 3145728, "inode_used": 3, "mount": "/opt/test1", "options": "rw,seclabel,relatime,attr2,inode64,logbufs=8,logbsize=32k,noquota", "size_available": 6353264640, "size_total": 6431965184, "uuid": "3db40ee5-790d-4909-8628-933255bd71de" } ], "storage_test_mount_expected_match_count": "1", "storage_test_mount_point_matches": [ { "block_available": 1551090, "block_size": 4096, "block_total": 1570304, "block_used": 19214, "device": "/dev/mapper/foo-test1", "fstype": "xfs", "inode_available": 3145725, "inode_total": 3145728, "inode_used": 3, "mount": "/opt/test1", "options": "rw,seclabel,relatime,attr2,inode64,logbufs=8,logbsize=32k,noquota", "size_available": 6353264640, "size_total": 6431965184, "uuid": "3db40ee5-790d-4909-8628-933255bd71de" } ], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Verify the current mount state by device] ******************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml:28 Thursday 21 July 2022 15:23:37 +0000 (0:00:00.054) 0:01:06.183 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify the current mount state by mount point] *************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml:37 Thursday 21 July 2022 15:23:37 +0000 (0:00:00.050) 0:01:06.234 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify the mount fs type] ************************************************ task path: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml:45 Thursday 21 July 2022 15:23:37 +0000 (0:00:00.046) 0:01:06.281 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [command] ***************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml:54 Thursday 21 July 2022 15:23:37 +0000 (0:00:00.049) 0:01:06.330 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml:58 Thursday 21 July 2022 15:23:37 +0000 (0:00:00.023) 0:01:06.354 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml:63 Thursday 21 July 2022 15:23:37 +0000 (0:00:00.024) 0:01:06.378 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml:75 Thursday 21 July 2022 15:23:37 +0000 (0:00:00.023) 0:01:06.402 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-fstab.yml:2 Thursday 21 July 2022 15:23:37 +0000 (0:00:00.034) 0:01:06.436 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "1", "storage_test_fstab_expected_mount_options_matches": "1", "storage_test_fstab_expected_mount_point_matches": "1", "storage_test_fstab_id_matches": [ "/dev/mapper/foo-test1 " ], "storage_test_fstab_mount_options_matches": [ " /opt/test1 xfs defaults " ], "storage_test_fstab_mount_point_matches": [ " /opt/test1 " ] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-fstab.yml:25 Thursday 21 July 2022 15:23:37 +0000 (0:00:00.062) 0:01:06.498 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify the fstab mount point] ******************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-fstab.yml:32 Thursday 21 July 2022 15:23:38 +0000 (0:00:00.048) 0:01:06.547 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-fstab.yml:39 Thursday 21 July 2022 15:23:38 +0000 (0:00:00.049) 0:01:06.596 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-fstab.yml:49 Thursday 21 July 2022 15:23:38 +0000 (0:00:00.034) 0:01:06.631 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-fs.yml:4 Thursday 21 July 2022 15:23:38 +0000 (0:00:00.036) 0:01:06.667 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify fs label] ********************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-fs.yml:10 Thursday 21 July 2022 15:23:38 +0000 (0:00:00.036) 0:01:06.704 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [See whether the device node is present] ********************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-device.yml:4 Thursday 21 July 2022 15:23:38 +0000 (0:00:00.038) 0:01:06.743 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "stat": { "atime": 1658416980.2722602, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1658416980.2722602, "dev": 6, "device_type": 64769, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 62659, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/symlink", "mode": "0660", "mtime": 1658416980.2722602, "nlink": 1, "path": "/dev/mapper/foo-test1", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-device.yml:10 Thursday 21 July 2022 15:23:38 +0000 (0:00:00.395) 0:01:07.138 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about this volume] ********************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-device.yml:18 Thursday 21 July 2022 15:23:38 +0000 (0:00:00.037) 0:01:07.175 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [(1/2) Process volume type (set initial value)] *************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-device.yml:24 Thursday 21 July 2022 15:23:38 +0000 (0:00:00.039) 0:01:07.215 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [(2/2) Process volume type (get RAID value)] ****************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-device.yml:28 Thursday 21 July 2022 15:23:38 +0000 (0:00:00.035) 0:01:07.251 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-device.yml:33 Thursday 21 July 2022 15:23:38 +0000 (0:00:00.024) 0:01:07.276 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Stat the LUKS device, if encrypted] ************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:3 Thursday 21 July 2022 15:23:38 +0000 (0:00:00.035) 0:01:07.312 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:10 Thursday 21 July 2022 15:23:38 +0000 (0:00:00.022) 0:01:07.334 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [Collect LUKS info for this volume] *************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:15 Thursday 21 July 2022 15:23:40 +0000 (0:00:01.390) 0:01:08.725 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:21 Thursday 21 July 2022 15:23:40 +0000 (0:00:00.024) 0:01:08.749 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:30 Thursday 21 July 2022 15:23:40 +0000 (0:00:00.022) 0:01:08.772 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:38 Thursday 21 July 2022 15:23:40 +0000 (0:00:00.120) 0:01:08.893 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:44 Thursday 21 July 2022 15:23:40 +0000 (0:00:00.024) 0:01:08.917 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:49 Thursday 21 July 2022 15:23:40 +0000 (0:00:00.023) 0:01:08.940 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:55 Thursday 21 July 2022 15:23:40 +0000 (0:00:00.023) 0:01:08.964 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:61 Thursday 21 July 2022 15:23:40 +0000 (0:00:00.024) 0:01:08.989 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:67 Thursday 21 July 2022 15:23:40 +0000 (0:00:00.023) 0:01:09.012 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:74 Thursday 21 July 2022 15:23:40 +0000 (0:00:00.049) 0:01:09.061 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:79 Thursday 21 July 2022 15:23:40 +0000 (0:00:00.048) 0:01:09.110 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:85 Thursday 21 July 2022 15:23:40 +0000 (0:00:00.034) 0:01:09.145 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:91 Thursday 21 July 2022 15:23:40 +0000 (0:00:00.037) 0:01:09.182 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:97 Thursday 21 July 2022 15:23:40 +0000 (0:00:00.034) 0:01:09.217 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [get information about RAID] ********************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-md.yml:7 Thursday 21 July 2022 15:23:40 +0000 (0:00:00.036) 0:01:09.253 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-md.yml:13 Thursday 21 July 2022 15:23:40 +0000 (0:00:00.035) 0:01:09.288 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-md.yml:17 Thursday 21 July 2022 15:23:40 +0000 (0:00:00.035) 0:01:09.324 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-md.yml:21 Thursday 21 July 2022 15:23:40 +0000 (0:00:00.035) 0:01:09.360 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-md.yml:25 Thursday 21 July 2022 15:23:40 +0000 (0:00:00.037) 0:01:09.398 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-md.yml:31 Thursday 21 July 2022 15:23:40 +0000 (0:00:00.038) 0:01:09.437 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-md.yml:37 Thursday 21 July 2022 15:23:40 +0000 (0:00:00.041) 0:01:09.478 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the actual size of the volume] ************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:3 Thursday 21 July 2022 15:23:41 +0000 (0:00:00.040) 0:01:09.518 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "bytes": 6442450944, "changed": false, "lvm": "6g", "parted": "6GiB", "size": "6 GiB" } TASK [parse the requested size of the volume] ********************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:9 Thursday 21 July 2022 15:23:41 +0000 (0:00:00.387) 0:01:09.906 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Establish base value for expected size] ********************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:15 Thursday 21 July 2022 15:23:41 +0000 (0:00:00.039) 0:01:09.946 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:20 Thursday 21 July 2022 15:23:41 +0000 (0:00:00.040) 0:01:09.986 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_expected_size": "4294967296.0" } TASK [debug] ******************************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:25 Thursday 21 July 2022 15:23:41 +0000 (0:00:00.039) 0:01:10.025 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_pool": { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "60%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test2", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test2", "_raw_device": "/dev/mapper/foo-test2", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "40%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } } TASK [debug] ******************************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:28 Thursday 21 July 2022 15:23:41 +0000 (0:00:00.053) 0:01:10.079 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_blkinfo": { "changed": false, "failed": false, "info": { "/dev/mapper/foo-test1": { "fstype": "xfs", "label": "", "name": "/dev/mapper/foo-test1", "size": "6G", "type": "lvm", "uuid": "3db40ee5-790d-4909-8628-933255bd71de" }, "/dev/mapper/foo-test2": { "fstype": "ext4", "label": "", "name": "/dev/mapper/foo-test2", "size": "4G", "type": "lvm", "uuid": "ee7d1310-5176-4ec6-80b8-28181ffffed3" }, "/dev/nvme0n1": { "fstype": "", "label": "", "name": "/dev/nvme0n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme1n1": { "fstype": "", "label": "", "name": "/dev/nvme1n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme2n1": { "fstype": "", "label": "", "name": "/dev/nvme2n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sda": { "fstype": "LVM2_member", "label": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "irzZ3c-YIeb-eoxj-fRdj-JSvm-ls7D-iZNahz" }, "/dev/sdb": { "fstype": "", "label": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc": { "fstype": "", "label": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sr0": { "fstype": "iso9660", "label": "cidata", "name": "/dev/sr0", "size": "364K", "type": "rom", "uuid": "2022-07-21-15-22-18-00" }, "/dev/vda": { "fstype": "", "label": "", "name": "/dev/vda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vda1": { "fstype": "", "label": "", "name": "/dev/vda1", "size": "1M", "type": "partition", "uuid": "" }, "/dev/vda2": { "fstype": "vfat", "label": "", "name": "/dev/vda2", "size": "100M", "type": "partition", "uuid": "7B77-95E7" }, "/dev/vda3": { "fstype": "xfs", "label": "root", "name": "/dev/vda3", "size": "9.9G", "type": "partition", "uuid": "58dca2db-4ccd-4171-a017-dd53ad8c1c54" }, "/dev/vdb": { "fstype": "", "label": "", "name": "/dev/vdb", "size": "2G", "type": "disk", "uuid": "" }, "/dev/vdb1": { "fstype": "ext4", "label": "yumcache", "name": "/dev/vdb1", "size": "2G", "type": "partition", "uuid": "4d1ed9a1-5608-4ff8-ad89-21d95caa66fa" }, "/dev/vdc": { "fstype": "", "label": "", "name": "/dev/vdc", "size": "2G", "type": "disk", "uuid": "" }, "/dev/vdc1": { "fstype": "ext4", "label": "yumvarlib", "name": "/dev/vdc1", "size": "2G", "type": "partition", "uuid": "a7abcc97-edc5-476c-8ccd-694b4621fd28" }, "/dev/vdd": { "fstype": "", "label": "", "name": "/dev/vdd", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vde": { "fstype": "", "label": "", "name": "/dev/vde", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vdf": { "fstype": "", "label": "", "name": "/dev/vdf", "size": "10G", "type": "disk", "uuid": "" } } } } TASK [Get the size of parent/pool device] ************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:31 Thursday 21 July 2022 15:23:41 +0000 (0:00:00.053) 0:01:10.133 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "bytes": 10737418240, "changed": false, "lvm": "10g", "parted": "10GiB", "size": "10 GiB" } TASK [debug] ******************************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:36 Thursday 21 July 2022 15:23:42 +0000 (0:00:00.396) 0:01:10.529 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_pool_size": { "bytes": 10737418240, "changed": false, "failed": false, "lvm": "10g", "parted": "10GiB", "size": "10 GiB" } } TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:39 Thursday 21 July 2022 15:23:42 +0000 (0:00:00.051) 0:01:10.580 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_expected_size": "6442450944.0" }, "changed": false } TASK [debug] ******************************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:44 Thursday 21 July 2022 15:23:42 +0000 (0:00:00.050) 0:01:10.631 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_actual_size": { "bytes": 6442450944, "changed": false, "failed": false, "lvm": "6g", "parted": "6GiB", "size": "6 GiB" } } TASK [debug] ******************************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:47 Thursday 21 July 2022 15:23:42 +0000 (0:00:00.037) 0:01:10.668 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_expected_size": "6442450944.0" } TASK [assert] ****************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:50 Thursday 21 July 2022 15:23:42 +0000 (0:00:00.038) 0:01:10.707 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Get information about the LV] ******************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-cache.yml:6 Thursday 21 July 2022 15:23:42 +0000 (0:00:00.081) 0:01:10.789 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "cmd": [ "lvs", "--noheadings", "--nameprefixes", "--units=b", "--nosuffix", "--unquoted", "-o", "name,attr,cache_total_blocks,chunk_size,segtype", "foo/test1" ], "delta": "0:00:00.032939", "end": "2022-07-21 11:23:42.634176", "rc": 0, "start": "2022-07-21 11:23:42.601237" } STDOUT: LVM2_LV_NAME=test1 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-cache.yml:14 Thursday 21 July 2022 15:23:42 +0000 (0:00:00.455) 0:01:11.245 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_lv_segtype": [ "linear" ] }, "changed": false } TASK [check segment type] ****************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-cache.yml:17 Thursday 21 July 2022 15:23:42 +0000 (0:00:00.100) 0:01:11.345 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-cache.yml:22 Thursday 21 July 2022 15:23:42 +0000 (0:00:00.126) 0:01:11.472 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the requested cache size] ****************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-cache.yml:26 Thursday 21 July 2022 15:23:42 +0000 (0:00:00.038) 0:01:11.510 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-cache.yml:32 Thursday 21 July 2022 15:23:43 +0000 (0:00:00.035) 0:01:11.546 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-cache.yml:36 Thursday 21 July 2022 15:23:43 +0000 (0:00:00.040) 0:01:11.586 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume.yml:16 Thursday 21 July 2022 15:23:43 +0000 (0:00:00.043) 0:01:11.630 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume.yml:2 Thursday 21 July 2022 15:23:43 +0000 (0:00:00.032) 0:01:11.663 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_volume_present": true, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume.yml:10 Thursday 21 July 2022 15:23:43 +0000 (0:00:00.049) 0:01:11.712 ********* included: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml for /cache/rhel-8-y.qcow2.snap => (item=mount) included: /tmp/tmpafnwmbol/tests/test-verify-volume-fstab.yml for /cache/rhel-8-y.qcow2.snap => (item=fstab) included: /tmp/tmpafnwmbol/tests/test-verify-volume-fs.yml for /cache/rhel-8-y.qcow2.snap => (item=fs) included: /tmp/tmpafnwmbol/tests/test-verify-volume-device.yml for /cache/rhel-8-y.qcow2.snap => (item=device) included: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml for /cache/rhel-8-y.qcow2.snap => (item=encryption) included: /tmp/tmpafnwmbol/tests/test-verify-volume-md.yml for /cache/rhel-8-y.qcow2.snap => (item=md) included: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml for /cache/rhel-8-y.qcow2.snap => (item=size) included: /tmp/tmpafnwmbol/tests/test-verify-volume-cache.yml for /cache/rhel-8-y.qcow2.snap => (item=cache) TASK [Get expected mount device based on device type] ************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml:6 Thursday 21 July 2022 15:23:43 +0000 (0:00:00.075) 0:01:11.788 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/foo-test2" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml:14 Thursday 21 July 2022 15:23:43 +0000 (0:00:00.039) 0:01:11.828 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_mount_device_matches": [ { "block_available": 954137, "block_size": 4096, "block_total": 1010616, "block_used": 56479, "device": "/dev/mapper/foo-test2", "fstype": "ext4", "inode_available": 262133, "inode_total": 262144, "inode_used": 11, "mount": "/opt/test2", "options": "rw,seclabel,relatime", "size_available": 3908145152, "size_total": 4139483136, "uuid": "ee7d1310-5176-4ec6-80b8-28181ffffed3" } ], "storage_test_mount_expected_match_count": "1", "storage_test_mount_point_matches": [ { "block_available": 954137, "block_size": 4096, "block_total": 1010616, "block_used": 56479, "device": "/dev/mapper/foo-test2", "fstype": "ext4", "inode_available": 262133, "inode_total": 262144, "inode_used": 11, "mount": "/opt/test2", "options": "rw,seclabel,relatime", "size_available": 3908145152, "size_total": 4139483136, "uuid": "ee7d1310-5176-4ec6-80b8-28181ffffed3" } ], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Verify the current mount state by device] ******************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml:28 Thursday 21 July 2022 15:23:43 +0000 (0:00:00.052) 0:01:11.880 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify the current mount state by mount point] *************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml:37 Thursday 21 July 2022 15:23:43 +0000 (0:00:00.049) 0:01:11.929 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify the mount fs type] ************************************************ task path: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml:45 Thursday 21 July 2022 15:23:43 +0000 (0:00:00.053) 0:01:11.983 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [command] ***************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml:54 Thursday 21 July 2022 15:23:43 +0000 (0:00:00.049) 0:01:12.032 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml:58 Thursday 21 July 2022 15:23:43 +0000 (0:00:00.025) 0:01:12.058 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml:63 Thursday 21 July 2022 15:23:43 +0000 (0:00:00.025) 0:01:12.083 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml:75 Thursday 21 July 2022 15:23:43 +0000 (0:00:00.030) 0:01:12.114 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-fstab.yml:2 Thursday 21 July 2022 15:23:43 +0000 (0:00:00.035) 0:01:12.150 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "1", "storage_test_fstab_expected_mount_options_matches": "1", "storage_test_fstab_expected_mount_point_matches": "1", "storage_test_fstab_id_matches": [ "/dev/mapper/foo-test2 " ], "storage_test_fstab_mount_options_matches": [ " /opt/test2 ext4 defaults " ], "storage_test_fstab_mount_point_matches": [ " /opt/test2 " ] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-fstab.yml:25 Thursday 21 July 2022 15:23:43 +0000 (0:00:00.065) 0:01:12.216 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify the fstab mount point] ******************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-fstab.yml:32 Thursday 21 July 2022 15:23:43 +0000 (0:00:00.056) 0:01:12.272 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-fstab.yml:39 Thursday 21 July 2022 15:23:43 +0000 (0:00:00.058) 0:01:12.331 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-fstab.yml:49 Thursday 21 July 2022 15:23:43 +0000 (0:00:00.035) 0:01:12.367 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-fs.yml:4 Thursday 21 July 2022 15:23:43 +0000 (0:00:00.034) 0:01:12.402 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify fs label] ********************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-fs.yml:10 Thursday 21 July 2022 15:23:43 +0000 (0:00:00.041) 0:01:12.443 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [See whether the device node is present] ********************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-device.yml:4 Thursday 21 July 2022 15:23:43 +0000 (0:00:00.040) 0:01:12.483 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "stat": { "atime": 1658417009.3442602, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1658417009.3442602, "dev": 6, "device_type": 64768, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 62525, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/symlink", "mode": "0660", "mtime": 1658417009.3442602, "nlink": 1, "path": "/dev/mapper/foo-test2", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-device.yml:10 Thursday 21 July 2022 15:23:44 +0000 (0:00:00.389) 0:01:12.873 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about this volume] ********************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-device.yml:18 Thursday 21 July 2022 15:23:44 +0000 (0:00:00.041) 0:01:12.914 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [(1/2) Process volume type (set initial value)] *************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-device.yml:24 Thursday 21 July 2022 15:23:44 +0000 (0:00:00.037) 0:01:12.951 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [(2/2) Process volume type (get RAID value)] ****************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-device.yml:28 Thursday 21 July 2022 15:23:44 +0000 (0:00:00.035) 0:01:12.987 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-device.yml:33 Thursday 21 July 2022 15:23:44 +0000 (0:00:00.023) 0:01:13.010 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Stat the LUKS device, if encrypted] ************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:3 Thursday 21 July 2022 15:23:44 +0000 (0:00:00.039) 0:01:13.050 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:10 Thursday 21 July 2022 15:23:44 +0000 (0:00:00.023) 0:01:13.074 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [Collect LUKS info for this volume] *************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:15 Thursday 21 July 2022 15:23:45 +0000 (0:00:01.349) 0:01:14.423 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:21 Thursday 21 July 2022 15:23:45 +0000 (0:00:00.027) 0:01:14.450 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:30 Thursday 21 July 2022 15:23:45 +0000 (0:00:00.024) 0:01:14.474 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:38 Thursday 21 July 2022 15:23:46 +0000 (0:00:00.050) 0:01:14.525 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:44 Thursday 21 July 2022 15:23:46 +0000 (0:00:00.022) 0:01:14.548 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:49 Thursday 21 July 2022 15:23:46 +0000 (0:00:00.025) 0:01:14.573 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:55 Thursday 21 July 2022 15:23:46 +0000 (0:00:00.024) 0:01:14.599 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:61 Thursday 21 July 2022 15:23:46 +0000 (0:00:00.023) 0:01:14.622 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:67 Thursday 21 July 2022 15:23:46 +0000 (0:00:00.026) 0:01:14.649 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:74 Thursday 21 July 2022 15:23:46 +0000 (0:00:00.055) 0:01:14.704 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:79 Thursday 21 July 2022 15:23:46 +0000 (0:00:00.050) 0:01:14.754 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:85 Thursday 21 July 2022 15:23:46 +0000 (0:00:00.037) 0:01:14.792 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:91 Thursday 21 July 2022 15:23:46 +0000 (0:00:00.038) 0:01:14.831 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:97 Thursday 21 July 2022 15:23:46 +0000 (0:00:00.041) 0:01:14.872 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [get information about RAID] ********************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-md.yml:7 Thursday 21 July 2022 15:23:46 +0000 (0:00:00.033) 0:01:14.905 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-md.yml:13 Thursday 21 July 2022 15:23:46 +0000 (0:00:00.036) 0:01:14.942 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-md.yml:17 Thursday 21 July 2022 15:23:46 +0000 (0:00:00.035) 0:01:14.978 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-md.yml:21 Thursday 21 July 2022 15:23:46 +0000 (0:00:00.041) 0:01:15.019 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-md.yml:25 Thursday 21 July 2022 15:23:46 +0000 (0:00:00.035) 0:01:15.055 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-md.yml:31 Thursday 21 July 2022 15:23:46 +0000 (0:00:00.035) 0:01:15.090 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-md.yml:37 Thursday 21 July 2022 15:23:46 +0000 (0:00:00.046) 0:01:15.137 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the actual size of the volume] ************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:3 Thursday 21 July 2022 15:23:46 +0000 (0:00:00.044) 0:01:15.181 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "bytes": 4294967296, "changed": false, "lvm": "4g", "parted": "4GiB", "size": "4 GiB" } TASK [parse the requested size of the volume] ********************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:9 Thursday 21 July 2022 15:23:47 +0000 (0:00:00.387) 0:01:15.569 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Establish base value for expected size] ********************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:15 Thursday 21 July 2022 15:23:47 +0000 (0:00:00.040) 0:01:15.609 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:20 Thursday 21 July 2022 15:23:47 +0000 (0:00:00.040) 0:01:15.650 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_expected_size": "6442450944.0" } TASK [debug] ******************************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:25 Thursday 21 July 2022 15:23:47 +0000 (0:00:00.038) 0:01:15.688 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_pool": { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "60%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test2", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test2", "_raw_device": "/dev/mapper/foo-test2", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "40%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } } TASK [debug] ******************************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:28 Thursday 21 July 2022 15:23:47 +0000 (0:00:00.089) 0:01:15.778 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_blkinfo": { "changed": false, "failed": false, "info": { "/dev/mapper/foo-test1": { "fstype": "xfs", "label": "", "name": "/dev/mapper/foo-test1", "size": "6G", "type": "lvm", "uuid": "3db40ee5-790d-4909-8628-933255bd71de" }, "/dev/mapper/foo-test2": { "fstype": "ext4", "label": "", "name": "/dev/mapper/foo-test2", "size": "4G", "type": "lvm", "uuid": "ee7d1310-5176-4ec6-80b8-28181ffffed3" }, "/dev/nvme0n1": { "fstype": "", "label": "", "name": "/dev/nvme0n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme1n1": { "fstype": "", "label": "", "name": "/dev/nvme1n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme2n1": { "fstype": "", "label": "", "name": "/dev/nvme2n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sda": { "fstype": "LVM2_member", "label": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "irzZ3c-YIeb-eoxj-fRdj-JSvm-ls7D-iZNahz" }, "/dev/sdb": { "fstype": "", "label": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc": { "fstype": "", "label": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sr0": { "fstype": "iso9660", "label": "cidata", "name": "/dev/sr0", "size": "364K", "type": "rom", "uuid": "2022-07-21-15-22-18-00" }, "/dev/vda": { "fstype": "", "label": "", "name": "/dev/vda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vda1": { "fstype": "", "label": "", "name": "/dev/vda1", "size": "1M", "type": "partition", "uuid": "" }, "/dev/vda2": { "fstype": "vfat", "label": "", "name": "/dev/vda2", "size": "100M", "type": "partition", "uuid": "7B77-95E7" }, "/dev/vda3": { "fstype": "xfs", "label": "root", "name": "/dev/vda3", "size": "9.9G", "type": "partition", "uuid": "58dca2db-4ccd-4171-a017-dd53ad8c1c54" }, "/dev/vdb": { "fstype": "", "label": "", "name": "/dev/vdb", "size": "2G", "type": "disk", "uuid": "" }, "/dev/vdb1": { "fstype": "ext4", "label": "yumcache", "name": "/dev/vdb1", "size": "2G", "type": "partition", "uuid": "4d1ed9a1-5608-4ff8-ad89-21d95caa66fa" }, "/dev/vdc": { "fstype": "", "label": "", "name": "/dev/vdc", "size": "2G", "type": "disk", "uuid": "" }, "/dev/vdc1": { "fstype": "ext4", "label": "yumvarlib", "name": "/dev/vdc1", "size": "2G", "type": "partition", "uuid": "a7abcc97-edc5-476c-8ccd-694b4621fd28" }, "/dev/vdd": { "fstype": "", "label": "", "name": "/dev/vdd", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vde": { "fstype": "", "label": "", "name": "/dev/vde", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vdf": { "fstype": "", "label": "", "name": "/dev/vdf", "size": "10G", "type": "disk", "uuid": "" } } } } TASK [Get the size of parent/pool device] ************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:31 Thursday 21 July 2022 15:23:47 +0000 (0:00:00.057) 0:01:15.836 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "bytes": 10737418240, "changed": false, "lvm": "10g", "parted": "10GiB", "size": "10 GiB" } TASK [debug] ******************************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:36 Thursday 21 July 2022 15:23:47 +0000 (0:00:00.430) 0:01:16.267 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_pool_size": { "bytes": 10737418240, "changed": false, "failed": false, "lvm": "10g", "parted": "10GiB", "size": "10 GiB" } } TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:39 Thursday 21 July 2022 15:23:47 +0000 (0:00:00.085) 0:01:16.352 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_expected_size": "4294967296.0" }, "changed": false } TASK [debug] ******************************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:44 Thursday 21 July 2022 15:23:47 +0000 (0:00:00.087) 0:01:16.440 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_actual_size": { "bytes": 4294967296, "changed": false, "failed": false, "lvm": "4g", "parted": "4GiB", "size": "4 GiB" } } TASK [debug] ******************************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:47 Thursday 21 July 2022 15:23:48 +0000 (0:00:00.079) 0:01:16.519 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_expected_size": "4294967296.0" } TASK [assert] ****************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:50 Thursday 21 July 2022 15:23:48 +0000 (0:00:00.102) 0:01:16.621 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Get information about the LV] ******************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-cache.yml:6 Thursday 21 July 2022 15:23:48 +0000 (0:00:00.055) 0:01:16.676 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "cmd": [ "lvs", "--noheadings", "--nameprefixes", "--units=b", "--nosuffix", "--unquoted", "-o", "name,attr,cache_total_blocks,chunk_size,segtype", "foo/test2" ], "delta": "0:00:00.039840", "end": "2022-07-21 11:23:48.483646", "rc": 0, "start": "2022-07-21 11:23:48.443806" } STDOUT: LVM2_LV_NAME=test2 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-cache.yml:14 Thursday 21 July 2022 15:23:48 +0000 (0:00:00.418) 0:01:17.095 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_lv_segtype": [ "linear" ] }, "changed": false } TASK [check segment type] ****************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-cache.yml:17 Thursday 21 July 2022 15:23:48 +0000 (0:00:00.049) 0:01:17.144 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-cache.yml:22 Thursday 21 July 2022 15:23:48 +0000 (0:00:00.050) 0:01:17.194 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the requested cache size] ****************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-cache.yml:26 Thursday 21 July 2022 15:23:48 +0000 (0:00:00.041) 0:01:17.236 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-cache.yml:32 Thursday 21 July 2022 15:23:48 +0000 (0:00:00.038) 0:01:17.274 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-cache.yml:36 Thursday 21 July 2022 15:23:48 +0000 (0:00:00.039) 0:01:17.313 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume.yml:16 Thursday 21 July 2022 15:23:48 +0000 (0:00:00.036) 0:01:17.350 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [Verify the volumes with no pool were correctly managed] ****************** task path: /tmp/tmpafnwmbol/tests/verify-role-results.yml:43 Thursday 21 July 2022 15:23:48 +0000 (0:00:00.036) 0:01:17.386 ********* TASK [Clean up variable namespace] ********************************************* task path: /tmp/tmpafnwmbol/tests/verify-role-results.yml:53 Thursday 21 July 2022 15:23:48 +0000 (0:00:00.022) 0:01:17.409 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_blkinfo": null, "storage_test_crypttab": null, "storage_test_fstab": null }, "changed": false } TASK [Shrink test2 volume via percentage-based size spec] ********************** task path: /tmp/tmpafnwmbol/tests/tests_lvm_percent_size.yml:97 Thursday 21 July 2022 15:23:48 +0000 (0:00:00.034) 0:01:17.443 ********* TASK [linux-system-roles.storage : set platform/version specific variables] **** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main.yml:2 Thursday 21 July 2022 15:23:48 +0000 (0:00:00.051) 0:01:17.495 ********* included: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for /cache/rhel-8-y.qcow2.snap TASK [linux-system-roles.storage : Ensure ansible_facts used by role] ********** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2 Thursday 21 July 2022 15:23:49 +0000 (0:00:00.034) 0:01:17.529 ********* ok: [/cache/rhel-8-y.qcow2.snap] TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8 Thursday 21 July 2022 15:23:49 +0000 (0:00:00.524) 0:01:18.054 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [/cache/rhel-8-y.qcow2.snap] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [/cache/rhel-8-y.qcow2.snap] => (item=RedHat_8.yml) => { "ansible_facts": { "blivet_package_list": [ "python3-blivet", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap", "vdo", "kmod-kvdo", "xfsprogs" ] }, "ansible_included_var_files": [ "/tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/vars/RedHat_8.yml" ], "ansible_loop_var": "item", "changed": false, "item": "RedHat_8.yml" } skipping: [/cache/rhel-8-y.qcow2.snap] => (item=RedHat_8.7.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat_8.7.yml", "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : define an empty list of pools to be used in testing] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main.yml:5 Thursday 21 July 2022 15:23:49 +0000 (0:00:00.065) 0:01:18.120 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : define an empty list of volumes to be used in testing] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main.yml:9 Thursday 21 July 2022 15:23:49 +0000 (0:00:00.036) 0:01:18.156 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : include the appropriate provider tasks] ***** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main.yml:13 Thursday 21 July 2022 15:23:49 +0000 (0:00:00.036) 0:01:18.193 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount included: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for /cache/rhel-8-y.qcow2.snap TASK [linux-system-roles.storage : get a list of rpm packages installed on host machine] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2 Thursday 21 July 2022 15:23:49 +0000 (0:00:00.052) 0:01:18.245 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : make sure blivet is available] ************** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:7 Thursday 21 July 2022 15:23:49 +0000 (0:00:00.022) 0:01:18.267 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [linux-system-roles.storage : show storage_pools] ************************* task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:14 Thursday 21 July 2022 15:23:51 +0000 (0:00:01.385) 0:01:19.652 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_pools": [ { "disks": [ "sda" ], "name": "foo", "volumes": [ { "mount_point": "/opt/test1", "name": "test1", "size": "60%" }, { "mount_point": "/opt/test2", "name": "test2", "size": "25%" } ] } ] } TASK [linux-system-roles.storage : show storage_volumes] *********************** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:19 Thursday 21 July 2022 15:23:51 +0000 (0:00:00.040) 0:01:19.693 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [linux-system-roles.storage : get required packages] ********************** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:24 Thursday 21 July 2022 15:23:51 +0000 (0:00:00.035) 0:01:19.728 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [ "lvm2" ], "pools": [], "volumes": [] } TASK [linux-system-roles.storage : enable copr repositories if needed] ********* task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:37 Thursday 21 July 2022 15:23:52 +0000 (0:00:01.571) 0:01:21.299 ********* included: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml for /cache/rhel-8-y.qcow2.snap TASK [linux-system-roles.storage : check if the COPR support packages should be installed] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:2 Thursday 21 July 2022 15:23:52 +0000 (0:00:00.045) 0:01:21.345 ********* TASK [linux-system-roles.storage : make sure COPR support packages are present] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:13 Thursday 21 July 2022 15:23:52 +0000 (0:00:00.035) 0:01:21.381 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : enable COPRs] ******************************* task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:18 Thursday 21 July 2022 15:23:52 +0000 (0:00:00.038) 0:01:21.420 ********* TASK [linux-system-roles.storage : make sure required packages are installed] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:44 Thursday 21 July 2022 15:23:52 +0000 (0:00:00.038) 0:01:21.458 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [linux-system-roles.storage : get service facts] ************************** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:51 Thursday 21 July 2022 15:23:54 +0000 (0:00:01.351) 0:01:22.809 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "services": { "NetworkManager-dispatcher.service": { "name": "NetworkManager-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "NetworkManager-wait-online.service": { "name": "NetworkManager-wait-online.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "NetworkManager.service": { "name": "NetworkManager.service", "source": "systemd", "state": "running", "status": "enabled" }, "arp-ethers.service": { "name": "arp-ethers.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "auditd.service": { "name": "auditd.service", "source": "systemd", "state": "running", "status": "enabled" }, "auth-rpcgss-module.service": { "name": "auth-rpcgss-module.service", "source": "systemd", "state": "stopped", "status": "static" }, "autovt@.service": { "name": "autovt@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "blivet.service": { "name": "blivet.service", "source": "systemd", "state": "inactive", "status": "static" }, "blk-availability.service": { "name": "blk-availability.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "chrony-dnssrv@.service": { "name": "chrony-dnssrv@.service", "source": "systemd", "state": "unknown", "status": "static" }, "chrony-wait.service": { "name": "chrony-wait.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd.service": { "name": "chronyd.service", "source": "systemd", "state": "running", "status": "enabled" }, "cloud-config.service": { "name": "cloud-config.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-final.service": { "name": "cloud-final.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init-hotplugd.service": { "name": "cloud-init-hotplugd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cloud-init-local.service": { "name": "cloud-init-local.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init.service": { "name": "cloud-init.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cockpit-motd.service": { "name": "cockpit-motd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cockpit-wsinstance-http.service": { "name": "cockpit-wsinstance-http.service", "source": "systemd", "state": "inactive", "status": "static" }, "cockpit-wsinstance-https-factory@.service": { "name": "cockpit-wsinstance-https-factory@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cockpit-wsinstance-https@.service": { "name": "cockpit-wsinstance-https@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cockpit.service": { "name": "cockpit.service", "source": "systemd", "state": "inactive", "status": "static" }, "console-getty.service": { "name": "console-getty.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "container-getty@.service": { "name": "container-getty@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cpupower.service": { "name": "cpupower.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "crond.service": { "name": "crond.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-org.freedesktop.hostname1.service": { "name": "dbus-org.freedesktop.hostname1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.locale1.service": { "name": "dbus-org.freedesktop.locale1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.login1.service": { "name": "dbus-org.freedesktop.login1.service", "source": "systemd", "state": "active", "status": "static" }, "dbus-org.freedesktop.nm-dispatcher.service": { "name": "dbus-org.freedesktop.nm-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "dbus-org.freedesktop.portable1.service": { "name": "dbus-org.freedesktop.portable1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.timedate1.service": { "name": "dbus-org.freedesktop.timedate1.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "dbus.service": { "name": "dbus.service", "source": "systemd", "state": "running", "status": "static" }, "dbxtool.service": { "name": "dbxtool.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "debug-shell.service": { "name": "debug-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "display-manager.service": { "name": "display-manager.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "dm-event.service": { "name": "dm-event.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-makecache.service": { "name": "dnf-makecache.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-cmdline.service": { "name": "dracut-cmdline.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-initqueue.service": { "name": "dracut-initqueue.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-mount.service": { "name": "dracut-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-mount.service": { "name": "dracut-pre-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-pivot.service": { "name": "dracut-pre-pivot.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-trigger.service": { "name": "dracut-pre-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-udev.service": { "name": "dracut-pre-udev.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown.service": { "name": "dracut-shutdown.service", "source": "systemd", "state": "stopped", "status": "static" }, "emergency.service": { "name": "emergency.service", "source": "systemd", "state": "stopped", "status": "static" }, "fcoe.service": { "name": "fcoe.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "fstrim.service": { "name": "fstrim.service", "source": "systemd", "state": "inactive", "status": "static" }, "getty@.service": { "name": "getty@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "getty@tty1.service": { "name": "getty@tty1.service", "source": "systemd", "state": "running", "status": "active" }, "grub-boot-indeterminate.service": { "name": "grub-boot-indeterminate.service", "source": "systemd", "state": "inactive", "status": "static" }, "gssproxy.service": { "name": "gssproxy.service", "source": "systemd", "state": "running", "status": "disabled" }, "halt-local.service": { "name": "halt-local.service", "source": "systemd", "state": "inactive", "status": "static" }, "hv_kvp_daemon.service": { "name": "hv_kvp_daemon.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "import-state.service": { "name": "import-state.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "initrd-cleanup.service": { "name": "initrd-cleanup.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-parse-etc.service": { "name": "initrd-parse-etc.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-switch-root.service": { "name": "initrd-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-udevadm-cleanup-db.service": { "name": "initrd-udevadm-cleanup-db.service", "source": "systemd", "state": "stopped", "status": "static" }, "insights-client-boot.service": { "name": "insights-client-boot.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "insights-client-results.service": { "name": "insights-client-results.service", "source": "systemd", "state": "inactive", "status": "static" }, "insights-client.service": { "name": "insights-client.service", "source": "systemd", "state": "inactive", "status": "static" }, "irqbalance.service": { "name": "irqbalance.service", "source": "systemd", "state": "running", "status": "enabled" }, "iscsi-shutdown.service": { "name": "iscsi-shutdown.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "iscsi.service": { "name": "iscsi.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "iscsid.service": { "name": "iscsid.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "kdump.service": { "name": "kdump.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "kmod-static-nodes.service": { "name": "kmod-static-nodes.service", "source": "systemd", "state": "stopped", "status": "static" }, "kvm_stat.service": { "name": "kvm_stat.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "ldconfig.service": { "name": "ldconfig.service", "source": "systemd", "state": "stopped", "status": "static" }, "loadmodules.service": { "name": "loadmodules.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "lvm2-activation-early.service": { "name": "lvm2-activation-early.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "lvm2-activation.service": { "name": "lvm2-activation.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "lvm2-lvmpolld.service": { "name": "lvm2-lvmpolld.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-monitor.service": { "name": "lvm2-monitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "lvm2-pvscan@.service": { "name": "lvm2-pvscan@.service", "source": "systemd", "state": "unknown", "status": "static" }, "lvm2-pvscan@8:0.service": { "name": "lvm2-pvscan@8:0.service", "source": "systemd", "state": "stopped", "status": "active" }, "man-db-cache-update.service": { "name": "man-db-cache-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "man-db-restart-cache-update.service": { "name": "man-db-restart-cache-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "mdadm-grow-continue@.service": { "name": "mdadm-grow-continue@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdadm-last-resort@.service": { "name": "mdadm-last-resort@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdcheck_continue.service": { "name": "mdcheck_continue.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdcheck_start.service": { "name": "mdcheck_start.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmon@.service": { "name": "mdmon@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdmonitor-oneshot.service": { "name": "mdmonitor-oneshot.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmonitor.service": { "name": "mdmonitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "messagebus.service": { "name": "messagebus.service", "source": "systemd", "state": "active", "status": "static" }, "microcode.service": { "name": "microcode.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "multipathd.service": { "name": "multipathd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "ndctl-monitor.service": { "name": "ndctl-monitor.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "network.service": { "name": "network.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "nfs-blkmap.service": { "name": "nfs-blkmap.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-convert.service": { "name": "nfs-convert.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-idmapd.service": { "name": "nfs-idmapd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-mountd.service": { "name": "nfs-mountd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-server.service": { "name": "nfs-server.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-utils.service": { "name": "nfs-utils.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfsdcld.service": { "name": "nfsdcld.service", "source": "systemd", "state": "stopped", "status": "static" }, "nis-domainname.service": { "name": "nis-domainname.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "ntpd.service": { "name": "ntpd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ntpdate.service": { "name": "ntpdate.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "oddjobd.service": { "name": "oddjobd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "packagekit-offline-update.service": { "name": "packagekit-offline-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "packagekit.service": { "name": "packagekit.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-quit-wait.service": { "name": "plymouth-quit-wait.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "plymouth-start.service": { "name": "plymouth-start.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "polkit.service": { "name": "polkit.service", "source": "systemd", "state": "running", "status": "static" }, "qemu-guest-agent.service": { "name": "qemu-guest-agent.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "quotaon.service": { "name": "quotaon.service", "source": "systemd", "state": "inactive", "status": "static" }, "rbdmap.service": { "name": "rbdmap.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "rc-local.service": { "name": "rc-local.service", "source": "systemd", "state": "stopped", "status": "static" }, "rdisc.service": { "name": "rdisc.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rescue.service": { "name": "rescue.service", "source": "systemd", "state": "stopped", "status": "static" }, "rhcd.service": { "name": "rhcd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhsm-facts.service": { "name": "rhsm-facts.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhsm.service": { "name": "rhsm.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhsmcertd.service": { "name": "rhsmcertd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpc-gssd.service": { "name": "rpc-gssd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd-notify.service": { "name": "rpc-statd-notify.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd.service": { "name": "rpc-statd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpcbind.service": { "name": "rpcbind.service", "source": "systemd", "state": "running", "status": "enabled" }, "rsyslog.service": { "name": "rsyslog.service", "source": "systemd", "state": "running", "status": "enabled" }, "selinux-autorelabel-mark.service": { "name": "selinux-autorelabel-mark.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "selinux-autorelabel.service": { "name": "selinux-autorelabel.service", "source": "systemd", "state": "inactive", "status": "static" }, "serial-getty@.service": { "name": "serial-getty@.service", "source": "systemd", "state": "unknown", "status": "indirect" }, "serial-getty@ttyS0.service": { "name": "serial-getty@ttyS0.service", "source": "systemd", "state": "running", "status": "active" }, "snapd.seeded.service": { "name": "snapd.seeded.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sntp.service": { "name": "sntp.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sshd-keygen.service": { "name": "sshd-keygen.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sshd-keygen@.service": { "name": "sshd-keygen@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@ecdsa.service": { "name": "sshd-keygen@ecdsa.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd-keygen@ed25519.service": { "name": "sshd-keygen@ed25519.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd-keygen@rsa.service": { "name": "sshd-keygen@rsa.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd.service": { "name": "sshd.service", "source": "systemd", "state": "running", "status": "enabled" }, "sshd@.service": { "name": "sshd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "sssd-autofs.service": { "name": "sssd-autofs.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-kcm.service": { "name": "sssd-kcm.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "sssd-nss.service": { "name": "sssd-nss.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pac.service": { "name": "sssd-pac.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pam.service": { "name": "sssd-pam.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-ssh.service": { "name": "sssd-ssh.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-sudo.service": { "name": "sssd-sudo.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd.service": { "name": "sssd.service", "source": "systemd", "state": "running", "status": "enabled" }, "syslog.service": { "name": "syslog.service", "source": "systemd", "state": "active", "status": "enabled" }, "system-update-cleanup.service": { "name": "system-update-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-ask-password-console.service": { "name": "systemd-ask-password-console.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-wall.service": { "name": "systemd-ask-password-wall.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-backlight@.service": { "name": "systemd-backlight@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-binfmt.service": { "name": "systemd-binfmt.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-coredump@.service": { "name": "systemd-coredump@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-exit.service": { "name": "systemd-exit.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-firstboot.service": { "name": "systemd-firstboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck-root.service": { "name": "systemd-fsck-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck@.service": { "name": "systemd-fsck@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-fsck@dev-disk-by\\x2duuid-7B77\\x2d95E7.service": { "name": "systemd-fsck@dev-disk-by\\x2duuid-7B77\\x2d95E7.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-fsck@dev-vdb1.service": { "name": "systemd-fsck@dev-vdb1.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-fsck@dev-vdc1.service": { "name": "systemd-fsck@dev-vdc1.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-halt.service": { "name": "systemd-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hibernate-resume@.service": { "name": "systemd-hibernate-resume@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-hibernate.service": { "name": "systemd-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hostnamed.service": { "name": "systemd-hostnamed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hwdb-update.service": { "name": "systemd-hwdb-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-hybrid-sleep.service": { "name": "systemd-hybrid-sleep.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-initctl.service": { "name": "systemd-initctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-catalog-update.service": { "name": "systemd-journal-catalog-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-flush.service": { "name": "systemd-journal-flush.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journald.service": { "name": "systemd-journald.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-kexec.service": { "name": "systemd-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-localed.service": { "name": "systemd-localed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-logind.service": { "name": "systemd-logind.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-machine-id-commit.service": { "name": "systemd-machine-id-commit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-modules-load.service": { "name": "systemd-modules-load.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-networkd-wait-online.service": { "name": "systemd-networkd-wait-online.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "systemd-portabled.service": { "name": "systemd-portabled.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-poweroff.service": { "name": "systemd-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-quotacheck.service": { "name": "systemd-quotacheck.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-random-seed.service": { "name": "systemd-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-reboot.service": { "name": "systemd-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-remount-fs.service": { "name": "systemd-remount-fs.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-resolved.service": { "name": "systemd-resolved.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-rfkill.service": { "name": "systemd-rfkill.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend-then-hibernate.service": { "name": "systemd-suspend-then-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend.service": { "name": "systemd-suspend.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-sysctl.service": { "name": "systemd-sysctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-sysusers.service": { "name": "systemd-sysusers.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-timedated.service": { "name": "systemd-timedated.service", "source": "systemd", "state": "inactive", "status": "masked" }, "systemd-timesyncd.service": { "name": "systemd-timesyncd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "systemd-tmpfiles-clean.service": { "name": "systemd-tmpfiles-clean.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev.service": { "name": "systemd-tmpfiles-setup-dev.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup.service": { "name": "systemd-tmpfiles-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-settle.service": { "name": "systemd-udev-settle.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-trigger.service": { "name": "systemd-udev-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udevd.service": { "name": "systemd-udevd.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-update-done.service": { "name": "systemd-update-done.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp-runlevel.service": { "name": "systemd-update-utmp-runlevel.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp.service": { "name": "systemd-update-utmp.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-user-sessions.service": { "name": "systemd-user-sessions.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-vconsole-setup.service": { "name": "systemd-vconsole-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-volatile-root.service": { "name": "systemd-volatile-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "tcsd.service": { "name": "tcsd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "teamd@.service": { "name": "teamd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "timedatex.service": { "name": "timedatex.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "tuned.service": { "name": "tuned.service", "source": "systemd", "state": "running", "status": "enabled" }, "unbound-anchor.service": { "name": "unbound-anchor.service", "source": "systemd", "state": "stopped", "status": "static" }, "user-runtime-dir@.service": { "name": "user-runtime-dir@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user-runtime-dir@0.service": { "name": "user-runtime-dir@0.service", "source": "systemd", "state": "stopped", "status": "active" }, "user@.service": { "name": "user@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user@0.service": { "name": "user@0.service", "source": "systemd", "state": "running", "status": "active" }, "vdo-start-by-dev@.service": { "name": "vdo-start-by-dev@.service", "source": "systemd", "state": "unknown", "status": "static" }, "vdo.service": { "name": "vdo.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "ypbind.service": { "name": "ypbind.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "yppasswdd.service": { "name": "yppasswdd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ypserv.service": { "name": "ypserv.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ypxfrd.service": { "name": "ypxfrd.service", "source": "systemd", "state": "stopped", "status": "not-found" } } }, "changed": false } TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************ task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:58 Thursday 21 July 2022 15:23:56 +0000 (0:00:01.746) 0:01:24.556 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] ******* task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:71 Thursday 21 July 2022 15:23:56 +0000 (0:00:00.059) 0:01:24.615 ********* TASK [linux-system-roles.storage : manage the pools and volumes to match the specified state] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77 Thursday 21 July 2022 15:23:56 +0000 (0:00:00.023) 0:01:24.639 ********* changed: [/cache/rhel-8-y.qcow2.snap] => { "actions": [ { "action": "resize format", "device": "/dev/mapper/foo-test2", "fs_type": "ext4" }, { "action": "resize device", "device": "/dev/mapper/foo-test2", "fs_type": null } ], "changed": true, "crypts": [], "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/vda2", "/dev/vda3", "/dev/mapper/foo-test1", "/dev/mapper/foo-test2", "/dev/sdb", "/dev/sdc", "/dev/nvme0n1", "/dev/nvme1n1", "/dev/nvme2n1", "/dev/vdb1", "/dev/vdc1", "/dev/vdd", "/dev/vde", "/dev/vdf" ], "mounts": [ { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" }, { "dump": 0, "fstype": "ext4", "opts": "defaults", "passno": 0, "path": "/opt/test2", "src": "/dev/mapper/foo-test2", "state": "mounted" } ], "packages": [ "lvm2", "e2fsprogs", "dosfstools", "xfsprogs" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "60%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test2", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test2", "_raw_device": "/dev/mapper/foo-test2", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "25%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } TASK [linux-system-roles.storage : Workaround for udev issue on some platforms] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:91 Thursday 21 July 2022 15:23:58 +0000 (0:00:02.293) 0:01:26.932 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] ***** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:103 Thursday 21 July 2022 15:23:58 +0000 (0:00:00.045) 0:01:26.978 ********* TASK [linux-system-roles.storage : show blivet_output] ************************* task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:109 Thursday 21 July 2022 15:23:58 +0000 (0:00:00.023) 0:01:27.002 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "blivet_output": { "actions": [ { "action": "resize format", "device": "/dev/mapper/foo-test2", "fs_type": "ext4" }, { "action": "resize device", "device": "/dev/mapper/foo-test2", "fs_type": null } ], "changed": true, "crypts": [], "failed": false, "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/vda2", "/dev/vda3", "/dev/mapper/foo-test1", "/dev/mapper/foo-test2", "/dev/sdb", "/dev/sdc", "/dev/nvme0n1", "/dev/nvme1n1", "/dev/nvme2n1", "/dev/vdb1", "/dev/vdc1", "/dev/vdd", "/dev/vde", "/dev/vdf" ], "mounts": [ { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" }, { "dump": 0, "fstype": "ext4", "opts": "defaults", "passno": 0, "path": "/opt/test2", "src": "/dev/mapper/foo-test2", "state": "mounted" } ], "packages": [ "lvm2", "e2fsprogs", "dosfstools", "xfsprogs" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "60%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test2", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test2", "_raw_device": "/dev/mapper/foo-test2", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "25%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } } TASK [linux-system-roles.storage : set the list of pools for test verification] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:114 Thursday 21 July 2022 15:23:58 +0000 (0:00:00.079) 0:01:27.081 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "60%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test2", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test2", "_raw_device": "/dev/mapper/foo-test2", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "25%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] }, "changed": false } TASK [linux-system-roles.storage : set the list of volumes for test verification] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:118 Thursday 21 July 2022 15:23:58 +0000 (0:00:00.040) 0:01:27.122 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : remove obsolete mounts] ********************* task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:134 Thursday 21 July 2022 15:23:58 +0000 (0:00:00.071) 0:01:27.194 ********* TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:146 Thursday 21 July 2022 15:23:58 +0000 (0:00:00.043) 0:01:27.237 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : set up new/current mounts] ****************** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:151 Thursday 21 July 2022 15:23:59 +0000 (0:00:00.687) 0:01:27.925 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount ok: [/cache/rhel-8-y.qcow2.snap] => (item={'src': '/dev/mapper/foo-test1', 'path': '/opt/test1', 'fstype': 'xfs', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted'}) => { "ansible_loop_var": "mount_info", "backup_file": "", "boot": "yes", "changed": false, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" }, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test1" } redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount changed: [/cache/rhel-8-y.qcow2.snap] => (item={'src': '/dev/mapper/foo-test2', 'path': '/opt/test2', 'fstype': 'ext4', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted'}) => { "ansible_loop_var": "mount_info", "backup_file": "", "boot": "yes", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "ext4", "mount_info": { "dump": 0, "fstype": "ext4", "opts": "defaults", "passno": 0, "path": "/opt/test2", "src": "/dev/mapper/foo-test2", "state": "mounted" }, "name": "/opt/test2", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test2" } TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:163 Thursday 21 July 2022 15:24:00 +0000 (0:00:00.782) 0:01:28.708 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : retrieve facts for the /etc/crypttab file] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:171 Thursday 21 July 2022 15:24:00 +0000 (0:00:00.621) 0:01:29.329 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "stat": { "exists": false } } TASK [linux-system-roles.storage : manage /etc/crypttab to account for changes we just made] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:176 Thursday 21 July 2022 15:24:01 +0000 (0:00:00.366) 0:01:29.695 ********* TASK [linux-system-roles.storage : Update facts] ******************************* task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:198 Thursday 21 July 2022 15:24:01 +0000 (0:00:00.022) 0:01:29.718 ********* ok: [/cache/rhel-8-y.qcow2.snap] META: role_complete for /cache/rhel-8-y.qcow2.snap TASK [include_tasks] *********************************************************** task path: /tmp/tmpafnwmbol/tests/tests_lvm_percent_size.yml:112 Thursday 21 July 2022 15:24:02 +0000 (0:00:00.974) 0:01:30.693 ********* included: /tmp/tmpafnwmbol/tests/verify-role-results.yml for /cache/rhel-8-y.qcow2.snap TASK [Print out pool information] ********************************************** task path: /tmp/tmpafnwmbol/tests/verify-role-results.yml:1 Thursday 21 July 2022 15:24:02 +0000 (0:00:00.047) 0:01:30.740 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "60%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test2", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test2", "_raw_device": "/dev/mapper/foo-test2", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "25%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] } TASK [Print out volume information] ******************************************** task path: /tmp/tmpafnwmbol/tests/verify-role-results.yml:6 Thursday 21 July 2022 15:24:02 +0000 (0:00:00.053) 0:01:30.794 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => {} TASK [Collect info about the volumes.] ***************************************** task path: /tmp/tmpafnwmbol/tests/verify-role-results.yml:14 Thursday 21 July 2022 15:24:02 +0000 (0:00:00.036) 0:01:30.830 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "info": { "/dev/mapper/foo-test1": { "fstype": "xfs", "label": "", "name": "/dev/mapper/foo-test1", "size": "6G", "type": "lvm", "uuid": "3db40ee5-790d-4909-8628-933255bd71de" }, "/dev/mapper/foo-test2": { "fstype": "ext4", "label": "", "name": "/dev/mapper/foo-test2", "size": "2.5G", "type": "lvm", "uuid": "ee7d1310-5176-4ec6-80b8-28181ffffed3" }, "/dev/nvme0n1": { "fstype": "", "label": "", "name": "/dev/nvme0n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme1n1": { "fstype": "", "label": "", "name": "/dev/nvme1n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme2n1": { "fstype": "", "label": "", "name": "/dev/nvme2n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sda": { "fstype": "LVM2_member", "label": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "irzZ3c-YIeb-eoxj-fRdj-JSvm-ls7D-iZNahz" }, "/dev/sdb": { "fstype": "", "label": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc": { "fstype": "", "label": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sr0": { "fstype": "iso9660", "label": "cidata", "name": "/dev/sr0", "size": "364K", "type": "rom", "uuid": "2022-07-21-15-22-18-00" }, "/dev/vda": { "fstype": "", "label": "", "name": "/dev/vda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vda1": { "fstype": "", "label": "", "name": "/dev/vda1", "size": "1M", "type": "partition", "uuid": "" }, "/dev/vda2": { "fstype": "vfat", "label": "", "name": "/dev/vda2", "size": "100M", "type": "partition", "uuid": "7B77-95E7" }, "/dev/vda3": { "fstype": "xfs", "label": "root", "name": "/dev/vda3", "size": "9.9G", "type": "partition", "uuid": "58dca2db-4ccd-4171-a017-dd53ad8c1c54" }, "/dev/vdb": { "fstype": "", "label": "", "name": "/dev/vdb", "size": "2G", "type": "disk", "uuid": "" }, "/dev/vdb1": { "fstype": "ext4", "label": "yumcache", "name": "/dev/vdb1", "size": "2G", "type": "partition", "uuid": "4d1ed9a1-5608-4ff8-ad89-21d95caa66fa" }, "/dev/vdc": { "fstype": "", "label": "", "name": "/dev/vdc", "size": "2G", "type": "disk", "uuid": "" }, "/dev/vdc1": { "fstype": "ext4", "label": "yumvarlib", "name": "/dev/vdc1", "size": "2G", "type": "partition", "uuid": "a7abcc97-edc5-476c-8ccd-694b4621fd28" }, "/dev/vdd": { "fstype": "", "label": "", "name": "/dev/vdd", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vde": { "fstype": "", "label": "", "name": "/dev/vde", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vdf": { "fstype": "", "label": "", "name": "/dev/vdf", "size": "10G", "type": "disk", "uuid": "" } } } TASK [Read the /etc/fstab file for volume existence] *************************** task path: /tmp/tmpafnwmbol/tests/verify-role-results.yml:19 Thursday 21 July 2022 15:24:02 +0000 (0:00:00.382) 0:01:31.213 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "cmd": [ "cat", "/etc/fstab" ], "delta": "0:00:00.002488", "end": "2022-07-21 11:24:02.976616", "rc": 0, "start": "2022-07-21 11:24:02.974128" } STDOUT: UUID=58dca2db-4ccd-4171-a017-dd53ad8c1c54 / xfs defaults 0 0 UUID=7B77-95E7 /boot/efi vfat defaults,uid=0,gid=0,umask=077,shortname=winnt 0 2 /dev/vdb1 /var/cache/dnf auto defaults,nofail,x-systemd.requires=cloud-init.service,_netdev,comment=cloudconfig 0 2 /dev/vdc1 /var/lib/dnf auto defaults,nofail,x-systemd.requires=cloud-init.service,_netdev,comment=cloudconfig 0 2 /dev/mapper/foo-test1 /opt/test1 xfs defaults 0 0 /dev/mapper/foo-test2 /opt/test2 ext4 defaults 0 0 TASK [Read the /etc/crypttab file] ********************************************* task path: /tmp/tmpafnwmbol/tests/verify-role-results.yml:24 Thursday 21 July 2022 15:24:03 +0000 (0:00:00.372) 0:01:31.585 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "cmd": [ "cat", "/etc/crypttab" ], "delta": "0:00:00.002651", "end": "2022-07-21 11:24:03.392332", "failed_when_result": false, "rc": 1, "start": "2022-07-21 11:24:03.389681" } STDERR: cat: /etc/crypttab: No such file or directory MSG: non-zero return code TASK [Verify the volumes listed in storage_pools were correctly managed] ******* task path: /tmp/tmpafnwmbol/tests/verify-role-results.yml:33 Thursday 21 July 2022 15:24:03 +0000 (0:00:00.414) 0:01:31.999 ********* included: /tmp/tmpafnwmbol/tests/test-verify-pool.yml for /cache/rhel-8-y.qcow2.snap => (item={'disks': ['sda'], 'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'name': 'foo', 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_metadata_version': None, 'state': 'present', 'type': 'lvm', 'volumes': [{'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'name': 'test1', 'raid_level': None, 'size': '60%', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test1', '_raw_device': '/dev/mapper/foo-test1', '_mount_id': '/dev/mapper/foo-test1', '_kernel_device': '/dev/dm-1', '_raw_kernel_device': '/dev/dm-1'}, {'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'ext4', 'mount_options': 'defaults', 'mount_point': '/opt/test2', 'name': 'test2', 'raid_level': None, 'size': '25%', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test2', '_raw_device': '/dev/mapper/foo-test2', '_mount_id': '/dev/mapper/foo-test2', '_kernel_device': '/dev/dm-0', '_raw_kernel_device': '/dev/dm-0'}], 'raid_chunk_size': None}) TASK [Set _storage_pool_tests] ************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-pool.yml:5 Thursday 21 July 2022 15:24:03 +0000 (0:00:00.103) 0:01:32.103 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_pool_tests": [ "members", "volumes" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool.yml:18 Thursday 21 July 2022 15:24:03 +0000 (0:00:00.082) 0:01:32.185 ********* included: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml for /cache/rhel-8-y.qcow2.snap => (item=members) included: /tmp/tmpafnwmbol/tests/test-verify-pool-volumes.yml for /cache/rhel-8-y.qcow2.snap => (item=volumes) TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:1 Thursday 21 July 2022 15:24:03 +0000 (0:00:00.091) 0:01:32.276 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_pv_count": "1", "_storage_test_pool_pvs_lvm": [ "/dev/sda" ] }, "changed": false } TASK [Get the canonical device path for each member device] ******************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:10 Thursday 21 July 2022 15:24:03 +0000 (0:00:00.062) 0:01:32.339 ********* ok: [/cache/rhel-8-y.qcow2.snap] => (item=/dev/sda) => { "ansible_loop_var": "pv", "changed": false, "device": "/dev/sda", "pv": "/dev/sda" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:19 Thursday 21 July 2022 15:24:04 +0000 (0:00:00.374) 0:01:32.714 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "__pvs_lvm_len": "1" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:23 Thursday 21 July 2022 15:24:04 +0000 (0:00:00.052) 0:01:32.767 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_pool_pvs": [ "/dev/sda" ] }, "changed": false } TASK [Verify PV count] ********************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:27 Thursday 21 July 2022 15:24:04 +0000 (0:00:00.054) 0:01:32.821 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:34 Thursday 21 July 2022 15:24:04 +0000 (0:00:00.052) 0:01:32.874 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:38 Thursday 21 July 2022 15:24:04 +0000 (0:00:00.037) 0:01:32.912 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:42 Thursday 21 July 2022 15:24:04 +0000 (0:00:00.049) 0:01:32.961 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check the type of each PV] *********************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:46 Thursday 21 July 2022 15:24:04 +0000 (0:00:00.024) 0:01:32.986 ********* ok: [/cache/rhel-8-y.qcow2.snap] => (item=/dev/sda) => { "ansible_loop_var": "pv", "changed": false, "pv": "/dev/sda" } MSG: All assertions passed TASK [Check MD RAID] *********************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:56 Thursday 21 July 2022 15:24:04 +0000 (0:00:00.040) 0:01:33.026 ********* included: /tmp/tmpafnwmbol/tests/verify-pool-md.yml for /cache/rhel-8-y.qcow2.snap TASK [get information about RAID] ********************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-md.yml:6 Thursday 21 July 2022 15:24:04 +0000 (0:00:00.041) 0:01:33.068 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-md.yml:12 Thursday 21 July 2022 15:24:04 +0000 (0:00:00.024) 0:01:33.092 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-md.yml:16 Thursday 21 July 2022 15:24:04 +0000 (0:00:00.023) 0:01:33.116 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-md.yml:20 Thursday 21 July 2022 15:24:04 +0000 (0:00:00.031) 0:01:33.148 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-md.yml:24 Thursday 21 July 2022 15:24:04 +0000 (0:00:00.029) 0:01:33.177 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-md.yml:30 Thursday 21 July 2022 15:24:04 +0000 (0:00:00.028) 0:01:33.206 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmpafnwmbol/tests/verify-pool-md.yml:36 Thursday 21 July 2022 15:24:04 +0000 (0:00:00.028) 0:01:33.235 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-md.yml:44 Thursday 21 July 2022 15:24:04 +0000 (0:00:00.026) 0:01:33.261 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_md_active_devices_re": null, "storage_test_md_metadata_version_re": null, "storage_test_md_spare_devices_re": null }, "changed": false } TASK [Check LVM RAID] ********************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:59 Thursday 21 July 2022 15:24:04 +0000 (0:00:00.044) 0:01:33.305 ********* included: /tmp/tmpafnwmbol/tests/verify-pool-members-lvmraid.yml for /cache/rhel-8-y.qcow2.snap TASK [Validate pool member LVM RAID settings] ********************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-members-lvmraid.yml:1 Thursday 21 July 2022 15:24:04 +0000 (0:00:00.048) 0:01:33.353 ********* included: /tmp/tmpafnwmbol/tests/verify-pool-member-lvmraid.yml for /cache/rhel-8-y.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'name': 'test1', 'raid_level': None, 'size': '60%', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test1', '_raw_device': '/dev/mapper/foo-test1', '_mount_id': '/dev/mapper/foo-test1', '_kernel_device': '/dev/dm-1', '_raw_kernel_device': '/dev/dm-1'}) included: /tmp/tmpafnwmbol/tests/verify-pool-member-lvmraid.yml for /cache/rhel-8-y.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'ext4', 'mount_options': 'defaults', 'mount_point': '/opt/test2', 'name': 'test2', 'raid_level': None, 'size': '25%', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test2', '_raw_device': '/dev/mapper/foo-test2', '_mount_id': '/dev/mapper/foo-test2', '_kernel_device': '/dev/dm-0', '_raw_kernel_device': '/dev/dm-0'}) TASK [Get information about LVM RAID] ****************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-lvmraid.yml:3 Thursday 21 July 2022 15:24:04 +0000 (0:00:00.052) 0:01:33.406 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is LVM RAID] ******************************************* task path: /tmp/tmpafnwmbol/tests/verify-pool-member-lvmraid.yml:8 Thursday 21 July 2022 15:24:04 +0000 (0:00:00.030) 0:01:33.436 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-lvmraid.yml:12 Thursday 21 July 2022 15:24:04 +0000 (0:00:00.026) 0:01:33.462 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about LVM RAID] ****************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-lvmraid.yml:3 Thursday 21 July 2022 15:24:04 +0000 (0:00:00.025) 0:01:33.488 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is LVM RAID] ******************************************* task path: /tmp/tmpafnwmbol/tests/verify-pool-member-lvmraid.yml:8 Thursday 21 July 2022 15:24:04 +0000 (0:00:00.027) 0:01:33.515 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-lvmraid.yml:12 Thursday 21 July 2022 15:24:05 +0000 (0:00:00.029) 0:01:33.544 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check Thin Pools] ******************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:62 Thursday 21 July 2022 15:24:05 +0000 (0:00:00.025) 0:01:33.570 ********* included: /tmp/tmpafnwmbol/tests/verify-pool-members-thin.yml for /cache/rhel-8-y.qcow2.snap TASK [Validate pool member thinpool settings] ********************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-members-thin.yml:1 Thursday 21 July 2022 15:24:05 +0000 (0:00:00.044) 0:01:33.615 ********* included: /tmp/tmpafnwmbol/tests/verify-pool-member-thin.yml for /cache/rhel-8-y.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'name': 'test1', 'raid_level': None, 'size': '60%', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test1', '_raw_device': '/dev/mapper/foo-test1', '_mount_id': '/dev/mapper/foo-test1', '_kernel_device': '/dev/dm-1', '_raw_kernel_device': '/dev/dm-1'}) included: /tmp/tmpafnwmbol/tests/verify-pool-member-thin.yml for /cache/rhel-8-y.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'ext4', 'mount_options': 'defaults', 'mount_point': '/opt/test2', 'name': 'test2', 'raid_level': None, 'size': '25%', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test2', '_raw_device': '/dev/mapper/foo-test2', '_mount_id': '/dev/mapper/foo-test2', '_kernel_device': '/dev/dm-0', '_raw_kernel_device': '/dev/dm-0'}) TASK [Get information about thinpool] ****************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-thin.yml:3 Thursday 21 July 2022 15:24:05 +0000 (0:00:00.050) 0:01:33.666 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-thin.yml:8 Thursday 21 July 2022 15:24:05 +0000 (0:00:00.054) 0:01:33.720 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-thin.yml:13 Thursday 21 July 2022 15:24:05 +0000 (0:00:00.025) 0:01:33.746 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-thin.yml:17 Thursday 21 July 2022 15:24:05 +0000 (0:00:00.025) 0:01:33.771 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about thinpool] ****************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-thin.yml:3 Thursday 21 July 2022 15:24:05 +0000 (0:00:00.024) 0:01:33.796 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-thin.yml:8 Thursday 21 July 2022 15:24:05 +0000 (0:00:00.026) 0:01:33.822 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-thin.yml:13 Thursday 21 July 2022 15:24:05 +0000 (0:00:00.024) 0:01:33.847 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-thin.yml:17 Thursday 21 July 2022 15:24:05 +0000 (0:00:00.023) 0:01:33.871 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check member encryption] ************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:65 Thursday 21 July 2022 15:24:05 +0000 (0:00:00.024) 0:01:33.895 ********* included: /tmp/tmpafnwmbol/tests/verify-pool-members-encryption.yml for /cache/rhel-8-y.qcow2.snap TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-members-encryption.yml:4 Thursday 21 July 2022 15:24:05 +0000 (0:00:00.048) 0:01:33.944 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Validate pool member LUKS settings] ************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-members-encryption.yml:8 Thursday 21 July 2022 15:24:05 +0000 (0:00:00.053) 0:01:33.998 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => (item=/dev/sda) => { "_storage_test_pool_member_path": "/dev/sda", "ansible_loop_var": "_storage_test_pool_member_path", "changed": false, "skip_reason": "Conditional result was False" } TASK [Validate pool member crypttab entries] *********************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-members-encryption.yml:15 Thursday 21 July 2022 15:24:05 +0000 (0:00:00.028) 0:01:34.027 ********* included: /tmp/tmpafnwmbol/tests/verify-pool-member-crypttab.yml for /cache/rhel-8-y.qcow2.snap => (item=/dev/sda) TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-crypttab.yml:1 Thursday 21 July 2022 15:24:05 +0000 (0:00:00.045) 0:01:34.072 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": [] }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpafnwmbol/tests/verify-pool-member-crypttab.yml:6 Thursday 21 July 2022 15:24:05 +0000 (0:00:00.050) 0:01:34.123 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpafnwmbol/tests/verify-pool-member-crypttab.yml:11 Thursday 21 July 2022 15:24:05 +0000 (0:00:00.052) 0:01:34.175 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-crypttab.yml:17 Thursday 21 July 2022 15:24:05 +0000 (0:00:00.036) 0:01:34.211 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-crypttab.yml:23 Thursday 21 July 2022 15:24:05 +0000 (0:00:00.036) 0:01:34.248 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-crypttab.yml:29 Thursday 21 July 2022 15:24:05 +0000 (0:00:00.037) 0:01:34.286 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-members-encryption.yml:22 Thursday 21 July 2022 15:24:05 +0000 (0:00:00.037) 0:01:34.324 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_crypttab_key_file": null }, "changed": false } TASK [Check VDO] *************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:68 Thursday 21 July 2022 15:24:05 +0000 (0:00:00.033) 0:01:34.358 ********* included: /tmp/tmpafnwmbol/tests/verify-pool-members-vdo.yml for /cache/rhel-8-y.qcow2.snap TASK [Validate pool member VDO settings] *************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-members-vdo.yml:1 Thursday 21 July 2022 15:24:05 +0000 (0:00:00.049) 0:01:34.407 ********* included: /tmp/tmpafnwmbol/tests/verify-pool-member-vdo.yml for /cache/rhel-8-y.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'name': 'test1', 'raid_level': None, 'size': '60%', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test1', '_raw_device': '/dev/mapper/foo-test1', '_mount_id': '/dev/mapper/foo-test1', '_kernel_device': '/dev/dm-1', '_raw_kernel_device': '/dev/dm-1'}) included: /tmp/tmpafnwmbol/tests/verify-pool-member-vdo.yml for /cache/rhel-8-y.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'ext4', 'mount_options': 'defaults', 'mount_point': '/opt/test2', 'name': 'test2', 'raid_level': None, 'size': '25%', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test2', '_raw_device': '/dev/mapper/foo-test2', '_mount_id': '/dev/mapper/foo-test2', '_kernel_device': '/dev/dm-0', '_raw_kernel_device': '/dev/dm-0'}) TASK [get information about VDO deduplication] ********************************* task path: /tmp/tmpafnwmbol/tests/verify-pool-member-vdo.yml:3 Thursday 21 July 2022 15:24:05 +0000 (0:00:00.057) 0:01:34.464 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-vdo.yml:8 Thursday 21 July 2022 15:24:05 +0000 (0:00:00.025) 0:01:34.489 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-vdo.yml:11 Thursday 21 July 2022 15:24:05 +0000 (0:00:00.025) 0:01:34.514 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-vdo.yml:16 Thursday 21 July 2022 15:24:06 +0000 (0:00:00.023) 0:01:34.538 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-vdo.yml:21 Thursday 21 July 2022 15:24:06 +0000 (0:00:00.025) 0:01:34.564 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-vdo.yml:24 Thursday 21 July 2022 15:24:06 +0000 (0:00:00.024) 0:01:34.589 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-vdo.yml:29 Thursday 21 July 2022 15:24:06 +0000 (0:00:00.024) 0:01:34.613 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-vdo.yml:39 Thursday 21 July 2022 15:24:06 +0000 (0:00:00.022) 0:01:34.636 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [get information about VDO deduplication] ********************************* task path: /tmp/tmpafnwmbol/tests/verify-pool-member-vdo.yml:3 Thursday 21 July 2022 15:24:06 +0000 (0:00:00.033) 0:01:34.670 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-vdo.yml:8 Thursday 21 July 2022 15:24:06 +0000 (0:00:00.028) 0:01:34.698 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-vdo.yml:11 Thursday 21 July 2022 15:24:06 +0000 (0:00:00.026) 0:01:34.724 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-vdo.yml:16 Thursday 21 July 2022 15:24:06 +0000 (0:00:00.027) 0:01:34.752 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-vdo.yml:21 Thursday 21 July 2022 15:24:06 +0000 (0:00:00.029) 0:01:34.781 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-vdo.yml:24 Thursday 21 July 2022 15:24:06 +0000 (0:00:00.025) 0:01:34.807 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-vdo.yml:29 Thursday 21 July 2022 15:24:06 +0000 (0:00:00.025) 0:01:34.832 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-vdo.yml:39 Thursday 21 July 2022 15:24:06 +0000 (0:00:00.026) 0:01:34.859 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:71 Thursday 21 July 2022 15:24:06 +0000 (0:00:00.092) 0:01:34.951 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "__pvs_lvm_len": null, "_storage_test_expected_pv_count": null, "_storage_test_expected_pv_type": null, "_storage_test_pool_pvs": [], "_storage_test_pool_pvs_lvm": [] }, "changed": false } TASK [verify the volumes] ****************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool-volumes.yml:3 Thursday 21 July 2022 15:24:06 +0000 (0:00:00.121) 0:01:35.072 ********* included: /tmp/tmpafnwmbol/tests/test-verify-volume.yml for /cache/rhel-8-y.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'name': 'test1', 'raid_level': None, 'size': '60%', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test1', '_raw_device': '/dev/mapper/foo-test1', '_mount_id': '/dev/mapper/foo-test1', '_kernel_device': '/dev/dm-1', '_raw_kernel_device': '/dev/dm-1'}) included: /tmp/tmpafnwmbol/tests/test-verify-volume.yml for /cache/rhel-8-y.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'ext4', 'mount_options': 'defaults', 'mount_point': '/opt/test2', 'name': 'test2', 'raid_level': None, 'size': '25%', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test2', '_raw_device': '/dev/mapper/foo-test2', '_mount_id': '/dev/mapper/foo-test2', '_kernel_device': '/dev/dm-0', '_raw_kernel_device': '/dev/dm-0'}) TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume.yml:2 Thursday 21 July 2022 15:24:06 +0000 (0:00:00.050) 0:01:35.123 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_volume_present": true, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume.yml:10 Thursday 21 July 2022 15:24:06 +0000 (0:00:00.051) 0:01:35.175 ********* included: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml for /cache/rhel-8-y.qcow2.snap => (item=mount) included: /tmp/tmpafnwmbol/tests/test-verify-volume-fstab.yml for /cache/rhel-8-y.qcow2.snap => (item=fstab) included: /tmp/tmpafnwmbol/tests/test-verify-volume-fs.yml for /cache/rhel-8-y.qcow2.snap => (item=fs) included: /tmp/tmpafnwmbol/tests/test-verify-volume-device.yml for /cache/rhel-8-y.qcow2.snap => (item=device) included: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml for /cache/rhel-8-y.qcow2.snap => (item=encryption) included: /tmp/tmpafnwmbol/tests/test-verify-volume-md.yml for /cache/rhel-8-y.qcow2.snap => (item=md) included: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml for /cache/rhel-8-y.qcow2.snap => (item=size) included: /tmp/tmpafnwmbol/tests/test-verify-volume-cache.yml for /cache/rhel-8-y.qcow2.snap => (item=cache) TASK [Get expected mount device based on device type] ************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml:6 Thursday 21 July 2022 15:24:06 +0000 (0:00:00.076) 0:01:35.252 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/foo-test1" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml:14 Thursday 21 July 2022 15:24:06 +0000 (0:00:00.043) 0:01:35.295 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_mount_device_matches": [ { "block_available": 1551090, "block_size": 4096, "block_total": 1570304, "block_used": 19214, "device": "/dev/mapper/foo-test1", "fstype": "xfs", "inode_available": 3145725, "inode_total": 3145728, "inode_used": 3, "mount": "/opt/test1", "options": "rw,seclabel,relatime,attr2,inode64,logbufs=8,logbsize=32k,noquota", "size_available": 6353264640, "size_total": 6431965184, "uuid": "3db40ee5-790d-4909-8628-933255bd71de" } ], "storage_test_mount_expected_match_count": "1", "storage_test_mount_point_matches": [ { "block_available": 1551090, "block_size": 4096, "block_total": 1570304, "block_used": 19214, "device": "/dev/mapper/foo-test1", "fstype": "xfs", "inode_available": 3145725, "inode_total": 3145728, "inode_used": 3, "mount": "/opt/test1", "options": "rw,seclabel,relatime,attr2,inode64,logbufs=8,logbsize=32k,noquota", "size_available": 6353264640, "size_total": 6431965184, "uuid": "3db40ee5-790d-4909-8628-933255bd71de" } ], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Verify the current mount state by device] ******************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml:28 Thursday 21 July 2022 15:24:06 +0000 (0:00:00.056) 0:01:35.351 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify the current mount state by mount point] *************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml:37 Thursday 21 July 2022 15:24:06 +0000 (0:00:00.052) 0:01:35.404 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify the mount fs type] ************************************************ task path: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml:45 Thursday 21 July 2022 15:24:06 +0000 (0:00:00.047) 0:01:35.451 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [command] ***************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml:54 Thursday 21 July 2022 15:24:06 +0000 (0:00:00.053) 0:01:35.505 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml:58 Thursday 21 July 2022 15:24:07 +0000 (0:00:00.024) 0:01:35.530 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml:63 Thursday 21 July 2022 15:24:07 +0000 (0:00:00.025) 0:01:35.555 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml:75 Thursday 21 July 2022 15:24:07 +0000 (0:00:00.025) 0:01:35.580 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-fstab.yml:2 Thursday 21 July 2022 15:24:07 +0000 (0:00:00.038) 0:01:35.619 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "1", "storage_test_fstab_expected_mount_options_matches": "1", "storage_test_fstab_expected_mount_point_matches": "1", "storage_test_fstab_id_matches": [ "/dev/mapper/foo-test1 " ], "storage_test_fstab_mount_options_matches": [ " /opt/test1 xfs defaults " ], "storage_test_fstab_mount_point_matches": [ " /opt/test1 " ] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-fstab.yml:25 Thursday 21 July 2022 15:24:07 +0000 (0:00:00.064) 0:01:35.684 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify the fstab mount point] ******************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-fstab.yml:32 Thursday 21 July 2022 15:24:07 +0000 (0:00:00.047) 0:01:35.731 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-fstab.yml:39 Thursday 21 July 2022 15:24:07 +0000 (0:00:00.047) 0:01:35.778 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-fstab.yml:49 Thursday 21 July 2022 15:24:07 +0000 (0:00:00.034) 0:01:35.812 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-fs.yml:4 Thursday 21 July 2022 15:24:07 +0000 (0:00:00.039) 0:01:35.852 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify fs label] ********************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-fs.yml:10 Thursday 21 July 2022 15:24:07 +0000 (0:00:00.039) 0:01:35.892 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [See whether the device node is present] ********************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-device.yml:4 Thursday 21 July 2022 15:24:07 +0000 (0:00:00.044) 0:01:35.936 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "stat": { "atime": 1658416980.2722602, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1658416980.2722602, "dev": 6, "device_type": 64769, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 62659, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/symlink", "mode": "0660", "mtime": 1658416980.2722602, "nlink": 1, "path": "/dev/mapper/foo-test1", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-device.yml:10 Thursday 21 July 2022 15:24:07 +0000 (0:00:00.381) 0:01:36.317 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about this volume] ********************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-device.yml:18 Thursday 21 July 2022 15:24:07 +0000 (0:00:00.040) 0:01:36.358 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [(1/2) Process volume type (set initial value)] *************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-device.yml:24 Thursday 21 July 2022 15:24:07 +0000 (0:00:00.041) 0:01:36.400 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [(2/2) Process volume type (get RAID value)] ****************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-device.yml:28 Thursday 21 July 2022 15:24:07 +0000 (0:00:00.038) 0:01:36.438 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-device.yml:33 Thursday 21 July 2022 15:24:07 +0000 (0:00:00.025) 0:01:36.463 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Stat the LUKS device, if encrypted] ************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:3 Thursday 21 July 2022 15:24:07 +0000 (0:00:00.035) 0:01:36.499 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:10 Thursday 21 July 2022 15:24:08 +0000 (0:00:00.022) 0:01:36.521 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [Collect LUKS info for this volume] *************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:15 Thursday 21 July 2022 15:24:09 +0000 (0:00:01.324) 0:01:37.846 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:21 Thursday 21 July 2022 15:24:09 +0000 (0:00:00.026) 0:01:37.872 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:30 Thursday 21 July 2022 15:24:09 +0000 (0:00:00.023) 0:01:37.896 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:38 Thursday 21 July 2022 15:24:09 +0000 (0:00:00.049) 0:01:37.946 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:44 Thursday 21 July 2022 15:24:09 +0000 (0:00:00.024) 0:01:37.970 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:49 Thursday 21 July 2022 15:24:09 +0000 (0:00:00.022) 0:01:37.992 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:55 Thursday 21 July 2022 15:24:09 +0000 (0:00:00.023) 0:01:38.016 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:61 Thursday 21 July 2022 15:24:09 +0000 (0:00:00.022) 0:01:38.038 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:67 Thursday 21 July 2022 15:24:09 +0000 (0:00:00.023) 0:01:38.062 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:74 Thursday 21 July 2022 15:24:09 +0000 (0:00:00.047) 0:01:38.109 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:79 Thursday 21 July 2022 15:24:09 +0000 (0:00:00.050) 0:01:38.160 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:85 Thursday 21 July 2022 15:24:09 +0000 (0:00:00.038) 0:01:38.198 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:91 Thursday 21 July 2022 15:24:09 +0000 (0:00:00.039) 0:01:38.238 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:97 Thursday 21 July 2022 15:24:09 +0000 (0:00:00.034) 0:01:38.272 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [get information about RAID] ********************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-md.yml:7 Thursday 21 July 2022 15:24:09 +0000 (0:00:00.033) 0:01:38.306 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-md.yml:13 Thursday 21 July 2022 15:24:09 +0000 (0:00:00.034) 0:01:38.340 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-md.yml:17 Thursday 21 July 2022 15:24:09 +0000 (0:00:00.040) 0:01:38.380 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-md.yml:21 Thursday 21 July 2022 15:24:09 +0000 (0:00:00.039) 0:01:38.420 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-md.yml:25 Thursday 21 July 2022 15:24:09 +0000 (0:00:00.036) 0:01:38.456 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-md.yml:31 Thursday 21 July 2022 15:24:09 +0000 (0:00:00.035) 0:01:38.492 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-md.yml:37 Thursday 21 July 2022 15:24:10 +0000 (0:00:00.040) 0:01:38.532 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the actual size of the volume] ************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:3 Thursday 21 July 2022 15:24:10 +0000 (0:00:00.039) 0:01:38.572 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "bytes": 6442450944, "changed": false, "lvm": "6g", "parted": "6GiB", "size": "6 GiB" } TASK [parse the requested size of the volume] ********************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:9 Thursday 21 July 2022 15:24:10 +0000 (0:00:00.385) 0:01:38.957 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Establish base value for expected size] ********************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:15 Thursday 21 July 2022 15:24:10 +0000 (0:00:00.041) 0:01:38.999 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:20 Thursday 21 July 2022 15:24:10 +0000 (0:00:00.039) 0:01:39.038 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_expected_size": "4294967296.0" } TASK [debug] ******************************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:25 Thursday 21 July 2022 15:24:10 +0000 (0:00:00.037) 0:01:39.075 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_pool": { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "60%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test2", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test2", "_raw_device": "/dev/mapper/foo-test2", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "25%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } } TASK [debug] ******************************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:28 Thursday 21 July 2022 15:24:10 +0000 (0:00:00.053) 0:01:39.129 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_blkinfo": { "changed": false, "failed": false, "info": { "/dev/mapper/foo-test1": { "fstype": "xfs", "label": "", "name": "/dev/mapper/foo-test1", "size": "6G", "type": "lvm", "uuid": "3db40ee5-790d-4909-8628-933255bd71de" }, "/dev/mapper/foo-test2": { "fstype": "ext4", "label": "", "name": "/dev/mapper/foo-test2", "size": "2.5G", "type": "lvm", "uuid": "ee7d1310-5176-4ec6-80b8-28181ffffed3" }, "/dev/nvme0n1": { "fstype": "", "label": "", "name": "/dev/nvme0n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme1n1": { "fstype": "", "label": "", "name": "/dev/nvme1n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme2n1": { "fstype": "", "label": "", "name": "/dev/nvme2n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sda": { "fstype": "LVM2_member", "label": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "irzZ3c-YIeb-eoxj-fRdj-JSvm-ls7D-iZNahz" }, "/dev/sdb": { "fstype": "", "label": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc": { "fstype": "", "label": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sr0": { "fstype": "iso9660", "label": "cidata", "name": "/dev/sr0", "size": "364K", "type": "rom", "uuid": "2022-07-21-15-22-18-00" }, "/dev/vda": { "fstype": "", "label": "", "name": "/dev/vda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vda1": { "fstype": "", "label": "", "name": "/dev/vda1", "size": "1M", "type": "partition", "uuid": "" }, "/dev/vda2": { "fstype": "vfat", "label": "", "name": "/dev/vda2", "size": "100M", "type": "partition", "uuid": "7B77-95E7" }, "/dev/vda3": { "fstype": "xfs", "label": "root", "name": "/dev/vda3", "size": "9.9G", "type": "partition", "uuid": "58dca2db-4ccd-4171-a017-dd53ad8c1c54" }, "/dev/vdb": { "fstype": "", "label": "", "name": "/dev/vdb", "size": "2G", "type": "disk", "uuid": "" }, "/dev/vdb1": { "fstype": "ext4", "label": "yumcache", "name": "/dev/vdb1", "size": "2G", "type": "partition", "uuid": "4d1ed9a1-5608-4ff8-ad89-21d95caa66fa" }, "/dev/vdc": { "fstype": "", "label": "", "name": "/dev/vdc", "size": "2G", "type": "disk", "uuid": "" }, "/dev/vdc1": { "fstype": "ext4", "label": "yumvarlib", "name": "/dev/vdc1", "size": "2G", "type": "partition", "uuid": "a7abcc97-edc5-476c-8ccd-694b4621fd28" }, "/dev/vdd": { "fstype": "", "label": "", "name": "/dev/vdd", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vde": { "fstype": "", "label": "", "name": "/dev/vde", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vdf": { "fstype": "", "label": "", "name": "/dev/vdf", "size": "10G", "type": "disk", "uuid": "" } } } } TASK [Get the size of parent/pool device] ************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:31 Thursday 21 July 2022 15:24:10 +0000 (0:00:00.054) 0:01:39.184 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "bytes": 10737418240, "changed": false, "lvm": "10g", "parted": "10GiB", "size": "10 GiB" } TASK [debug] ******************************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:36 Thursday 21 July 2022 15:24:11 +0000 (0:00:00.425) 0:01:39.610 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_pool_size": { "bytes": 10737418240, "changed": false, "failed": false, "lvm": "10g", "parted": "10GiB", "size": "10 GiB" } } TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:39 Thursday 21 July 2022 15:24:11 +0000 (0:00:00.088) 0:01:39.699 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_expected_size": "6442450944.0" }, "changed": false } TASK [debug] ******************************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:44 Thursday 21 July 2022 15:24:11 +0000 (0:00:00.089) 0:01:39.788 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_actual_size": { "bytes": 6442450944, "changed": false, "failed": false, "lvm": "6g", "parted": "6GiB", "size": "6 GiB" } } TASK [debug] ******************************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:47 Thursday 21 July 2022 15:24:11 +0000 (0:00:00.075) 0:01:39.864 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_expected_size": "6442450944.0" } TASK [assert] ****************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:50 Thursday 21 July 2022 15:24:11 +0000 (0:00:00.090) 0:01:39.955 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Get information about the LV] ******************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-cache.yml:6 Thursday 21 July 2022 15:24:11 +0000 (0:00:00.051) 0:01:40.007 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "cmd": [ "lvs", "--noheadings", "--nameprefixes", "--units=b", "--nosuffix", "--unquoted", "-o", "name,attr,cache_total_blocks,chunk_size,segtype", "foo/test1" ], "delta": "0:00:00.037902", "end": "2022-07-21 11:24:11.809920", "rc": 0, "start": "2022-07-21 11:24:11.772018" } STDOUT: LVM2_LV_NAME=test1 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-cache.yml:14 Thursday 21 July 2022 15:24:11 +0000 (0:00:00.415) 0:01:40.423 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_lv_segtype": [ "linear" ] }, "changed": false } TASK [check segment type] ****************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-cache.yml:17 Thursday 21 July 2022 15:24:11 +0000 (0:00:00.068) 0:01:40.491 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-cache.yml:22 Thursday 21 July 2022 15:24:12 +0000 (0:00:00.053) 0:01:40.545 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the requested cache size] ****************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-cache.yml:26 Thursday 21 July 2022 15:24:12 +0000 (0:00:00.038) 0:01:40.583 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-cache.yml:32 Thursday 21 July 2022 15:24:12 +0000 (0:00:00.035) 0:01:40.619 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-cache.yml:36 Thursday 21 July 2022 15:24:12 +0000 (0:00:00.035) 0:01:40.655 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume.yml:16 Thursday 21 July 2022 15:24:12 +0000 (0:00:00.040) 0:01:40.695 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume.yml:2 Thursday 21 July 2022 15:24:12 +0000 (0:00:00.040) 0:01:40.736 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_volume_present": true, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume.yml:10 Thursday 21 July 2022 15:24:12 +0000 (0:00:00.050) 0:01:40.786 ********* included: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml for /cache/rhel-8-y.qcow2.snap => (item=mount) included: /tmp/tmpafnwmbol/tests/test-verify-volume-fstab.yml for /cache/rhel-8-y.qcow2.snap => (item=fstab) included: /tmp/tmpafnwmbol/tests/test-verify-volume-fs.yml for /cache/rhel-8-y.qcow2.snap => (item=fs) included: /tmp/tmpafnwmbol/tests/test-verify-volume-device.yml for /cache/rhel-8-y.qcow2.snap => (item=device) included: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml for /cache/rhel-8-y.qcow2.snap => (item=encryption) included: /tmp/tmpafnwmbol/tests/test-verify-volume-md.yml for /cache/rhel-8-y.qcow2.snap => (item=md) included: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml for /cache/rhel-8-y.qcow2.snap => (item=size) included: /tmp/tmpafnwmbol/tests/test-verify-volume-cache.yml for /cache/rhel-8-y.qcow2.snap => (item=cache) TASK [Get expected mount device based on device type] ************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml:6 Thursday 21 July 2022 15:24:12 +0000 (0:00:00.077) 0:01:40.864 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/foo-test2" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml:14 Thursday 21 July 2022 15:24:12 +0000 (0:00:00.046) 0:01:40.910 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_mount_device_matches": [ { "block_available": 588749, "block_size": 4096, "block_total": 625618, "block_used": 36869, "device": "/dev/mapper/foo-test2", "fstype": "ext4", "inode_available": 163829, "inode_total": 163840, "inode_used": 11, "mount": "/opt/test2", "options": "rw,seclabel,relatime", "size_available": 2411515904, "size_total": 2562531328, "uuid": "ee7d1310-5176-4ec6-80b8-28181ffffed3" } ], "storage_test_mount_expected_match_count": "1", "storage_test_mount_point_matches": [ { "block_available": 588749, "block_size": 4096, "block_total": 625618, "block_used": 36869, "device": "/dev/mapper/foo-test2", "fstype": "ext4", "inode_available": 163829, "inode_total": 163840, "inode_used": 11, "mount": "/opt/test2", "options": "rw,seclabel,relatime", "size_available": 2411515904, "size_total": 2562531328, "uuid": "ee7d1310-5176-4ec6-80b8-28181ffffed3" } ], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Verify the current mount state by device] ******************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml:28 Thursday 21 July 2022 15:24:12 +0000 (0:00:00.074) 0:01:40.984 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify the current mount state by mount point] *************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml:37 Thursday 21 July 2022 15:24:12 +0000 (0:00:00.052) 0:01:41.036 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify the mount fs type] ************************************************ task path: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml:45 Thursday 21 July 2022 15:24:12 +0000 (0:00:00.053) 0:01:41.090 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [command] ***************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml:54 Thursday 21 July 2022 15:24:12 +0000 (0:00:00.053) 0:01:41.144 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml:58 Thursday 21 July 2022 15:24:12 +0000 (0:00:00.026) 0:01:41.170 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml:63 Thursday 21 July 2022 15:24:12 +0000 (0:00:00.025) 0:01:41.196 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml:75 Thursday 21 July 2022 15:24:12 +0000 (0:00:00.025) 0:01:41.221 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-fstab.yml:2 Thursday 21 July 2022 15:24:12 +0000 (0:00:00.039) 0:01:41.261 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "1", "storage_test_fstab_expected_mount_options_matches": "1", "storage_test_fstab_expected_mount_point_matches": "1", "storage_test_fstab_id_matches": [ "/dev/mapper/foo-test2 " ], "storage_test_fstab_mount_options_matches": [ " /opt/test2 ext4 defaults " ], "storage_test_fstab_mount_point_matches": [ " /opt/test2 " ] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-fstab.yml:25 Thursday 21 July 2022 15:24:12 +0000 (0:00:00.064) 0:01:41.325 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify the fstab mount point] ******************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-fstab.yml:32 Thursday 21 July 2022 15:24:12 +0000 (0:00:00.052) 0:01:41.378 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-fstab.yml:39 Thursday 21 July 2022 15:24:12 +0000 (0:00:00.055) 0:01:41.433 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-fstab.yml:49 Thursday 21 July 2022 15:24:12 +0000 (0:00:00.039) 0:01:41.473 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-fs.yml:4 Thursday 21 July 2022 15:24:12 +0000 (0:00:00.038) 0:01:41.512 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify fs label] ********************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-fs.yml:10 Thursday 21 July 2022 15:24:13 +0000 (0:00:00.042) 0:01:41.555 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [See whether the device node is present] ********************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-device.yml:4 Thursday 21 July 2022 15:24:13 +0000 (0:00:00.045) 0:01:41.600 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "stat": { "atime": 1658417038.22426, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1658417038.22426, "dev": 6, "device_type": 64768, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 62525, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/symlink", "mode": "0660", "mtime": 1658417038.22426, "nlink": 1, "path": "/dev/mapper/foo-test2", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-device.yml:10 Thursday 21 July 2022 15:24:13 +0000 (0:00:00.402) 0:01:42.003 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about this volume] ********************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-device.yml:18 Thursday 21 July 2022 15:24:13 +0000 (0:00:00.037) 0:01:42.041 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [(1/2) Process volume type (set initial value)] *************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-device.yml:24 Thursday 21 July 2022 15:24:13 +0000 (0:00:00.037) 0:01:42.079 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [(2/2) Process volume type (get RAID value)] ****************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-device.yml:28 Thursday 21 July 2022 15:24:13 +0000 (0:00:00.032) 0:01:42.112 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-device.yml:33 Thursday 21 July 2022 15:24:13 +0000 (0:00:00.022) 0:01:42.134 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Stat the LUKS device, if encrypted] ************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:3 Thursday 21 July 2022 15:24:13 +0000 (0:00:00.036) 0:01:42.171 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:10 Thursday 21 July 2022 15:24:13 +0000 (0:00:00.023) 0:01:42.194 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [Collect LUKS info for this volume] *************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:15 Thursday 21 July 2022 15:24:14 +0000 (0:00:01.305) 0:01:43.500 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:21 Thursday 21 July 2022 15:24:15 +0000 (0:00:00.026) 0:01:43.526 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:30 Thursday 21 July 2022 15:24:15 +0000 (0:00:00.025) 0:01:43.552 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:38 Thursday 21 July 2022 15:24:15 +0000 (0:00:00.054) 0:01:43.607 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:44 Thursday 21 July 2022 15:24:15 +0000 (0:00:00.026) 0:01:43.633 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:49 Thursday 21 July 2022 15:24:15 +0000 (0:00:00.025) 0:01:43.659 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:55 Thursday 21 July 2022 15:24:15 +0000 (0:00:00.024) 0:01:43.683 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:61 Thursday 21 July 2022 15:24:15 +0000 (0:00:00.022) 0:01:43.706 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:67 Thursday 21 July 2022 15:24:15 +0000 (0:00:00.022) 0:01:43.728 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:74 Thursday 21 July 2022 15:24:15 +0000 (0:00:00.051) 0:01:43.779 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:79 Thursday 21 July 2022 15:24:15 +0000 (0:00:00.051) 0:01:43.831 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:85 Thursday 21 July 2022 15:24:15 +0000 (0:00:00.035) 0:01:43.866 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:91 Thursday 21 July 2022 15:24:15 +0000 (0:00:00.034) 0:01:43.901 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:97 Thursday 21 July 2022 15:24:15 +0000 (0:00:00.036) 0:01:43.937 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [get information about RAID] ********************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-md.yml:7 Thursday 21 July 2022 15:24:15 +0000 (0:00:00.036) 0:01:43.973 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-md.yml:13 Thursday 21 July 2022 15:24:15 +0000 (0:00:00.035) 0:01:44.008 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-md.yml:17 Thursday 21 July 2022 15:24:15 +0000 (0:00:00.034) 0:01:44.043 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-md.yml:21 Thursday 21 July 2022 15:24:15 +0000 (0:00:00.035) 0:01:44.079 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-md.yml:25 Thursday 21 July 2022 15:24:15 +0000 (0:00:00.038) 0:01:44.117 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-md.yml:31 Thursday 21 July 2022 15:24:15 +0000 (0:00:00.035) 0:01:44.152 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-md.yml:37 Thursday 21 July 2022 15:24:15 +0000 (0:00:00.034) 0:01:44.187 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the actual size of the volume] ************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:3 Thursday 21 July 2022 15:24:15 +0000 (0:00:00.035) 0:01:44.222 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "bytes": 2684354560, "changed": false, "lvm": "2g", "parted": "2GiB", "size": "2 GiB" } TASK [parse the requested size of the volume] ********************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:9 Thursday 21 July 2022 15:24:16 +0000 (0:00:00.368) 0:01:44.590 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Establish base value for expected size] ********************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:15 Thursday 21 July 2022 15:24:16 +0000 (0:00:00.038) 0:01:44.629 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:20 Thursday 21 July 2022 15:24:16 +0000 (0:00:00.039) 0:01:44.668 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_expected_size": "6442450944.0" } TASK [debug] ******************************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:25 Thursday 21 July 2022 15:24:16 +0000 (0:00:00.043) 0:01:44.712 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_pool": { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "60%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test2", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test2", "_raw_device": "/dev/mapper/foo-test2", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "25%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } } TASK [debug] ******************************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:28 Thursday 21 July 2022 15:24:16 +0000 (0:00:00.056) 0:01:44.768 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_blkinfo": { "changed": false, "failed": false, "info": { "/dev/mapper/foo-test1": { "fstype": "xfs", "label": "", "name": "/dev/mapper/foo-test1", "size": "6G", "type": "lvm", "uuid": "3db40ee5-790d-4909-8628-933255bd71de" }, "/dev/mapper/foo-test2": { "fstype": "ext4", "label": "", "name": "/dev/mapper/foo-test2", "size": "2.5G", "type": "lvm", "uuid": "ee7d1310-5176-4ec6-80b8-28181ffffed3" }, "/dev/nvme0n1": { "fstype": "", "label": "", "name": "/dev/nvme0n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme1n1": { "fstype": "", "label": "", "name": "/dev/nvme1n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme2n1": { "fstype": "", "label": "", "name": "/dev/nvme2n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sda": { "fstype": "LVM2_member", "label": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "irzZ3c-YIeb-eoxj-fRdj-JSvm-ls7D-iZNahz" }, "/dev/sdb": { "fstype": "", "label": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc": { "fstype": "", "label": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sr0": { "fstype": "iso9660", "label": "cidata", "name": "/dev/sr0", "size": "364K", "type": "rom", "uuid": "2022-07-21-15-22-18-00" }, "/dev/vda": { "fstype": "", "label": "", "name": "/dev/vda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vda1": { "fstype": "", "label": "", "name": "/dev/vda1", "size": "1M", "type": "partition", "uuid": "" }, "/dev/vda2": { "fstype": "vfat", "label": "", "name": "/dev/vda2", "size": "100M", "type": "partition", "uuid": "7B77-95E7" }, "/dev/vda3": { "fstype": "xfs", "label": "root", "name": "/dev/vda3", "size": "9.9G", "type": "partition", "uuid": "58dca2db-4ccd-4171-a017-dd53ad8c1c54" }, "/dev/vdb": { "fstype": "", "label": "", "name": "/dev/vdb", "size": "2G", "type": "disk", "uuid": "" }, "/dev/vdb1": { "fstype": "ext4", "label": "yumcache", "name": "/dev/vdb1", "size": "2G", "type": "partition", "uuid": "4d1ed9a1-5608-4ff8-ad89-21d95caa66fa" }, "/dev/vdc": { "fstype": "", "label": "", "name": "/dev/vdc", "size": "2G", "type": "disk", "uuid": "" }, "/dev/vdc1": { "fstype": "ext4", "label": "yumvarlib", "name": "/dev/vdc1", "size": "2G", "type": "partition", "uuid": "a7abcc97-edc5-476c-8ccd-694b4621fd28" }, "/dev/vdd": { "fstype": "", "label": "", "name": "/dev/vdd", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vde": { "fstype": "", "label": "", "name": "/dev/vde", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vdf": { "fstype": "", "label": "", "name": "/dev/vdf", "size": "10G", "type": "disk", "uuid": "" } } } } TASK [Get the size of parent/pool device] ************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:31 Thursday 21 July 2022 15:24:16 +0000 (0:00:00.103) 0:01:44.872 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "bytes": 10737418240, "changed": false, "lvm": "10g", "parted": "10GiB", "size": "10 GiB" } TASK [debug] ******************************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:36 Thursday 21 July 2022 15:24:16 +0000 (0:00:00.438) 0:01:45.311 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_pool_size": { "bytes": 10737418240, "changed": false, "failed": false, "lvm": "10g", "parted": "10GiB", "size": "10 GiB" } } TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:39 Thursday 21 July 2022 15:24:16 +0000 (0:00:00.053) 0:01:45.364 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_expected_size": "2684354560.0" }, "changed": false } TASK [debug] ******************************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:44 Thursday 21 July 2022 15:24:16 +0000 (0:00:00.054) 0:01:45.418 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_actual_size": { "bytes": 2684354560, "changed": false, "failed": false, "lvm": "2g", "parted": "2GiB", "size": "2 GiB" } } TASK [debug] ******************************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:47 Thursday 21 July 2022 15:24:16 +0000 (0:00:00.037) 0:01:45.456 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_expected_size": "2684354560.0" } TASK [assert] ****************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:50 Thursday 21 July 2022 15:24:16 +0000 (0:00:00.033) 0:01:45.490 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Get information about the LV] ******************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-cache.yml:6 Thursday 21 July 2022 15:24:17 +0000 (0:00:00.050) 0:01:45.540 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "cmd": [ "lvs", "--noheadings", "--nameprefixes", "--units=b", "--nosuffix", "--unquoted", "-o", "name,attr,cache_total_blocks,chunk_size,segtype", "foo/test2" ], "delta": "0:00:00.036539", "end": "2022-07-21 11:24:17.355710", "rc": 0, "start": "2022-07-21 11:24:17.319171" } STDOUT: LVM2_LV_NAME=test2 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-cache.yml:14 Thursday 21 July 2022 15:24:17 +0000 (0:00:00.429) 0:01:45.969 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_lv_segtype": [ "linear" ] }, "changed": false } TASK [check segment type] ****************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-cache.yml:17 Thursday 21 July 2022 15:24:17 +0000 (0:00:00.051) 0:01:46.021 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-cache.yml:22 Thursday 21 July 2022 15:24:17 +0000 (0:00:00.055) 0:01:46.076 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the requested cache size] ****************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-cache.yml:26 Thursday 21 July 2022 15:24:17 +0000 (0:00:00.042) 0:01:46.119 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-cache.yml:32 Thursday 21 July 2022 15:24:17 +0000 (0:00:00.039) 0:01:46.159 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-cache.yml:36 Thursday 21 July 2022 15:24:17 +0000 (0:00:00.044) 0:01:46.204 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume.yml:16 Thursday 21 July 2022 15:24:17 +0000 (0:00:00.037) 0:01:46.241 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [Verify the volumes with no pool were correctly managed] ****************** task path: /tmp/tmpafnwmbol/tests/verify-role-results.yml:43 Thursday 21 July 2022 15:24:17 +0000 (0:00:00.035) 0:01:46.277 ********* TASK [Clean up variable namespace] ********************************************* task path: /tmp/tmpafnwmbol/tests/verify-role-results.yml:53 Thursday 21 July 2022 15:24:17 +0000 (0:00:00.020) 0:01:46.298 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_blkinfo": null, "storage_test_crypttab": null, "storage_test_fstab": null }, "changed": false } TASK [Get the size of test2 volume] ******************************************** task path: /tmp/tmpafnwmbol/tests/tests_lvm_percent_size.yml:114 Thursday 21 July 2022 15:24:17 +0000 (0:00:00.036) 0:01:46.334 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "cmd": [ "lsblk", "--noheadings", "-o", "SIZE", "/dev/mapper/foo-test2" ], "delta": "0:00:00.006420", "end": "2022-07-21 11:24:18.115785", "rc": 0, "start": "2022-07-21 11:24:18.109365" } STDOUT: 2.5G TASK [Remove the test1 volume without changing its size] *********************** task path: /tmp/tmpafnwmbol/tests/tests_lvm_percent_size.yml:119 Thursday 21 July 2022 15:24:18 +0000 (0:00:00.394) 0:01:46.728 ********* TASK [linux-system-roles.storage : set platform/version specific variables] **** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main.yml:2 Thursday 21 July 2022 15:24:18 +0000 (0:00:00.060) 0:01:46.788 ********* included: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for /cache/rhel-8-y.qcow2.snap TASK [linux-system-roles.storage : Ensure ansible_facts used by role] ********** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2 Thursday 21 July 2022 15:24:18 +0000 (0:00:00.038) 0:01:46.827 ********* ok: [/cache/rhel-8-y.qcow2.snap] TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8 Thursday 21 July 2022 15:24:18 +0000 (0:00:00.540) 0:01:47.367 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [/cache/rhel-8-y.qcow2.snap] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [/cache/rhel-8-y.qcow2.snap] => (item=RedHat_8.yml) => { "ansible_facts": { "blivet_package_list": [ "python3-blivet", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap", "vdo", "kmod-kvdo", "xfsprogs" ] }, "ansible_included_var_files": [ "/tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/vars/RedHat_8.yml" ], "ansible_loop_var": "item", "changed": false, "item": "RedHat_8.yml" } skipping: [/cache/rhel-8-y.qcow2.snap] => (item=RedHat_8.7.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat_8.7.yml", "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : define an empty list of pools to be used in testing] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main.yml:5 Thursday 21 July 2022 15:24:18 +0000 (0:00:00.098) 0:01:47.465 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : define an empty list of volumes to be used in testing] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main.yml:9 Thursday 21 July 2022 15:24:18 +0000 (0:00:00.037) 0:01:47.503 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : include the appropriate provider tasks] ***** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main.yml:13 Thursday 21 July 2022 15:24:19 +0000 (0:00:00.034) 0:01:47.538 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount included: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for /cache/rhel-8-y.qcow2.snap TASK [linux-system-roles.storage : get a list of rpm packages installed on host machine] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2 Thursday 21 July 2022 15:24:19 +0000 (0:00:00.050) 0:01:47.588 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : make sure blivet is available] ************** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:7 Thursday 21 July 2022 15:24:19 +0000 (0:00:00.022) 0:01:47.611 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [linux-system-roles.storage : show storage_pools] ************************* task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:14 Thursday 21 July 2022 15:24:20 +0000 (0:00:01.431) 0:01:49.042 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_pools": [ { "disks": [ "sda" ], "name": "foo", "state": "present", "volumes": [ { "mount_point": "/opt/test1", "name": "test1", "size": "60%", "state": "absent" }, { "mount_point": "/opt/test2", "name": "test2", "size": "25%" } ] } ] } TASK [linux-system-roles.storage : show storage_volumes] *********************** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:19 Thursday 21 July 2022 15:24:20 +0000 (0:00:00.041) 0:01:49.083 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [linux-system-roles.storage : get required packages] ********************** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:24 Thursday 21 July 2022 15:24:20 +0000 (0:00:00.037) 0:01:49.120 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [ "lvm2" ], "pools": [], "volumes": [] } TASK [linux-system-roles.storage : enable copr repositories if needed] ********* task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:37 Thursday 21 July 2022 15:24:22 +0000 (0:00:01.607) 0:01:50.728 ********* included: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml for /cache/rhel-8-y.qcow2.snap TASK [linux-system-roles.storage : check if the COPR support packages should be installed] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:2 Thursday 21 July 2022 15:24:22 +0000 (0:00:00.048) 0:01:50.776 ********* TASK [linux-system-roles.storage : make sure COPR support packages are present] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:13 Thursday 21 July 2022 15:24:22 +0000 (0:00:00.039) 0:01:50.816 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : enable COPRs] ******************************* task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:18 Thursday 21 July 2022 15:24:22 +0000 (0:00:00.047) 0:01:50.864 ********* TASK [linux-system-roles.storage : make sure required packages are installed] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:44 Thursday 21 July 2022 15:24:22 +0000 (0:00:00.035) 0:01:50.899 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [linux-system-roles.storage : get service facts] ************************** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:51 Thursday 21 July 2022 15:24:23 +0000 (0:00:01.349) 0:01:52.249 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "services": { "NetworkManager-dispatcher.service": { "name": "NetworkManager-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "NetworkManager-wait-online.service": { "name": "NetworkManager-wait-online.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "NetworkManager.service": { "name": "NetworkManager.service", "source": "systemd", "state": "running", "status": "enabled" }, "arp-ethers.service": { "name": "arp-ethers.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "auditd.service": { "name": "auditd.service", "source": "systemd", "state": "running", "status": "enabled" }, "auth-rpcgss-module.service": { "name": "auth-rpcgss-module.service", "source": "systemd", "state": "stopped", "status": "static" }, "autovt@.service": { "name": "autovt@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "blivet.service": { "name": "blivet.service", "source": "systemd", "state": "inactive", "status": "static" }, "blk-availability.service": { "name": "blk-availability.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "chrony-dnssrv@.service": { "name": "chrony-dnssrv@.service", "source": "systemd", "state": "unknown", "status": "static" }, "chrony-wait.service": { "name": "chrony-wait.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd.service": { "name": "chronyd.service", "source": "systemd", "state": "running", "status": "enabled" }, "cloud-config.service": { "name": "cloud-config.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-final.service": { "name": "cloud-final.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init-hotplugd.service": { "name": "cloud-init-hotplugd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cloud-init-local.service": { "name": "cloud-init-local.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init.service": { "name": "cloud-init.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cockpit-motd.service": { "name": "cockpit-motd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cockpit-wsinstance-http.service": { "name": "cockpit-wsinstance-http.service", "source": "systemd", "state": "inactive", "status": "static" }, "cockpit-wsinstance-https-factory@.service": { "name": "cockpit-wsinstance-https-factory@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cockpit-wsinstance-https@.service": { "name": "cockpit-wsinstance-https@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cockpit.service": { "name": "cockpit.service", "source": "systemd", "state": "inactive", "status": "static" }, "console-getty.service": { "name": "console-getty.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "container-getty@.service": { "name": "container-getty@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cpupower.service": { "name": "cpupower.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "crond.service": { "name": "crond.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-org.freedesktop.hostname1.service": { "name": "dbus-org.freedesktop.hostname1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.locale1.service": { "name": "dbus-org.freedesktop.locale1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.login1.service": { "name": "dbus-org.freedesktop.login1.service", "source": "systemd", "state": "active", "status": "static" }, "dbus-org.freedesktop.nm-dispatcher.service": { "name": "dbus-org.freedesktop.nm-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "dbus-org.freedesktop.portable1.service": { "name": "dbus-org.freedesktop.portable1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.timedate1.service": { "name": "dbus-org.freedesktop.timedate1.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "dbus.service": { "name": "dbus.service", "source": "systemd", "state": "running", "status": "static" }, "dbxtool.service": { "name": "dbxtool.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "debug-shell.service": { "name": "debug-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "display-manager.service": { "name": "display-manager.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "dm-event.service": { "name": "dm-event.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-makecache.service": { "name": "dnf-makecache.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-cmdline.service": { "name": "dracut-cmdline.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-initqueue.service": { "name": "dracut-initqueue.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-mount.service": { "name": "dracut-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-mount.service": { "name": "dracut-pre-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-pivot.service": { "name": "dracut-pre-pivot.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-trigger.service": { "name": "dracut-pre-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-udev.service": { "name": "dracut-pre-udev.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown.service": { "name": "dracut-shutdown.service", "source": "systemd", "state": "stopped", "status": "static" }, "emergency.service": { "name": "emergency.service", "source": "systemd", "state": "stopped", "status": "static" }, "fcoe.service": { "name": "fcoe.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "fstrim.service": { "name": "fstrim.service", "source": "systemd", "state": "inactive", "status": "static" }, "getty@.service": { "name": "getty@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "getty@tty1.service": { "name": "getty@tty1.service", "source": "systemd", "state": "running", "status": "active" }, "grub-boot-indeterminate.service": { "name": "grub-boot-indeterminate.service", "source": "systemd", "state": "inactive", "status": "static" }, "gssproxy.service": { "name": "gssproxy.service", "source": "systemd", "state": "running", "status": "disabled" }, "halt-local.service": { "name": "halt-local.service", "source": "systemd", "state": "inactive", "status": "static" }, "hv_kvp_daemon.service": { "name": "hv_kvp_daemon.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "import-state.service": { "name": "import-state.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "initrd-cleanup.service": { "name": "initrd-cleanup.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-parse-etc.service": { "name": "initrd-parse-etc.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-switch-root.service": { "name": "initrd-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-udevadm-cleanup-db.service": { "name": "initrd-udevadm-cleanup-db.service", "source": "systemd", "state": "stopped", "status": "static" }, "insights-client-boot.service": { "name": "insights-client-boot.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "insights-client-results.service": { "name": "insights-client-results.service", "source": "systemd", "state": "inactive", "status": "static" }, "insights-client.service": { "name": "insights-client.service", "source": "systemd", "state": "inactive", "status": "static" }, "irqbalance.service": { "name": "irqbalance.service", "source": "systemd", "state": "running", "status": "enabled" }, "iscsi-shutdown.service": { "name": "iscsi-shutdown.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "iscsi.service": { "name": "iscsi.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "iscsid.service": { "name": "iscsid.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "kdump.service": { "name": "kdump.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "kmod-static-nodes.service": { "name": "kmod-static-nodes.service", "source": "systemd", "state": "stopped", "status": "static" }, "kvm_stat.service": { "name": "kvm_stat.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "ldconfig.service": { "name": "ldconfig.service", "source": "systemd", "state": "stopped", "status": "static" }, "loadmodules.service": { "name": "loadmodules.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "lvm2-activation-early.service": { "name": "lvm2-activation-early.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "lvm2-activation.service": { "name": "lvm2-activation.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "lvm2-lvmpolld.service": { "name": "lvm2-lvmpolld.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-monitor.service": { "name": "lvm2-monitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "lvm2-pvscan@.service": { "name": "lvm2-pvscan@.service", "source": "systemd", "state": "unknown", "status": "static" }, "lvm2-pvscan@8:0.service": { "name": "lvm2-pvscan@8:0.service", "source": "systemd", "state": "stopped", "status": "active" }, "man-db-cache-update.service": { "name": "man-db-cache-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "man-db-restart-cache-update.service": { "name": "man-db-restart-cache-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "mdadm-grow-continue@.service": { "name": "mdadm-grow-continue@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdadm-last-resort@.service": { "name": "mdadm-last-resort@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdcheck_continue.service": { "name": "mdcheck_continue.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdcheck_start.service": { "name": "mdcheck_start.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmon@.service": { "name": "mdmon@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdmonitor-oneshot.service": { "name": "mdmonitor-oneshot.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmonitor.service": { "name": "mdmonitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "messagebus.service": { "name": "messagebus.service", "source": "systemd", "state": "active", "status": "static" }, "microcode.service": { "name": "microcode.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "multipathd.service": { "name": "multipathd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "ndctl-monitor.service": { "name": "ndctl-monitor.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "network.service": { "name": "network.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "nfs-blkmap.service": { "name": "nfs-blkmap.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-convert.service": { "name": "nfs-convert.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-idmapd.service": { "name": "nfs-idmapd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-mountd.service": { "name": "nfs-mountd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-server.service": { "name": "nfs-server.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-utils.service": { "name": "nfs-utils.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfsdcld.service": { "name": "nfsdcld.service", "source": "systemd", "state": "stopped", "status": "static" }, "nis-domainname.service": { "name": "nis-domainname.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "ntpd.service": { "name": "ntpd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ntpdate.service": { "name": "ntpdate.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "oddjobd.service": { "name": "oddjobd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "packagekit-offline-update.service": { "name": "packagekit-offline-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "packagekit.service": { "name": "packagekit.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-quit-wait.service": { "name": "plymouth-quit-wait.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "plymouth-start.service": { "name": "plymouth-start.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "polkit.service": { "name": "polkit.service", "source": "systemd", "state": "running", "status": "static" }, "qemu-guest-agent.service": { "name": "qemu-guest-agent.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "quotaon.service": { "name": "quotaon.service", "source": "systemd", "state": "inactive", "status": "static" }, "rbdmap.service": { "name": "rbdmap.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "rc-local.service": { "name": "rc-local.service", "source": "systemd", "state": "stopped", "status": "static" }, "rdisc.service": { "name": "rdisc.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rescue.service": { "name": "rescue.service", "source": "systemd", "state": "stopped", "status": "static" }, "rhcd.service": { "name": "rhcd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhsm-facts.service": { "name": "rhsm-facts.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhsm.service": { "name": "rhsm.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhsmcertd.service": { "name": "rhsmcertd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpc-gssd.service": { "name": "rpc-gssd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd-notify.service": { "name": "rpc-statd-notify.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd.service": { "name": "rpc-statd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpcbind.service": { "name": "rpcbind.service", "source": "systemd", "state": "running", "status": "enabled" }, "rsyslog.service": { "name": "rsyslog.service", "source": "systemd", "state": "running", "status": "enabled" }, "selinux-autorelabel-mark.service": { "name": "selinux-autorelabel-mark.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "selinux-autorelabel.service": { "name": "selinux-autorelabel.service", "source": "systemd", "state": "inactive", "status": "static" }, "serial-getty@.service": { "name": "serial-getty@.service", "source": "systemd", "state": "unknown", "status": "indirect" }, "serial-getty@ttyS0.service": { "name": "serial-getty@ttyS0.service", "source": "systemd", "state": "running", "status": "active" }, "snapd.seeded.service": { "name": "snapd.seeded.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sntp.service": { "name": "sntp.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sshd-keygen.service": { "name": "sshd-keygen.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sshd-keygen@.service": { "name": "sshd-keygen@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@ecdsa.service": { "name": "sshd-keygen@ecdsa.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd-keygen@ed25519.service": { "name": "sshd-keygen@ed25519.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd-keygen@rsa.service": { "name": "sshd-keygen@rsa.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd.service": { "name": "sshd.service", "source": "systemd", "state": "running", "status": "enabled" }, "sshd@.service": { "name": "sshd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "sssd-autofs.service": { "name": "sssd-autofs.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-kcm.service": { "name": "sssd-kcm.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "sssd-nss.service": { "name": "sssd-nss.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pac.service": { "name": "sssd-pac.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pam.service": { "name": "sssd-pam.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-ssh.service": { "name": "sssd-ssh.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-sudo.service": { "name": "sssd-sudo.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd.service": { "name": "sssd.service", "source": "systemd", "state": "running", "status": "enabled" }, "syslog.service": { "name": "syslog.service", "source": "systemd", "state": "active", "status": "enabled" }, "system-update-cleanup.service": { "name": "system-update-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-ask-password-console.service": { "name": "systemd-ask-password-console.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-wall.service": { "name": "systemd-ask-password-wall.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-backlight@.service": { "name": "systemd-backlight@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-binfmt.service": { "name": "systemd-binfmt.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-coredump@.service": { "name": "systemd-coredump@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-exit.service": { "name": "systemd-exit.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-firstboot.service": { "name": "systemd-firstboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck-root.service": { "name": "systemd-fsck-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck@.service": { "name": "systemd-fsck@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-fsck@dev-disk-by\\x2duuid-7B77\\x2d95E7.service": { "name": "systemd-fsck@dev-disk-by\\x2duuid-7B77\\x2d95E7.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-fsck@dev-vdb1.service": { "name": "systemd-fsck@dev-vdb1.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-fsck@dev-vdc1.service": { "name": "systemd-fsck@dev-vdc1.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-halt.service": { "name": "systemd-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hibernate-resume@.service": { "name": "systemd-hibernate-resume@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-hibernate.service": { "name": "systemd-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hostnamed.service": { "name": "systemd-hostnamed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hwdb-update.service": { "name": "systemd-hwdb-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-hybrid-sleep.service": { "name": "systemd-hybrid-sleep.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-initctl.service": { "name": "systemd-initctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-catalog-update.service": { "name": "systemd-journal-catalog-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-flush.service": { "name": "systemd-journal-flush.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journald.service": { "name": "systemd-journald.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-kexec.service": { "name": "systemd-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-localed.service": { "name": "systemd-localed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-logind.service": { "name": "systemd-logind.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-machine-id-commit.service": { "name": "systemd-machine-id-commit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-modules-load.service": { "name": "systemd-modules-load.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-networkd-wait-online.service": { "name": "systemd-networkd-wait-online.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "systemd-portabled.service": { "name": "systemd-portabled.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-poweroff.service": { "name": "systemd-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-quotacheck.service": { "name": "systemd-quotacheck.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-random-seed.service": { "name": "systemd-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-reboot.service": { "name": "systemd-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-remount-fs.service": { "name": "systemd-remount-fs.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-resolved.service": { "name": "systemd-resolved.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-rfkill.service": { "name": "systemd-rfkill.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend-then-hibernate.service": { "name": "systemd-suspend-then-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend.service": { "name": "systemd-suspend.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-sysctl.service": { "name": "systemd-sysctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-sysusers.service": { "name": "systemd-sysusers.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-timedated.service": { "name": "systemd-timedated.service", "source": "systemd", "state": "inactive", "status": "masked" }, "systemd-timesyncd.service": { "name": "systemd-timesyncd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "systemd-tmpfiles-clean.service": { "name": "systemd-tmpfiles-clean.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev.service": { "name": "systemd-tmpfiles-setup-dev.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup.service": { "name": "systemd-tmpfiles-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-settle.service": { "name": "systemd-udev-settle.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-trigger.service": { "name": "systemd-udev-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udevd.service": { "name": "systemd-udevd.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-update-done.service": { "name": "systemd-update-done.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp-runlevel.service": { "name": "systemd-update-utmp-runlevel.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp.service": { "name": "systemd-update-utmp.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-user-sessions.service": { "name": "systemd-user-sessions.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-vconsole-setup.service": { "name": "systemd-vconsole-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-volatile-root.service": { "name": "systemd-volatile-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "tcsd.service": { "name": "tcsd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "teamd@.service": { "name": "teamd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "timedatex.service": { "name": "timedatex.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "tuned.service": { "name": "tuned.service", "source": "systemd", "state": "running", "status": "enabled" }, "unbound-anchor.service": { "name": "unbound-anchor.service", "source": "systemd", "state": "stopped", "status": "static" }, "user-runtime-dir@.service": { "name": "user-runtime-dir@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user-runtime-dir@0.service": { "name": "user-runtime-dir@0.service", "source": "systemd", "state": "stopped", "status": "active" }, "user@.service": { "name": "user@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user@0.service": { "name": "user@0.service", "source": "systemd", "state": "running", "status": "active" }, "vdo-start-by-dev@.service": { "name": "vdo-start-by-dev@.service", "source": "systemd", "state": "unknown", "status": "static" }, "vdo.service": { "name": "vdo.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "ypbind.service": { "name": "ypbind.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "yppasswdd.service": { "name": "yppasswdd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ypserv.service": { "name": "ypserv.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ypxfrd.service": { "name": "ypxfrd.service", "source": "systemd", "state": "stopped", "status": "not-found" } } }, "changed": false } TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************ task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:58 Thursday 21 July 2022 15:24:26 +0000 (0:00:02.745) 0:01:54.994 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] ******* task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:71 Thursday 21 July 2022 15:24:26 +0000 (0:00:00.104) 0:01:55.099 ********* TASK [linux-system-roles.storage : manage the pools and volumes to match the specified state] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77 Thursday 21 July 2022 15:24:26 +0000 (0:00:00.021) 0:01:55.121 ********* changed: [/cache/rhel-8-y.qcow2.snap] => { "actions": [ { "action": "destroy format", "device": "/dev/mapper/foo-test1", "fs_type": "xfs" }, { "action": "destroy device", "device": "/dev/mapper/foo-test1", "fs_type": null } ], "changed": true, "crypts": [], "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/vda2", "/dev/vda3", "/dev/mapper/foo-test2", "/dev/sdb", "/dev/sdc", "/dev/nvme0n1", "/dev/nvme1n1", "/dev/nvme2n1", "/dev/vdb1", "/dev/vdc1", "/dev/vdd", "/dev/vde", "/dev/vdf" ], "mounts": [ { "fstype": "xfs", "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "absent" }, { "dump": 0, "fstype": "ext4", "opts": "defaults", "passno": 0, "path": "/opt/test2", "src": "/dev/mapper/foo-test2", "state": "mounted" } ], "packages": [ "xfsprogs", "dosfstools", "e2fsprogs", "lvm2" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "60%", "state": "absent", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test2", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test2", "_raw_device": "/dev/mapper/foo-test2", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "25%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } TASK [linux-system-roles.storage : Workaround for udev issue on some platforms] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:91 Thursday 21 July 2022 15:24:28 +0000 (0:00:02.293) 0:01:57.415 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] ***** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:103 Thursday 21 July 2022 15:24:28 +0000 (0:00:00.039) 0:01:57.454 ********* TASK [linux-system-roles.storage : show blivet_output] ************************* task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:109 Thursday 21 July 2022 15:24:28 +0000 (0:00:00.052) 0:01:57.507 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "blivet_output": { "actions": [ { "action": "destroy format", "device": "/dev/mapper/foo-test1", "fs_type": "xfs" }, { "action": "destroy device", "device": "/dev/mapper/foo-test1", "fs_type": null } ], "changed": true, "crypts": [], "failed": false, "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/vda2", "/dev/vda3", "/dev/mapper/foo-test2", "/dev/sdb", "/dev/sdc", "/dev/nvme0n1", "/dev/nvme1n1", "/dev/nvme2n1", "/dev/vdb1", "/dev/vdc1", "/dev/vdd", "/dev/vde", "/dev/vdf" ], "mounts": [ { "fstype": "xfs", "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "absent" }, { "dump": 0, "fstype": "ext4", "opts": "defaults", "passno": 0, "path": "/opt/test2", "src": "/dev/mapper/foo-test2", "state": "mounted" } ], "packages": [ "xfsprogs", "dosfstools", "e2fsprogs", "lvm2" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "60%", "state": "absent", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test2", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test2", "_raw_device": "/dev/mapper/foo-test2", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "25%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } } TASK [linux-system-roles.storage : set the list of pools for test verification] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:114 Thursday 21 July 2022 15:24:29 +0000 (0:00:00.040) 0:01:57.547 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "60%", "state": "absent", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test2", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test2", "_raw_device": "/dev/mapper/foo-test2", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "25%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] }, "changed": false } TASK [linux-system-roles.storage : set the list of volumes for test verification] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:118 Thursday 21 July 2022 15:24:29 +0000 (0:00:00.040) 0:01:57.587 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : remove obsolete mounts] ********************* task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:134 Thursday 21 July 2022 15:24:29 +0000 (0:00:00.041) 0:01:57.628 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount changed: [/cache/rhel-8-y.qcow2.snap] => (item={'src': '/dev/mapper/foo-test1', 'path': '/opt/test1', 'state': 'absent', 'fstype': 'xfs'}) => { "ansible_loop_var": "mount_info", "backup_file": "", "boot": "yes", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": { "fstype": "xfs", "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "absent" }, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test1" } TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:146 Thursday 21 July 2022 15:24:29 +0000 (0:00:00.418) 0:01:58.047 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : set up new/current mounts] ****************** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:151 Thursday 21 July 2022 15:24:30 +0000 (0:00:00.641) 0:01:58.689 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount ok: [/cache/rhel-8-y.qcow2.snap] => (item={'src': '/dev/mapper/foo-test2', 'path': '/opt/test2', 'fstype': 'ext4', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted'}) => { "ansible_loop_var": "mount_info", "backup_file": "", "boot": "yes", "changed": false, "dump": "0", "fstab": "/etc/fstab", "fstype": "ext4", "mount_info": { "dump": 0, "fstype": "ext4", "opts": "defaults", "passno": 0, "path": "/opt/test2", "src": "/dev/mapper/foo-test2", "state": "mounted" }, "name": "/opt/test2", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test2" } TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:163 Thursday 21 July 2022 15:24:30 +0000 (0:00:00.413) 0:01:59.102 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : retrieve facts for the /etc/crypttab file] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:171 Thursday 21 July 2022 15:24:31 +0000 (0:00:00.639) 0:01:59.742 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "stat": { "exists": false } } TASK [linux-system-roles.storage : manage /etc/crypttab to account for changes we just made] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:176 Thursday 21 July 2022 15:24:31 +0000 (0:00:00.392) 0:02:00.135 ********* TASK [linux-system-roles.storage : Update facts] ******************************* task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:198 Thursday 21 July 2022 15:24:31 +0000 (0:00:00.026) 0:02:00.161 ********* ok: [/cache/rhel-8-y.qcow2.snap] META: role_complete for /cache/rhel-8-y.qcow2.snap TASK [include_tasks] *********************************************************** task path: /tmp/tmpafnwmbol/tests/tests_lvm_percent_size.yml:136 Thursday 21 July 2022 15:24:32 +0000 (0:00:00.973) 0:02:01.135 ********* included: /tmp/tmpafnwmbol/tests/verify-role-results.yml for /cache/rhel-8-y.qcow2.snap TASK [Print out pool information] ********************************************** task path: /tmp/tmpafnwmbol/tests/verify-role-results.yml:1 Thursday 21 July 2022 15:24:32 +0000 (0:00:00.048) 0:02:01.183 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "60%", "state": "absent", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test2", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test2", "_raw_device": "/dev/mapper/foo-test2", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "25%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] } TASK [Print out volume information] ******************************************** task path: /tmp/tmpafnwmbol/tests/verify-role-results.yml:6 Thursday 21 July 2022 15:24:32 +0000 (0:00:00.088) 0:02:01.272 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => {} TASK [Collect info about the volumes.] ***************************************** task path: /tmp/tmpafnwmbol/tests/verify-role-results.yml:14 Thursday 21 July 2022 15:24:32 +0000 (0:00:00.072) 0:02:01.344 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "info": { "/dev/mapper/foo-test2": { "fstype": "ext4", "label": "", "name": "/dev/mapper/foo-test2", "size": "2.5G", "type": "lvm", "uuid": "ee7d1310-5176-4ec6-80b8-28181ffffed3" }, "/dev/nvme0n1": { "fstype": "", "label": "", "name": "/dev/nvme0n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme1n1": { "fstype": "", "label": "", "name": "/dev/nvme1n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme2n1": { "fstype": "", "label": "", "name": "/dev/nvme2n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sda": { "fstype": "LVM2_member", "label": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "irzZ3c-YIeb-eoxj-fRdj-JSvm-ls7D-iZNahz" }, "/dev/sdb": { "fstype": "", "label": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc": { "fstype": "", "label": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sr0": { "fstype": "iso9660", "label": "cidata", "name": "/dev/sr0", "size": "364K", "type": "rom", "uuid": "2022-07-21-15-22-18-00" }, "/dev/vda": { "fstype": "", "label": "", "name": "/dev/vda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vda1": { "fstype": "", "label": "", "name": "/dev/vda1", "size": "1M", "type": "partition", "uuid": "" }, "/dev/vda2": { "fstype": "vfat", "label": "", "name": "/dev/vda2", "size": "100M", "type": "partition", "uuid": "7B77-95E7" }, "/dev/vda3": { "fstype": "xfs", "label": "root", "name": "/dev/vda3", "size": "9.9G", "type": "partition", "uuid": "58dca2db-4ccd-4171-a017-dd53ad8c1c54" }, "/dev/vdb": { "fstype": "", "label": "", "name": "/dev/vdb", "size": "2G", "type": "disk", "uuid": "" }, "/dev/vdb1": { "fstype": "ext4", "label": "yumcache", "name": "/dev/vdb1", "size": "2G", "type": "partition", "uuid": "4d1ed9a1-5608-4ff8-ad89-21d95caa66fa" }, "/dev/vdc": { "fstype": "", "label": "", "name": "/dev/vdc", "size": "2G", "type": "disk", "uuid": "" }, "/dev/vdc1": { "fstype": "ext4", "label": "yumvarlib", "name": "/dev/vdc1", "size": "2G", "type": "partition", "uuid": "a7abcc97-edc5-476c-8ccd-694b4621fd28" }, "/dev/vdd": { "fstype": "", "label": "", "name": "/dev/vdd", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vde": { "fstype": "", "label": "", "name": "/dev/vde", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vdf": { "fstype": "", "label": "", "name": "/dev/vdf", "size": "10G", "type": "disk", "uuid": "" } } } TASK [Read the /etc/fstab file for volume existence] *************************** task path: /tmp/tmpafnwmbol/tests/verify-role-results.yml:19 Thursday 21 July 2022 15:24:33 +0000 (0:00:00.390) 0:02:01.734 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "cmd": [ "cat", "/etc/fstab" ], "delta": "0:00:01.004292", "end": "2022-07-21 11:24:34.513787", "rc": 0, "start": "2022-07-21 11:24:33.509495" } STDOUT: UUID=58dca2db-4ccd-4171-a017-dd53ad8c1c54 / xfs defaults 0 0 UUID=7B77-95E7 /boot/efi vfat defaults,uid=0,gid=0,umask=077,shortname=winnt 0 2 /dev/vdb1 /var/cache/dnf auto defaults,nofail,x-systemd.requires=cloud-init.service,_netdev,comment=cloudconfig 0 2 /dev/vdc1 /var/lib/dnf auto defaults,nofail,x-systemd.requires=cloud-init.service,_netdev,comment=cloudconfig 0 2 /dev/mapper/foo-test2 /opt/test2 ext4 defaults 0 0 TASK [Read the /etc/crypttab file] ********************************************* task path: /tmp/tmpafnwmbol/tests/verify-role-results.yml:24 Thursday 21 July 2022 15:24:34 +0000 (0:00:01.390) 0:02:03.125 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "cmd": [ "cat", "/etc/crypttab" ], "delta": "0:00:00.002877", "end": "2022-07-21 11:24:34.892541", "failed_when_result": false, "rc": 1, "start": "2022-07-21 11:24:34.889664" } STDERR: cat: /etc/crypttab: No such file or directory MSG: non-zero return code TASK [Verify the volumes listed in storage_pools were correctly managed] ******* task path: /tmp/tmpafnwmbol/tests/verify-role-results.yml:33 Thursday 21 July 2022 15:24:34 +0000 (0:00:00.379) 0:02:03.504 ********* included: /tmp/tmpafnwmbol/tests/test-verify-pool.yml for /cache/rhel-8-y.qcow2.snap => (item={'disks': ['sda'], 'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'name': 'foo', 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_metadata_version': None, 'state': 'present', 'type': 'lvm', 'volumes': [{'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'name': 'test1', 'raid_level': None, 'size': '60%', 'state': 'absent', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test1', '_raw_device': '/dev/mapper/foo-test1', '_mount_id': '/dev/mapper/foo-test1'}, {'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'ext4', 'mount_options': 'defaults', 'mount_point': '/opt/test2', 'name': 'test2', 'raid_level': None, 'size': '25%', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test2', '_raw_device': '/dev/mapper/foo-test2', '_mount_id': '/dev/mapper/foo-test2', '_kernel_device': '/dev/dm-0', '_raw_kernel_device': '/dev/dm-0'}], 'raid_chunk_size': None}) TASK [Set _storage_pool_tests] ************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-pool.yml:5 Thursday 21 July 2022 15:24:35 +0000 (0:00:00.065) 0:02:03.570 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_pool_tests": [ "members", "volumes" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool.yml:18 Thursday 21 July 2022 15:24:35 +0000 (0:00:00.037) 0:02:03.608 ********* included: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml for /cache/rhel-8-y.qcow2.snap => (item=members) included: /tmp/tmpafnwmbol/tests/test-verify-pool-volumes.yml for /cache/rhel-8-y.qcow2.snap => (item=volumes) TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:1 Thursday 21 July 2022 15:24:35 +0000 (0:00:00.051) 0:02:03.659 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_pv_count": "1", "_storage_test_pool_pvs_lvm": [ "/dev/sda" ] }, "changed": false } TASK [Get the canonical device path for each member device] ******************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:10 Thursday 21 July 2022 15:24:35 +0000 (0:00:00.061) 0:02:03.720 ********* ok: [/cache/rhel-8-y.qcow2.snap] => (item=/dev/sda) => { "ansible_loop_var": "pv", "changed": false, "device": "/dev/sda", "pv": "/dev/sda" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:19 Thursday 21 July 2022 15:24:35 +0000 (0:00:00.388) 0:02:04.109 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "__pvs_lvm_len": "1" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:23 Thursday 21 July 2022 15:24:35 +0000 (0:00:00.053) 0:02:04.163 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_pool_pvs": [ "/dev/sda" ] }, "changed": false } TASK [Verify PV count] ********************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:27 Thursday 21 July 2022 15:24:35 +0000 (0:00:00.051) 0:02:04.214 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:34 Thursday 21 July 2022 15:24:35 +0000 (0:00:00.099) 0:02:04.313 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:38 Thursday 21 July 2022 15:24:35 +0000 (0:00:00.041) 0:02:04.355 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:42 Thursday 21 July 2022 15:24:35 +0000 (0:00:00.099) 0:02:04.455 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check the type of each PV] *********************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:46 Thursday 21 July 2022 15:24:35 +0000 (0:00:00.024) 0:02:04.480 ********* ok: [/cache/rhel-8-y.qcow2.snap] => (item=/dev/sda) => { "ansible_loop_var": "pv", "changed": false, "pv": "/dev/sda" } MSG: All assertions passed TASK [Check MD RAID] *********************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:56 Thursday 21 July 2022 15:24:36 +0000 (0:00:00.078) 0:02:04.558 ********* included: /tmp/tmpafnwmbol/tests/verify-pool-md.yml for /cache/rhel-8-y.qcow2.snap TASK [get information about RAID] ********************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-md.yml:6 Thursday 21 July 2022 15:24:36 +0000 (0:00:00.043) 0:02:04.602 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-md.yml:12 Thursday 21 July 2022 15:24:36 +0000 (0:00:00.027) 0:02:04.630 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-md.yml:16 Thursday 21 July 2022 15:24:36 +0000 (0:00:00.059) 0:02:04.689 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-md.yml:20 Thursday 21 July 2022 15:24:36 +0000 (0:00:00.025) 0:02:04.715 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-md.yml:24 Thursday 21 July 2022 15:24:36 +0000 (0:00:00.023) 0:02:04.739 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-md.yml:30 Thursday 21 July 2022 15:24:36 +0000 (0:00:00.023) 0:02:04.762 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmpafnwmbol/tests/verify-pool-md.yml:36 Thursday 21 July 2022 15:24:36 +0000 (0:00:00.025) 0:02:04.787 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-md.yml:44 Thursday 21 July 2022 15:24:36 +0000 (0:00:00.024) 0:02:04.812 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_md_active_devices_re": null, "storage_test_md_metadata_version_re": null, "storage_test_md_spare_devices_re": null }, "changed": false } TASK [Check LVM RAID] ********************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:59 Thursday 21 July 2022 15:24:36 +0000 (0:00:00.037) 0:02:04.849 ********* included: /tmp/tmpafnwmbol/tests/verify-pool-members-lvmraid.yml for /cache/rhel-8-y.qcow2.snap TASK [Validate pool member LVM RAID settings] ********************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-members-lvmraid.yml:1 Thursday 21 July 2022 15:24:36 +0000 (0:00:00.046) 0:02:04.895 ********* included: /tmp/tmpafnwmbol/tests/verify-pool-member-lvmraid.yml for /cache/rhel-8-y.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'name': 'test1', 'raid_level': None, 'size': '60%', 'state': 'absent', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test1', '_raw_device': '/dev/mapper/foo-test1', '_mount_id': '/dev/mapper/foo-test1'}) included: /tmp/tmpafnwmbol/tests/verify-pool-member-lvmraid.yml for /cache/rhel-8-y.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'ext4', 'mount_options': 'defaults', 'mount_point': '/opt/test2', 'name': 'test2', 'raid_level': None, 'size': '25%', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test2', '_raw_device': '/dev/mapper/foo-test2', '_mount_id': '/dev/mapper/foo-test2', '_kernel_device': '/dev/dm-0', '_raw_kernel_device': '/dev/dm-0'}) TASK [Get information about LVM RAID] ****************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-lvmraid.yml:3 Thursday 21 July 2022 15:24:36 +0000 (0:00:00.050) 0:02:04.946 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is LVM RAID] ******************************************* task path: /tmp/tmpafnwmbol/tests/verify-pool-member-lvmraid.yml:8 Thursday 21 July 2022 15:24:36 +0000 (0:00:00.031) 0:02:04.978 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-lvmraid.yml:12 Thursday 21 July 2022 15:24:36 +0000 (0:00:00.029) 0:02:05.007 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about LVM RAID] ****************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-lvmraid.yml:3 Thursday 21 July 2022 15:24:36 +0000 (0:00:00.029) 0:02:05.037 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is LVM RAID] ******************************************* task path: /tmp/tmpafnwmbol/tests/verify-pool-member-lvmraid.yml:8 Thursday 21 July 2022 15:24:36 +0000 (0:00:00.028) 0:02:05.065 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-lvmraid.yml:12 Thursday 21 July 2022 15:24:36 +0000 (0:00:00.029) 0:02:05.094 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check Thin Pools] ******************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:62 Thursday 21 July 2022 15:24:36 +0000 (0:00:00.028) 0:02:05.123 ********* included: /tmp/tmpafnwmbol/tests/verify-pool-members-thin.yml for /cache/rhel-8-y.qcow2.snap TASK [Validate pool member thinpool settings] ********************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-members-thin.yml:1 Thursday 21 July 2022 15:24:36 +0000 (0:00:00.045) 0:02:05.168 ********* included: /tmp/tmpafnwmbol/tests/verify-pool-member-thin.yml for /cache/rhel-8-y.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'name': 'test1', 'raid_level': None, 'size': '60%', 'state': 'absent', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test1', '_raw_device': '/dev/mapper/foo-test1', '_mount_id': '/dev/mapper/foo-test1'}) included: /tmp/tmpafnwmbol/tests/verify-pool-member-thin.yml for /cache/rhel-8-y.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'ext4', 'mount_options': 'defaults', 'mount_point': '/opt/test2', 'name': 'test2', 'raid_level': None, 'size': '25%', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test2', '_raw_device': '/dev/mapper/foo-test2', '_mount_id': '/dev/mapper/foo-test2', '_kernel_device': '/dev/dm-0', '_raw_kernel_device': '/dev/dm-0'}) TASK [Get information about thinpool] ****************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-thin.yml:3 Thursday 21 July 2022 15:24:36 +0000 (0:00:00.054) 0:02:05.222 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-thin.yml:8 Thursday 21 July 2022 15:24:36 +0000 (0:00:00.025) 0:02:05.248 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-thin.yml:13 Thursday 21 July 2022 15:24:36 +0000 (0:00:00.025) 0:02:05.273 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-thin.yml:17 Thursday 21 July 2022 15:24:36 +0000 (0:00:00.024) 0:02:05.297 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about thinpool] ****************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-thin.yml:3 Thursday 21 July 2022 15:24:36 +0000 (0:00:00.026) 0:02:05.324 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-thin.yml:8 Thursday 21 July 2022 15:24:36 +0000 (0:00:00.025) 0:02:05.350 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-thin.yml:13 Thursday 21 July 2022 15:24:36 +0000 (0:00:00.027) 0:02:05.378 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-thin.yml:17 Thursday 21 July 2022 15:24:36 +0000 (0:00:00.025) 0:02:05.404 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check member encryption] ************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:65 Thursday 21 July 2022 15:24:36 +0000 (0:00:00.025) 0:02:05.429 ********* included: /tmp/tmpafnwmbol/tests/verify-pool-members-encryption.yml for /cache/rhel-8-y.qcow2.snap TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-members-encryption.yml:4 Thursday 21 July 2022 15:24:36 +0000 (0:00:00.050) 0:02:05.479 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Validate pool member LUKS settings] ************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-members-encryption.yml:8 Thursday 21 July 2022 15:24:37 +0000 (0:00:00.049) 0:02:05.529 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => (item=/dev/sda) => { "_storage_test_pool_member_path": "/dev/sda", "ansible_loop_var": "_storage_test_pool_member_path", "changed": false, "skip_reason": "Conditional result was False" } TASK [Validate pool member crypttab entries] *********************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-members-encryption.yml:15 Thursday 21 July 2022 15:24:37 +0000 (0:00:00.027) 0:02:05.557 ********* included: /tmp/tmpafnwmbol/tests/verify-pool-member-crypttab.yml for /cache/rhel-8-y.qcow2.snap => (item=/dev/sda) TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-crypttab.yml:1 Thursday 21 July 2022 15:24:37 +0000 (0:00:00.045) 0:02:05.603 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": [] }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpafnwmbol/tests/verify-pool-member-crypttab.yml:6 Thursday 21 July 2022 15:24:37 +0000 (0:00:00.050) 0:02:05.654 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpafnwmbol/tests/verify-pool-member-crypttab.yml:11 Thursday 21 July 2022 15:24:37 +0000 (0:00:00.055) 0:02:05.709 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-crypttab.yml:17 Thursday 21 July 2022 15:24:37 +0000 (0:00:00.045) 0:02:05.754 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-crypttab.yml:23 Thursday 21 July 2022 15:24:37 +0000 (0:00:00.036) 0:02:05.791 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-crypttab.yml:29 Thursday 21 July 2022 15:24:37 +0000 (0:00:00.038) 0:02:05.829 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-members-encryption.yml:22 Thursday 21 July 2022 15:24:37 +0000 (0:00:00.087) 0:02:05.917 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_crypttab_key_file": null }, "changed": false } TASK [Check VDO] *************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:68 Thursday 21 July 2022 15:24:37 +0000 (0:00:00.091) 0:02:06.008 ********* included: /tmp/tmpafnwmbol/tests/verify-pool-members-vdo.yml for /cache/rhel-8-y.qcow2.snap TASK [Validate pool member VDO settings] *************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-members-vdo.yml:1 Thursday 21 July 2022 15:24:37 +0000 (0:00:00.048) 0:02:06.057 ********* included: /tmp/tmpafnwmbol/tests/verify-pool-member-vdo.yml for /cache/rhel-8-y.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'name': 'test1', 'raid_level': None, 'size': '60%', 'state': 'absent', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test1', '_raw_device': '/dev/mapper/foo-test1', '_mount_id': '/dev/mapper/foo-test1'}) included: /tmp/tmpafnwmbol/tests/verify-pool-member-vdo.yml for /cache/rhel-8-y.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'ext4', 'mount_options': 'defaults', 'mount_point': '/opt/test2', 'name': 'test2', 'raid_level': None, 'size': '25%', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test2', '_raw_device': '/dev/mapper/foo-test2', '_mount_id': '/dev/mapper/foo-test2', '_kernel_device': '/dev/dm-0', '_raw_kernel_device': '/dev/dm-0'}) TASK [get information about VDO deduplication] ********************************* task path: /tmp/tmpafnwmbol/tests/verify-pool-member-vdo.yml:3 Thursday 21 July 2022 15:24:37 +0000 (0:00:00.055) 0:02:06.113 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-vdo.yml:8 Thursday 21 July 2022 15:24:37 +0000 (0:00:00.025) 0:02:06.138 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-vdo.yml:11 Thursday 21 July 2022 15:24:37 +0000 (0:00:00.025) 0:02:06.164 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-vdo.yml:16 Thursday 21 July 2022 15:24:37 +0000 (0:00:00.025) 0:02:06.189 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-vdo.yml:21 Thursday 21 July 2022 15:24:37 +0000 (0:00:00.025) 0:02:06.215 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-vdo.yml:24 Thursday 21 July 2022 15:24:37 +0000 (0:00:00.025) 0:02:06.240 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-vdo.yml:29 Thursday 21 July 2022 15:24:37 +0000 (0:00:00.026) 0:02:06.267 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-vdo.yml:39 Thursday 21 July 2022 15:24:37 +0000 (0:00:00.026) 0:02:06.293 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [get information about VDO deduplication] ********************************* task path: /tmp/tmpafnwmbol/tests/verify-pool-member-vdo.yml:3 Thursday 21 July 2022 15:24:37 +0000 (0:00:00.038) 0:02:06.331 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-vdo.yml:8 Thursday 21 July 2022 15:24:37 +0000 (0:00:00.024) 0:02:06.355 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-vdo.yml:11 Thursday 21 July 2022 15:24:37 +0000 (0:00:00.027) 0:02:06.382 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-vdo.yml:16 Thursday 21 July 2022 15:24:37 +0000 (0:00:00.025) 0:02:06.408 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-vdo.yml:21 Thursday 21 July 2022 15:24:37 +0000 (0:00:00.025) 0:02:06.433 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-vdo.yml:24 Thursday 21 July 2022 15:24:37 +0000 (0:00:00.024) 0:02:06.457 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-vdo.yml:29 Thursday 21 July 2022 15:24:37 +0000 (0:00:00.025) 0:02:06.482 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-vdo.yml:39 Thursday 21 July 2022 15:24:37 +0000 (0:00:00.023) 0:02:06.506 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:71 Thursday 21 July 2022 15:24:38 +0000 (0:00:00.037) 0:02:06.544 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "__pvs_lvm_len": null, "_storage_test_expected_pv_count": null, "_storage_test_expected_pv_type": null, "_storage_test_pool_pvs": [], "_storage_test_pool_pvs_lvm": [] }, "changed": false } TASK [verify the volumes] ****************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool-volumes.yml:3 Thursday 21 July 2022 15:24:38 +0000 (0:00:00.038) 0:02:06.582 ********* included: /tmp/tmpafnwmbol/tests/test-verify-volume.yml for /cache/rhel-8-y.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'name': 'test1', 'raid_level': None, 'size': '60%', 'state': 'absent', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test1', '_raw_device': '/dev/mapper/foo-test1', '_mount_id': '/dev/mapper/foo-test1'}) included: /tmp/tmpafnwmbol/tests/test-verify-volume.yml for /cache/rhel-8-y.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'ext4', 'mount_options': 'defaults', 'mount_point': '/opt/test2', 'name': 'test2', 'raid_level': None, 'size': '25%', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test2', '_raw_device': '/dev/mapper/foo-test2', '_mount_id': '/dev/mapper/foo-test2', '_kernel_device': '/dev/dm-0', '_raw_kernel_device': '/dev/dm-0'}) TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume.yml:2 Thursday 21 July 2022 15:24:38 +0000 (0:00:00.052) 0:02:06.635 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_volume_present": false, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume.yml:10 Thursday 21 July 2022 15:24:38 +0000 (0:00:00.050) 0:02:06.685 ********* included: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml for /cache/rhel-8-y.qcow2.snap => (item=mount) included: /tmp/tmpafnwmbol/tests/test-verify-volume-fstab.yml for /cache/rhel-8-y.qcow2.snap => (item=fstab) included: /tmp/tmpafnwmbol/tests/test-verify-volume-fs.yml for /cache/rhel-8-y.qcow2.snap => (item=fs) included: /tmp/tmpafnwmbol/tests/test-verify-volume-device.yml for /cache/rhel-8-y.qcow2.snap => (item=device) included: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml for /cache/rhel-8-y.qcow2.snap => (item=encryption) included: /tmp/tmpafnwmbol/tests/test-verify-volume-md.yml for /cache/rhel-8-y.qcow2.snap => (item=md) included: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml for /cache/rhel-8-y.qcow2.snap => (item=size) included: /tmp/tmpafnwmbol/tests/test-verify-volume-cache.yml for /cache/rhel-8-y.qcow2.snap => (item=cache) TASK [Get expected mount device based on device type] ************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml:6 Thursday 21 July 2022 15:24:38 +0000 (0:00:00.100) 0:02:06.786 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/foo-test1" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml:14 Thursday 21 July 2022 15:24:38 +0000 (0:00:00.041) 0:02:06.827 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_mount_device_matches": [], "storage_test_mount_expected_match_count": "0", "storage_test_mount_point_matches": [], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Verify the current mount state by device] ******************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml:28 Thursday 21 July 2022 15:24:38 +0000 (0:00:00.056) 0:02:06.884 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the current mount state by mount point] *************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml:37 Thursday 21 July 2022 15:24:38 +0000 (0:00:00.023) 0:02:06.908 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify the mount fs type] ************************************************ task path: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml:45 Thursday 21 July 2022 15:24:38 +0000 (0:00:00.052) 0:02:06.960 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [command] ***************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml:54 Thursday 21 July 2022 15:24:38 +0000 (0:00:00.038) 0:02:06.999 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml:58 Thursday 21 July 2022 15:24:38 +0000 (0:00:00.023) 0:02:07.022 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml:63 Thursday 21 July 2022 15:24:38 +0000 (0:00:00.023) 0:02:07.046 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml:75 Thursday 21 July 2022 15:24:38 +0000 (0:00:00.023) 0:02:07.069 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-fstab.yml:2 Thursday 21 July 2022 15:24:38 +0000 (0:00:00.036) 0:02:07.106 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "0", "storage_test_fstab_expected_mount_options_matches": "0", "storage_test_fstab_expected_mount_point_matches": "0", "storage_test_fstab_id_matches": [], "storage_test_fstab_mount_options_matches": [], "storage_test_fstab_mount_point_matches": [] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-fstab.yml:25 Thursday 21 July 2022 15:24:38 +0000 (0:00:00.106) 0:02:07.212 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the fstab mount point] ******************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-fstab.yml:32 Thursday 21 July 2022 15:24:38 +0000 (0:00:00.024) 0:02:07.236 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-fstab.yml:39 Thursday 21 July 2022 15:24:38 +0000 (0:00:00.127) 0:02:07.363 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-fstab.yml:49 Thursday 21 July 2022 15:24:38 +0000 (0:00:00.036) 0:02:07.400 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-fs.yml:4 Thursday 21 July 2022 15:24:38 +0000 (0:00:00.034) 0:02:07.434 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify fs label] ********************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-fs.yml:10 Thursday 21 July 2022 15:24:38 +0000 (0:00:00.024) 0:02:07.458 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [See whether the device node is present] ********************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-device.yml:4 Thursday 21 July 2022 15:24:38 +0000 (0:00:00.029) 0:02:07.488 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "stat": { "exists": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-device.yml:10 Thursday 21 July 2022 15:24:39 +0000 (0:00:00.377) 0:02:07.866 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about this volume] ********************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-device.yml:18 Thursday 21 July 2022 15:24:39 +0000 (0:00:00.042) 0:02:07.909 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [(1/2) Process volume type (set initial value)] *************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-device.yml:24 Thursday 21 July 2022 15:24:39 +0000 (0:00:00.026) 0:02:07.935 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [(2/2) Process volume type (get RAID value)] ****************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-device.yml:28 Thursday 21 July 2022 15:24:39 +0000 (0:00:00.038) 0:02:07.974 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-device.yml:33 Thursday 21 July 2022 15:24:39 +0000 (0:00:00.026) 0:02:08.001 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Stat the LUKS device, if encrypted] ************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:3 Thursday 21 July 2022 15:24:39 +0000 (0:00:00.025) 0:02:08.026 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:10 Thursday 21 July 2022 15:24:39 +0000 (0:00:00.029) 0:02:08.055 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [Collect LUKS info for this volume] *************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:15 Thursday 21 July 2022 15:24:40 +0000 (0:00:01.342) 0:02:09.397 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:21 Thursday 21 July 2022 15:24:40 +0000 (0:00:00.023) 0:02:09.421 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:30 Thursday 21 July 2022 15:24:40 +0000 (0:00:00.025) 0:02:09.446 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:38 Thursday 21 July 2022 15:24:40 +0000 (0:00:00.025) 0:02:09.472 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:44 Thursday 21 July 2022 15:24:40 +0000 (0:00:00.025) 0:02:09.497 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:49 Thursday 21 July 2022 15:24:41 +0000 (0:00:00.024) 0:02:09.521 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:55 Thursday 21 July 2022 15:24:41 +0000 (0:00:00.025) 0:02:09.547 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:61 Thursday 21 July 2022 15:24:41 +0000 (0:00:00.024) 0:02:09.571 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:67 Thursday 21 July 2022 15:24:41 +0000 (0:00:00.025) 0:02:09.596 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:74 Thursday 21 July 2022 15:24:41 +0000 (0:00:00.051) 0:02:09.648 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:79 Thursday 21 July 2022 15:24:41 +0000 (0:00:00.048) 0:02:09.696 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:85 Thursday 21 July 2022 15:24:41 +0000 (0:00:00.054) 0:02:09.751 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:91 Thursday 21 July 2022 15:24:41 +0000 (0:00:00.041) 0:02:09.793 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:97 Thursday 21 July 2022 15:24:41 +0000 (0:00:00.040) 0:02:09.833 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [get information about RAID] ********************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-md.yml:7 Thursday 21 July 2022 15:24:41 +0000 (0:00:00.035) 0:02:09.868 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-md.yml:13 Thursday 21 July 2022 15:24:41 +0000 (0:00:00.038) 0:02:09.907 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-md.yml:17 Thursday 21 July 2022 15:24:41 +0000 (0:00:00.037) 0:02:09.945 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-md.yml:21 Thursday 21 July 2022 15:24:41 +0000 (0:00:00.037) 0:02:09.983 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-md.yml:25 Thursday 21 July 2022 15:24:41 +0000 (0:00:00.036) 0:02:10.019 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-md.yml:31 Thursday 21 July 2022 15:24:41 +0000 (0:00:00.038) 0:02:10.057 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-md.yml:37 Thursday 21 July 2022 15:24:41 +0000 (0:00:00.037) 0:02:10.094 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the actual size of the volume] ************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:3 Thursday 21 July 2022 15:24:41 +0000 (0:00:00.040) 0:02:10.134 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the requested size of the volume] ********************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:9 Thursday 21 July 2022 15:24:41 +0000 (0:00:00.063) 0:02:10.198 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Establish base value for expected size] ********************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:15 Thursday 21 July 2022 15:24:41 +0000 (0:00:00.040) 0:02:10.238 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:20 Thursday 21 July 2022 15:24:41 +0000 (0:00:00.036) 0:02:10.275 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_expected_size": "2684354560.0" } TASK [debug] ******************************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:25 Thursday 21 July 2022 15:24:41 +0000 (0:00:00.037) 0:02:10.312 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => {} TASK [debug] ******************************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:28 Thursday 21 July 2022 15:24:41 +0000 (0:00:00.040) 0:02:10.353 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => {} TASK [Get the size of parent/pool device] ************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:31 Thursday 21 July 2022 15:24:41 +0000 (0:00:00.038) 0:02:10.392 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:36 Thursday 21 July 2022 15:24:41 +0000 (0:00:00.037) 0:02:10.430 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => {} TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:39 Thursday 21 July 2022 15:24:41 +0000 (0:00:00.037) 0:02:10.467 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:44 Thursday 21 July 2022 15:24:41 +0000 (0:00:00.039) 0:02:10.507 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_actual_size": { "changed": false, "skip_reason": "Conditional result was False", "skipped": true } } TASK [debug] ******************************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:47 Thursday 21 July 2022 15:24:42 +0000 (0:00:00.035) 0:02:10.543 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_expected_size": "2684354560.0" } TASK [assert] ****************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:50 Thursday 21 July 2022 15:24:42 +0000 (0:00:00.035) 0:02:10.579 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about the LV] ******************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-cache.yml:6 Thursday 21 July 2022 15:24:42 +0000 (0:00:00.024) 0:02:10.603 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-cache.yml:14 Thursday 21 July 2022 15:24:42 +0000 (0:00:00.023) 0:02:10.627 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check segment type] ****************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-cache.yml:17 Thursday 21 July 2022 15:24:42 +0000 (0:00:00.022) 0:02:10.649 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-cache.yml:22 Thursday 21 July 2022 15:24:42 +0000 (0:00:00.023) 0:02:10.672 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the requested cache size] ****************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-cache.yml:26 Thursday 21 July 2022 15:24:42 +0000 (0:00:00.023) 0:02:10.695 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-cache.yml:32 Thursday 21 July 2022 15:24:42 +0000 (0:00:00.023) 0:02:10.718 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-cache.yml:36 Thursday 21 July 2022 15:24:42 +0000 (0:00:00.026) 0:02:10.745 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume.yml:16 Thursday 21 July 2022 15:24:42 +0000 (0:00:00.025) 0:02:10.771 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume.yml:2 Thursday 21 July 2022 15:24:42 +0000 (0:00:00.035) 0:02:10.807 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_volume_present": true, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume.yml:10 Thursday 21 July 2022 15:24:42 +0000 (0:00:00.051) 0:02:10.858 ********* included: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml for /cache/rhel-8-y.qcow2.snap => (item=mount) included: /tmp/tmpafnwmbol/tests/test-verify-volume-fstab.yml for /cache/rhel-8-y.qcow2.snap => (item=fstab) included: /tmp/tmpafnwmbol/tests/test-verify-volume-fs.yml for /cache/rhel-8-y.qcow2.snap => (item=fs) included: /tmp/tmpafnwmbol/tests/test-verify-volume-device.yml for /cache/rhel-8-y.qcow2.snap => (item=device) included: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml for /cache/rhel-8-y.qcow2.snap => (item=encryption) included: /tmp/tmpafnwmbol/tests/test-verify-volume-md.yml for /cache/rhel-8-y.qcow2.snap => (item=md) included: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml for /cache/rhel-8-y.qcow2.snap => (item=size) included: /tmp/tmpafnwmbol/tests/test-verify-volume-cache.yml for /cache/rhel-8-y.qcow2.snap => (item=cache) TASK [Get expected mount device based on device type] ************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml:6 Thursday 21 July 2022 15:24:42 +0000 (0:00:00.078) 0:02:10.936 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/foo-test2" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml:14 Thursday 21 July 2022 15:24:42 +0000 (0:00:00.044) 0:02:10.981 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_mount_device_matches": [ { "block_available": 588749, "block_size": 4096, "block_total": 625618, "block_used": 36869, "device": "/dev/mapper/foo-test2", "fstype": "ext4", "inode_available": 163829, "inode_total": 163840, "inode_used": 11, "mount": "/opt/test2", "options": "rw,seclabel,relatime", "size_available": 2411515904, "size_total": 2562531328, "uuid": "ee7d1310-5176-4ec6-80b8-28181ffffed3" } ], "storage_test_mount_expected_match_count": "1", "storage_test_mount_point_matches": [ { "block_available": 588749, "block_size": 4096, "block_total": 625618, "block_used": 36869, "device": "/dev/mapper/foo-test2", "fstype": "ext4", "inode_available": 163829, "inode_total": 163840, "inode_used": 11, "mount": "/opt/test2", "options": "rw,seclabel,relatime", "size_available": 2411515904, "size_total": 2562531328, "uuid": "ee7d1310-5176-4ec6-80b8-28181ffffed3" } ], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Verify the current mount state by device] ******************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml:28 Thursday 21 July 2022 15:24:42 +0000 (0:00:00.056) 0:02:11.037 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify the current mount state by mount point] *************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml:37 Thursday 21 July 2022 15:24:42 +0000 (0:00:00.051) 0:02:11.088 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify the mount fs type] ************************************************ task path: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml:45 Thursday 21 July 2022 15:24:42 +0000 (0:00:00.050) 0:02:11.138 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [command] ***************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml:54 Thursday 21 July 2022 15:24:42 +0000 (0:00:00.053) 0:02:11.192 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml:58 Thursday 21 July 2022 15:24:42 +0000 (0:00:00.026) 0:02:11.218 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml:63 Thursday 21 July 2022 15:24:42 +0000 (0:00:00.026) 0:02:11.244 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml:75 Thursday 21 July 2022 15:24:42 +0000 (0:00:00.023) 0:02:11.268 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-fstab.yml:2 Thursday 21 July 2022 15:24:42 +0000 (0:00:00.035) 0:02:11.303 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "1", "storage_test_fstab_expected_mount_options_matches": "1", "storage_test_fstab_expected_mount_point_matches": "1", "storage_test_fstab_id_matches": [ "/dev/mapper/foo-test2 " ], "storage_test_fstab_mount_options_matches": [ " /opt/test2 ext4 defaults " ], "storage_test_fstab_mount_point_matches": [ " /opt/test2 " ] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-fstab.yml:25 Thursday 21 July 2022 15:24:42 +0000 (0:00:00.062) 0:02:11.366 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify the fstab mount point] ******************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-fstab.yml:32 Thursday 21 July 2022 15:24:42 +0000 (0:00:00.051) 0:02:11.418 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-fstab.yml:39 Thursday 21 July 2022 15:24:42 +0000 (0:00:00.053) 0:02:11.471 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-fstab.yml:49 Thursday 21 July 2022 15:24:42 +0000 (0:00:00.041) 0:02:11.513 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-fs.yml:4 Thursday 21 July 2022 15:24:43 +0000 (0:00:00.035) 0:02:11.549 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify fs label] ********************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-fs.yml:10 Thursday 21 July 2022 15:24:43 +0000 (0:00:00.072) 0:02:11.621 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [See whether the device node is present] ********************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-device.yml:4 Thursday 21 July 2022 15:24:43 +0000 (0:00:00.044) 0:02:11.665 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "stat": { "atime": 1658417038.22426, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1658417038.22426, "dev": 6, "device_type": 64768, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 62525, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/symlink", "mode": "0660", "mtime": 1658417038.22426, "nlink": 1, "path": "/dev/mapper/foo-test2", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-device.yml:10 Thursday 21 July 2022 15:24:43 +0000 (0:00:00.387) 0:02:12.053 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about this volume] ********************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-device.yml:18 Thursday 21 July 2022 15:24:43 +0000 (0:00:00.075) 0:02:12.129 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [(1/2) Process volume type (set initial value)] *************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-device.yml:24 Thursday 21 July 2022 15:24:43 +0000 (0:00:00.077) 0:02:12.206 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [(2/2) Process volume type (get RAID value)] ****************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-device.yml:28 Thursday 21 July 2022 15:24:43 +0000 (0:00:00.081) 0:02:12.287 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-device.yml:33 Thursday 21 July 2022 15:24:43 +0000 (0:00:00.032) 0:02:12.320 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Stat the LUKS device, if encrypted] ************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:3 Thursday 21 July 2022 15:24:43 +0000 (0:00:00.100) 0:02:12.420 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:10 Thursday 21 July 2022 15:24:43 +0000 (0:00:00.023) 0:02:12.444 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [Collect LUKS info for this volume] *************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:15 Thursday 21 July 2022 15:24:45 +0000 (0:00:01.283) 0:02:13.727 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:21 Thursday 21 July 2022 15:24:45 +0000 (0:00:00.027) 0:02:13.754 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:30 Thursday 21 July 2022 15:24:45 +0000 (0:00:00.026) 0:02:13.781 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:38 Thursday 21 July 2022 15:24:45 +0000 (0:00:00.054) 0:02:13.836 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:44 Thursday 21 July 2022 15:24:45 +0000 (0:00:00.023) 0:02:13.859 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:49 Thursday 21 July 2022 15:24:45 +0000 (0:00:00.023) 0:02:13.883 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:55 Thursday 21 July 2022 15:24:45 +0000 (0:00:00.023) 0:02:13.907 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:61 Thursday 21 July 2022 15:24:45 +0000 (0:00:00.023) 0:02:13.930 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:67 Thursday 21 July 2022 15:24:45 +0000 (0:00:00.025) 0:02:13.956 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:74 Thursday 21 July 2022 15:24:45 +0000 (0:00:00.053) 0:02:14.009 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:79 Thursday 21 July 2022 15:24:45 +0000 (0:00:00.048) 0:02:14.057 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:85 Thursday 21 July 2022 15:24:45 +0000 (0:00:00.038) 0:02:14.095 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:91 Thursday 21 July 2022 15:24:45 +0000 (0:00:00.042) 0:02:14.138 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:97 Thursday 21 July 2022 15:24:45 +0000 (0:00:00.037) 0:02:14.176 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [get information about RAID] ********************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-md.yml:7 Thursday 21 July 2022 15:24:45 +0000 (0:00:00.036) 0:02:14.213 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-md.yml:13 Thursday 21 July 2022 15:24:45 +0000 (0:00:00.035) 0:02:14.248 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-md.yml:17 Thursday 21 July 2022 15:24:45 +0000 (0:00:00.038) 0:02:14.287 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-md.yml:21 Thursday 21 July 2022 15:24:45 +0000 (0:00:00.038) 0:02:14.325 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-md.yml:25 Thursday 21 July 2022 15:24:45 +0000 (0:00:00.039) 0:02:14.364 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-md.yml:31 Thursday 21 July 2022 15:24:45 +0000 (0:00:00.037) 0:02:14.402 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-md.yml:37 Thursday 21 July 2022 15:24:45 +0000 (0:00:00.040) 0:02:14.442 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the actual size of the volume] ************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:3 Thursday 21 July 2022 15:24:45 +0000 (0:00:00.035) 0:02:14.478 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "bytes": 2684354560, "changed": false, "lvm": "2g", "parted": "2GiB", "size": "2 GiB" } TASK [parse the requested size of the volume] ********************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:9 Thursday 21 July 2022 15:24:46 +0000 (0:00:00.372) 0:02:14.851 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Establish base value for expected size] ********************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:15 Thursday 21 July 2022 15:24:46 +0000 (0:00:00.039) 0:02:14.890 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:20 Thursday 21 July 2022 15:24:46 +0000 (0:00:00.040) 0:02:14.931 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_expected_size": "2684354560.0" } TASK [debug] ******************************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:25 Thursday 21 July 2022 15:24:46 +0000 (0:00:00.038) 0:02:14.970 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_pool": { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "60%", "state": "absent", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test2", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test2", "_raw_device": "/dev/mapper/foo-test2", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "25%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } } TASK [debug] ******************************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:28 Thursday 21 July 2022 15:24:46 +0000 (0:00:00.091) 0:02:15.062 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_blkinfo": { "changed": false, "failed": false, "info": { "/dev/mapper/foo-test2": { "fstype": "ext4", "label": "", "name": "/dev/mapper/foo-test2", "size": "2.5G", "type": "lvm", "uuid": "ee7d1310-5176-4ec6-80b8-28181ffffed3" }, "/dev/nvme0n1": { "fstype": "", "label": "", "name": "/dev/nvme0n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme1n1": { "fstype": "", "label": "", "name": "/dev/nvme1n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme2n1": { "fstype": "", "label": "", "name": "/dev/nvme2n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sda": { "fstype": "LVM2_member", "label": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "irzZ3c-YIeb-eoxj-fRdj-JSvm-ls7D-iZNahz" }, "/dev/sdb": { "fstype": "", "label": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc": { "fstype": "", "label": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sr0": { "fstype": "iso9660", "label": "cidata", "name": "/dev/sr0", "size": "364K", "type": "rom", "uuid": "2022-07-21-15-22-18-00" }, "/dev/vda": { "fstype": "", "label": "", "name": "/dev/vda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vda1": { "fstype": "", "label": "", "name": "/dev/vda1", "size": "1M", "type": "partition", "uuid": "" }, "/dev/vda2": { "fstype": "vfat", "label": "", "name": "/dev/vda2", "size": "100M", "type": "partition", "uuid": "7B77-95E7" }, "/dev/vda3": { "fstype": "xfs", "label": "root", "name": "/dev/vda3", "size": "9.9G", "type": "partition", "uuid": "58dca2db-4ccd-4171-a017-dd53ad8c1c54" }, "/dev/vdb": { "fstype": "", "label": "", "name": "/dev/vdb", "size": "2G", "type": "disk", "uuid": "" }, "/dev/vdb1": { "fstype": "ext4", "label": "yumcache", "name": "/dev/vdb1", "size": "2G", "type": "partition", "uuid": "4d1ed9a1-5608-4ff8-ad89-21d95caa66fa" }, "/dev/vdc": { "fstype": "", "label": "", "name": "/dev/vdc", "size": "2G", "type": "disk", "uuid": "" }, "/dev/vdc1": { "fstype": "ext4", "label": "yumvarlib", "name": "/dev/vdc1", "size": "2G", "type": "partition", "uuid": "a7abcc97-edc5-476c-8ccd-694b4621fd28" }, "/dev/vdd": { "fstype": "", "label": "", "name": "/dev/vdd", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vde": { "fstype": "", "label": "", "name": "/dev/vde", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vdf": { "fstype": "", "label": "", "name": "/dev/vdf", "size": "10G", "type": "disk", "uuid": "" } } } } TASK [Get the size of parent/pool device] ************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:31 Thursday 21 July 2022 15:24:46 +0000 (0:00:00.055) 0:02:15.118 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "bytes": 10737418240, "changed": false, "lvm": "10g", "parted": "10GiB", "size": "10 GiB" } TASK [debug] ******************************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:36 Thursday 21 July 2022 15:24:47 +0000 (0:00:00.423) 0:02:15.541 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_pool_size": { "bytes": 10737418240, "changed": false, "failed": false, "lvm": "10g", "parted": "10GiB", "size": "10 GiB" } } TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:39 Thursday 21 July 2022 15:24:47 +0000 (0:00:00.089) 0:02:15.630 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_expected_size": "2684354560.0" }, "changed": false } TASK [debug] ******************************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:44 Thursday 21 July 2022 15:24:47 +0000 (0:00:00.092) 0:02:15.722 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_actual_size": { "bytes": 2684354560, "changed": false, "failed": false, "lvm": "2g", "parted": "2GiB", "size": "2 GiB" } } TASK [debug] ******************************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:47 Thursday 21 July 2022 15:24:47 +0000 (0:00:00.081) 0:02:15.804 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_expected_size": "2684354560.0" } TASK [assert] ****************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:50 Thursday 21 July 2022 15:24:47 +0000 (0:00:00.078) 0:02:15.883 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Get information about the LV] ******************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-cache.yml:6 Thursday 21 July 2022 15:24:47 +0000 (0:00:00.050) 0:02:15.934 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "cmd": [ "lvs", "--noheadings", "--nameprefixes", "--units=b", "--nosuffix", "--unquoted", "-o", "name,attr,cache_total_blocks,chunk_size,segtype", "foo/test2" ], "delta": "0:00:00.036002", "end": "2022-07-21 11:24:47.740499", "rc": 0, "start": "2022-07-21 11:24:47.704497" } STDOUT: LVM2_LV_NAME=test2 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-cache.yml:14 Thursday 21 July 2022 15:24:47 +0000 (0:00:00.418) 0:02:16.353 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_lv_segtype": [ "linear" ] }, "changed": false } TASK [check segment type] ****************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-cache.yml:17 Thursday 21 July 2022 15:24:47 +0000 (0:00:00.051) 0:02:16.404 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-cache.yml:22 Thursday 21 July 2022 15:24:47 +0000 (0:00:00.049) 0:02:16.453 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the requested cache size] ****************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-cache.yml:26 Thursday 21 July 2022 15:24:47 +0000 (0:00:00.037) 0:02:16.491 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-cache.yml:32 Thursday 21 July 2022 15:24:48 +0000 (0:00:00.039) 0:02:16.530 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-cache.yml:36 Thursday 21 July 2022 15:24:48 +0000 (0:00:00.037) 0:02:16.567 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume.yml:16 Thursday 21 July 2022 15:24:48 +0000 (0:00:00.039) 0:02:16.607 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [Verify the volumes with no pool were correctly managed] ****************** task path: /tmp/tmpafnwmbol/tests/verify-role-results.yml:43 Thursday 21 July 2022 15:24:48 +0000 (0:00:00.040) 0:02:16.647 ********* TASK [Clean up variable namespace] ********************************************* task path: /tmp/tmpafnwmbol/tests/verify-role-results.yml:53 Thursday 21 July 2022 15:24:48 +0000 (0:00:00.023) 0:02:16.671 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_blkinfo": null, "storage_test_crypttab": null, "storage_test_fstab": null }, "changed": false } TASK [Get the size of test2 volume again] ************************************** task path: /tmp/tmpafnwmbol/tests/tests_lvm_percent_size.yml:138 Thursday 21 July 2022 15:24:48 +0000 (0:00:00.038) 0:02:16.710 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "cmd": [ "lsblk", "--noheadings", "-o", "SIZE", "/dev/mapper/foo-test2" ], "delta": "0:00:00.004208", "end": "2022-07-21 11:24:48.485408", "rc": 0, "start": "2022-07-21 11:24:48.481200" } STDOUT: 2.5G TASK [Verify that removing test1 didn't cause a change in test2 size] ********** task path: /tmp/tmpafnwmbol/tests/tests_lvm_percent_size.yml:143 Thursday 21 July 2022 15:24:48 +0000 (0:00:00.398) 0:02:17.109 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Grow test2 using a percentage-based size spec] *************************** task path: /tmp/tmpafnwmbol/tests/tests_lvm_percent_size.yml:147 Thursday 21 July 2022 15:24:48 +0000 (0:00:00.040) 0:02:17.149 ********* TASK [linux-system-roles.storage : set platform/version specific variables] **** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main.yml:2 Thursday 21 July 2022 15:24:48 +0000 (0:00:00.063) 0:02:17.212 ********* included: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for /cache/rhel-8-y.qcow2.snap TASK [linux-system-roles.storage : Ensure ansible_facts used by role] ********** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2 Thursday 21 July 2022 15:24:48 +0000 (0:00:00.036) 0:02:17.248 ********* ok: [/cache/rhel-8-y.qcow2.snap] TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8 Thursday 21 July 2022 15:24:49 +0000 (0:00:00.531) 0:02:17.780 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [/cache/rhel-8-y.qcow2.snap] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [/cache/rhel-8-y.qcow2.snap] => (item=RedHat_8.yml) => { "ansible_facts": { "blivet_package_list": [ "python3-blivet", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap", "vdo", "kmod-kvdo", "xfsprogs" ] }, "ansible_included_var_files": [ "/tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/vars/RedHat_8.yml" ], "ansible_loop_var": "item", "changed": false, "item": "RedHat_8.yml" } skipping: [/cache/rhel-8-y.qcow2.snap] => (item=RedHat_8.7.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat_8.7.yml", "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : define an empty list of pools to be used in testing] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main.yml:5 Thursday 21 July 2022 15:24:49 +0000 (0:00:00.063) 0:02:17.844 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : define an empty list of volumes to be used in testing] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main.yml:9 Thursday 21 July 2022 15:24:49 +0000 (0:00:00.035) 0:02:17.880 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : include the appropriate provider tasks] ***** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main.yml:13 Thursday 21 July 2022 15:24:49 +0000 (0:00:00.041) 0:02:17.921 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount included: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for /cache/rhel-8-y.qcow2.snap TASK [linux-system-roles.storage : get a list of rpm packages installed on host machine] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2 Thursday 21 July 2022 15:24:49 +0000 (0:00:00.090) 0:02:18.011 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : make sure blivet is available] ************** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:7 Thursday 21 July 2022 15:24:49 +0000 (0:00:00.023) 0:02:18.034 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [linux-system-roles.storage : show storage_pools] ************************* task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:14 Thursday 21 July 2022 15:24:50 +0000 (0:00:01.328) 0:02:19.362 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_pools": [ { "disks": [ "sda" ], "name": "foo", "state": "present", "volumes": [ { "mount_point": "/opt/test2", "name": "test2", "size": "50%" } ] } ] } TASK [linux-system-roles.storage : show storage_volumes] *********************** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:19 Thursday 21 July 2022 15:24:50 +0000 (0:00:00.043) 0:02:19.406 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [linux-system-roles.storage : get required packages] ********************** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:24 Thursday 21 July 2022 15:24:50 +0000 (0:00:00.037) 0:02:19.443 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [ "lvm2" ], "pools": [], "volumes": [] } TASK [linux-system-roles.storage : enable copr repositories if needed] ********* task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:37 Thursday 21 July 2022 15:24:52 +0000 (0:00:01.522) 0:02:20.966 ********* included: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml for /cache/rhel-8-y.qcow2.snap TASK [linux-system-roles.storage : check if the COPR support packages should be installed] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:2 Thursday 21 July 2022 15:24:52 +0000 (0:00:00.048) 0:02:21.014 ********* TASK [linux-system-roles.storage : make sure COPR support packages are present] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:13 Thursday 21 July 2022 15:24:52 +0000 (0:00:00.037) 0:02:21.052 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : enable COPRs] ******************************* task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:18 Thursday 21 July 2022 15:24:52 +0000 (0:00:00.039) 0:02:21.091 ********* TASK [linux-system-roles.storage : make sure required packages are installed] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:44 Thursday 21 July 2022 15:24:52 +0000 (0:00:00.037) 0:02:21.129 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [linux-system-roles.storage : get service facts] ************************** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:51 Thursday 21 July 2022 15:24:53 +0000 (0:00:01.344) 0:02:22.474 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "services": { "NetworkManager-dispatcher.service": { "name": "NetworkManager-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "NetworkManager-wait-online.service": { "name": "NetworkManager-wait-online.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "NetworkManager.service": { "name": "NetworkManager.service", "source": "systemd", "state": "running", "status": "enabled" }, "arp-ethers.service": { "name": "arp-ethers.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "auditd.service": { "name": "auditd.service", "source": "systemd", "state": "running", "status": "enabled" }, "auth-rpcgss-module.service": { "name": "auth-rpcgss-module.service", "source": "systemd", "state": "stopped", "status": "static" }, "autovt@.service": { "name": "autovt@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "blivet.service": { "name": "blivet.service", "source": "systemd", "state": "inactive", "status": "static" }, "blk-availability.service": { "name": "blk-availability.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "chrony-dnssrv@.service": { "name": "chrony-dnssrv@.service", "source": "systemd", "state": "unknown", "status": "static" }, "chrony-wait.service": { "name": "chrony-wait.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd.service": { "name": "chronyd.service", "source": "systemd", "state": "running", "status": "enabled" }, "cloud-config.service": { "name": "cloud-config.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-final.service": { "name": "cloud-final.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init-hotplugd.service": { "name": "cloud-init-hotplugd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cloud-init-local.service": { "name": "cloud-init-local.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init.service": { "name": "cloud-init.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cockpit-motd.service": { "name": "cockpit-motd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cockpit-wsinstance-http.service": { "name": "cockpit-wsinstance-http.service", "source": "systemd", "state": "inactive", "status": "static" }, "cockpit-wsinstance-https-factory@.service": { "name": "cockpit-wsinstance-https-factory@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cockpit-wsinstance-https@.service": { "name": "cockpit-wsinstance-https@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cockpit.service": { "name": "cockpit.service", "source": "systemd", "state": "inactive", "status": "static" }, "console-getty.service": { "name": "console-getty.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "container-getty@.service": { "name": "container-getty@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cpupower.service": { "name": "cpupower.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "crond.service": { "name": "crond.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-org.freedesktop.hostname1.service": { "name": "dbus-org.freedesktop.hostname1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.locale1.service": { "name": "dbus-org.freedesktop.locale1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.login1.service": { "name": "dbus-org.freedesktop.login1.service", "source": "systemd", "state": "active", "status": "static" }, "dbus-org.freedesktop.nm-dispatcher.service": { "name": "dbus-org.freedesktop.nm-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "dbus-org.freedesktop.portable1.service": { "name": "dbus-org.freedesktop.portable1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.timedate1.service": { "name": "dbus-org.freedesktop.timedate1.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "dbus.service": { "name": "dbus.service", "source": "systemd", "state": "running", "status": "static" }, "dbxtool.service": { "name": "dbxtool.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "debug-shell.service": { "name": "debug-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "display-manager.service": { "name": "display-manager.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "dm-event.service": { "name": "dm-event.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-makecache.service": { "name": "dnf-makecache.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-cmdline.service": { "name": "dracut-cmdline.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-initqueue.service": { "name": "dracut-initqueue.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-mount.service": { "name": "dracut-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-mount.service": { "name": "dracut-pre-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-pivot.service": { "name": "dracut-pre-pivot.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-trigger.service": { "name": "dracut-pre-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-udev.service": { "name": "dracut-pre-udev.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown.service": { "name": "dracut-shutdown.service", "source": "systemd", "state": "stopped", "status": "static" }, "emergency.service": { "name": "emergency.service", "source": "systemd", "state": "stopped", "status": "static" }, "fcoe.service": { "name": "fcoe.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "fstrim.service": { "name": "fstrim.service", "source": "systemd", "state": "inactive", "status": "static" }, "getty@.service": { "name": "getty@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "getty@tty1.service": { "name": "getty@tty1.service", "source": "systemd", "state": "running", "status": "active" }, "grub-boot-indeterminate.service": { "name": "grub-boot-indeterminate.service", "source": "systemd", "state": "inactive", "status": "static" }, "gssproxy.service": { "name": "gssproxy.service", "source": "systemd", "state": "running", "status": "disabled" }, "halt-local.service": { "name": "halt-local.service", "source": "systemd", "state": "inactive", "status": "static" }, "hv_kvp_daemon.service": { "name": "hv_kvp_daemon.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "import-state.service": { "name": "import-state.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "initrd-cleanup.service": { "name": "initrd-cleanup.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-parse-etc.service": { "name": "initrd-parse-etc.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-switch-root.service": { "name": "initrd-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-udevadm-cleanup-db.service": { "name": "initrd-udevadm-cleanup-db.service", "source": "systemd", "state": "stopped", "status": "static" }, "insights-client-boot.service": { "name": "insights-client-boot.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "insights-client-results.service": { "name": "insights-client-results.service", "source": "systemd", "state": "inactive", "status": "static" }, "insights-client.service": { "name": "insights-client.service", "source": "systemd", "state": "inactive", "status": "static" }, "irqbalance.service": { "name": "irqbalance.service", "source": "systemd", "state": "running", "status": "enabled" }, "iscsi-shutdown.service": { "name": "iscsi-shutdown.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "iscsi.service": { "name": "iscsi.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "iscsid.service": { "name": "iscsid.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "kdump.service": { "name": "kdump.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "kmod-static-nodes.service": { "name": "kmod-static-nodes.service", "source": "systemd", "state": "stopped", "status": "static" }, "kvm_stat.service": { "name": "kvm_stat.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "ldconfig.service": { "name": "ldconfig.service", "source": "systemd", "state": "stopped", "status": "static" }, "loadmodules.service": { "name": "loadmodules.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "lvm2-activation-early.service": { "name": "lvm2-activation-early.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "lvm2-activation.service": { "name": "lvm2-activation.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "lvm2-lvmpolld.service": { "name": "lvm2-lvmpolld.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-monitor.service": { "name": "lvm2-monitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "lvm2-pvscan@.service": { "name": "lvm2-pvscan@.service", "source": "systemd", "state": "unknown", "status": "static" }, "lvm2-pvscan@8:0.service": { "name": "lvm2-pvscan@8:0.service", "source": "systemd", "state": "stopped", "status": "active" }, "man-db-cache-update.service": { "name": "man-db-cache-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "man-db-restart-cache-update.service": { "name": "man-db-restart-cache-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "mdadm-grow-continue@.service": { "name": "mdadm-grow-continue@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdadm-last-resort@.service": { "name": "mdadm-last-resort@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdcheck_continue.service": { "name": "mdcheck_continue.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdcheck_start.service": { "name": "mdcheck_start.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmon@.service": { "name": "mdmon@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdmonitor-oneshot.service": { "name": "mdmonitor-oneshot.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmonitor.service": { "name": "mdmonitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "messagebus.service": { "name": "messagebus.service", "source": "systemd", "state": "active", "status": "static" }, "microcode.service": { "name": "microcode.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "multipathd.service": { "name": "multipathd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "ndctl-monitor.service": { "name": "ndctl-monitor.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "network.service": { "name": "network.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "nfs-blkmap.service": { "name": "nfs-blkmap.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-convert.service": { "name": "nfs-convert.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-idmapd.service": { "name": "nfs-idmapd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-mountd.service": { "name": "nfs-mountd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-server.service": { "name": "nfs-server.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-utils.service": { "name": "nfs-utils.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfsdcld.service": { "name": "nfsdcld.service", "source": "systemd", "state": "stopped", "status": "static" }, "nis-domainname.service": { "name": "nis-domainname.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "ntpd.service": { "name": "ntpd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ntpdate.service": { "name": "ntpdate.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "oddjobd.service": { "name": "oddjobd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "packagekit-offline-update.service": { "name": "packagekit-offline-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "packagekit.service": { "name": "packagekit.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-quit-wait.service": { "name": "plymouth-quit-wait.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "plymouth-start.service": { "name": "plymouth-start.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "polkit.service": { "name": "polkit.service", "source": "systemd", "state": "running", "status": "static" }, "qemu-guest-agent.service": { "name": "qemu-guest-agent.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "quotaon.service": { "name": "quotaon.service", "source": "systemd", "state": "inactive", "status": "static" }, "rbdmap.service": { "name": "rbdmap.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "rc-local.service": { "name": "rc-local.service", "source": "systemd", "state": "stopped", "status": "static" }, "rdisc.service": { "name": "rdisc.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rescue.service": { "name": "rescue.service", "source": "systemd", "state": "stopped", "status": "static" }, "rhcd.service": { "name": "rhcd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhsm-facts.service": { "name": "rhsm-facts.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhsm.service": { "name": "rhsm.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhsmcertd.service": { "name": "rhsmcertd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpc-gssd.service": { "name": "rpc-gssd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd-notify.service": { "name": "rpc-statd-notify.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd.service": { "name": "rpc-statd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpcbind.service": { "name": "rpcbind.service", "source": "systemd", "state": "running", "status": "enabled" }, "rsyslog.service": { "name": "rsyslog.service", "source": "systemd", "state": "running", "status": "enabled" }, "selinux-autorelabel-mark.service": { "name": "selinux-autorelabel-mark.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "selinux-autorelabel.service": { "name": "selinux-autorelabel.service", "source": "systemd", "state": "inactive", "status": "static" }, "serial-getty@.service": { "name": "serial-getty@.service", "source": "systemd", "state": "unknown", "status": "indirect" }, "serial-getty@ttyS0.service": { "name": "serial-getty@ttyS0.service", "source": "systemd", "state": "running", "status": "active" }, "snapd.seeded.service": { "name": "snapd.seeded.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sntp.service": { "name": "sntp.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sshd-keygen.service": { "name": "sshd-keygen.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sshd-keygen@.service": { "name": "sshd-keygen@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@ecdsa.service": { "name": "sshd-keygen@ecdsa.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd-keygen@ed25519.service": { "name": "sshd-keygen@ed25519.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd-keygen@rsa.service": { "name": "sshd-keygen@rsa.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd.service": { "name": "sshd.service", "source": "systemd", "state": "running", "status": "enabled" }, "sshd@.service": { "name": "sshd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "sssd-autofs.service": { "name": "sssd-autofs.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-kcm.service": { "name": "sssd-kcm.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "sssd-nss.service": { "name": "sssd-nss.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pac.service": { "name": "sssd-pac.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pam.service": { "name": "sssd-pam.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-ssh.service": { "name": "sssd-ssh.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-sudo.service": { "name": "sssd-sudo.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd.service": { "name": "sssd.service", "source": "systemd", "state": "running", "status": "enabled" }, "syslog.service": { "name": "syslog.service", "source": "systemd", "state": "active", "status": "enabled" }, "system-update-cleanup.service": { "name": "system-update-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-ask-password-console.service": { "name": "systemd-ask-password-console.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-wall.service": { "name": "systemd-ask-password-wall.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-backlight@.service": { "name": "systemd-backlight@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-binfmt.service": { "name": "systemd-binfmt.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-coredump@.service": { "name": "systemd-coredump@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-exit.service": { "name": "systemd-exit.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-firstboot.service": { "name": "systemd-firstboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck-root.service": { "name": "systemd-fsck-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck@.service": { "name": "systemd-fsck@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-fsck@dev-disk-by\\x2duuid-7B77\\x2d95E7.service": { "name": "systemd-fsck@dev-disk-by\\x2duuid-7B77\\x2d95E7.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-fsck@dev-vdb1.service": { "name": "systemd-fsck@dev-vdb1.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-fsck@dev-vdc1.service": { "name": "systemd-fsck@dev-vdc1.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-halt.service": { "name": "systemd-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hibernate-resume@.service": { "name": "systemd-hibernate-resume@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-hibernate.service": { "name": "systemd-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hostnamed.service": { "name": "systemd-hostnamed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hwdb-update.service": { "name": "systemd-hwdb-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-hybrid-sleep.service": { "name": "systemd-hybrid-sleep.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-initctl.service": { "name": "systemd-initctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-catalog-update.service": { "name": "systemd-journal-catalog-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-flush.service": { "name": "systemd-journal-flush.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journald.service": { "name": "systemd-journald.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-kexec.service": { "name": "systemd-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-localed.service": { "name": "systemd-localed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-logind.service": { "name": "systemd-logind.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-machine-id-commit.service": { "name": "systemd-machine-id-commit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-modules-load.service": { "name": "systemd-modules-load.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-networkd-wait-online.service": { "name": "systemd-networkd-wait-online.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "systemd-portabled.service": { "name": "systemd-portabled.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-poweroff.service": { "name": "systemd-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-quotacheck.service": { "name": "systemd-quotacheck.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-random-seed.service": { "name": "systemd-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-reboot.service": { "name": "systemd-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-remount-fs.service": { "name": "systemd-remount-fs.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-resolved.service": { "name": "systemd-resolved.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-rfkill.service": { "name": "systemd-rfkill.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend-then-hibernate.service": { "name": "systemd-suspend-then-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend.service": { "name": "systemd-suspend.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-sysctl.service": { "name": "systemd-sysctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-sysusers.service": { "name": "systemd-sysusers.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-timedated.service": { "name": "systemd-timedated.service", "source": "systemd", "state": "inactive", "status": "masked" }, "systemd-timesyncd.service": { "name": "systemd-timesyncd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "systemd-tmpfiles-clean.service": { "name": "systemd-tmpfiles-clean.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev.service": { "name": "systemd-tmpfiles-setup-dev.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup.service": { "name": "systemd-tmpfiles-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-settle.service": { "name": "systemd-udev-settle.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-trigger.service": { "name": "systemd-udev-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udevd.service": { "name": "systemd-udevd.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-update-done.service": { "name": "systemd-update-done.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp-runlevel.service": { "name": "systemd-update-utmp-runlevel.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp.service": { "name": "systemd-update-utmp.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-user-sessions.service": { "name": "systemd-user-sessions.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-vconsole-setup.service": { "name": "systemd-vconsole-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-volatile-root.service": { "name": "systemd-volatile-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "tcsd.service": { "name": "tcsd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "teamd@.service": { "name": "teamd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "timedatex.service": { "name": "timedatex.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "tuned.service": { "name": "tuned.service", "source": "systemd", "state": "running", "status": "enabled" }, "unbound-anchor.service": { "name": "unbound-anchor.service", "source": "systemd", "state": "stopped", "status": "static" }, "user-runtime-dir@.service": { "name": "user-runtime-dir@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user-runtime-dir@0.service": { "name": "user-runtime-dir@0.service", "source": "systemd", "state": "stopped", "status": "active" }, "user@.service": { "name": "user@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user@0.service": { "name": "user@0.service", "source": "systemd", "state": "running", "status": "active" }, "vdo-start-by-dev@.service": { "name": "vdo-start-by-dev@.service", "source": "systemd", "state": "unknown", "status": "static" }, "vdo.service": { "name": "vdo.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "ypbind.service": { "name": "ypbind.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "yppasswdd.service": { "name": "yppasswdd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ypserv.service": { "name": "ypserv.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ypxfrd.service": { "name": "ypxfrd.service", "source": "systemd", "state": "stopped", "status": "not-found" } } }, "changed": false } TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************ task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:58 Thursday 21 July 2022 15:24:55 +0000 (0:00:01.738) 0:02:24.212 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] ******* task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:71 Thursday 21 July 2022 15:24:55 +0000 (0:00:00.120) 0:02:24.333 ********* TASK [linux-system-roles.storage : manage the pools and volumes to match the specified state] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77 Thursday 21 July 2022 15:24:55 +0000 (0:00:00.024) 0:02:24.358 ********* changed: [/cache/rhel-8-y.qcow2.snap] => { "actions": [ { "action": "resize device", "device": "/dev/mapper/foo-test2", "fs_type": null }, { "action": "resize format", "device": "/dev/mapper/foo-test2", "fs_type": "ext4" } ], "changed": true, "crypts": [], "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/vda2", "/dev/vda3", "/dev/mapper/foo-test2", "/dev/sdb", "/dev/sdc", "/dev/nvme0n1", "/dev/nvme1n1", "/dev/nvme2n1", "/dev/vdb1", "/dev/vdc1", "/dev/vdd", "/dev/vde", "/dev/vdf" ], "mounts": [ { "dump": 0, "fstype": "ext4", "opts": "defaults", "passno": 0, "path": "/opt/test2", "src": "/dev/mapper/foo-test2", "state": "mounted" } ], "packages": [ "e2fsprogs", "xfsprogs", "dosfstools", "lvm2" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test2", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test2", "_raw_device": "/dev/mapper/foo-test2", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "50%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } TASK [linux-system-roles.storage : Workaround for udev issue on some platforms] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:91 Thursday 21 July 2022 15:24:57 +0000 (0:00:02.086) 0:02:26.444 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] ***** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:103 Thursday 21 July 2022 15:24:57 +0000 (0:00:00.039) 0:02:26.484 ********* TASK [linux-system-roles.storage : show blivet_output] ************************* task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:109 Thursday 21 July 2022 15:24:57 +0000 (0:00:00.023) 0:02:26.507 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "blivet_output": { "actions": [ { "action": "resize device", "device": "/dev/mapper/foo-test2", "fs_type": null }, { "action": "resize format", "device": "/dev/mapper/foo-test2", "fs_type": "ext4" } ], "changed": true, "crypts": [], "failed": false, "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/vda2", "/dev/vda3", "/dev/mapper/foo-test2", "/dev/sdb", "/dev/sdc", "/dev/nvme0n1", "/dev/nvme1n1", "/dev/nvme2n1", "/dev/vdb1", "/dev/vdc1", "/dev/vdd", "/dev/vde", "/dev/vdf" ], "mounts": [ { "dump": 0, "fstype": "ext4", "opts": "defaults", "passno": 0, "path": "/opt/test2", "src": "/dev/mapper/foo-test2", "state": "mounted" } ], "packages": [ "e2fsprogs", "xfsprogs", "dosfstools", "lvm2" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test2", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test2", "_raw_device": "/dev/mapper/foo-test2", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "50%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } } TASK [linux-system-roles.storage : set the list of pools for test verification] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:114 Thursday 21 July 2022 15:24:58 +0000 (0:00:00.075) 0:02:26.583 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test2", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test2", "_raw_device": "/dev/mapper/foo-test2", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "50%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] }, "changed": false } TASK [linux-system-roles.storage : set the list of volumes for test verification] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:118 Thursday 21 July 2022 15:24:58 +0000 (0:00:00.103) 0:02:26.686 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : remove obsolete mounts] ********************* task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:134 Thursday 21 July 2022 15:24:58 +0000 (0:00:00.039) 0:02:26.726 ********* TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:146 Thursday 21 July 2022 15:24:58 +0000 (0:00:00.036) 0:02:26.763 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : set up new/current mounts] ****************** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:151 Thursday 21 July 2022 15:24:58 +0000 (0:00:00.642) 0:02:27.405 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount changed: [/cache/rhel-8-y.qcow2.snap] => (item={'src': '/dev/mapper/foo-test2', 'path': '/opt/test2', 'fstype': 'ext4', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted'}) => { "ansible_loop_var": "mount_info", "backup_file": "", "boot": "yes", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "ext4", "mount_info": { "dump": 0, "fstype": "ext4", "opts": "defaults", "passno": 0, "path": "/opt/test2", "src": "/dev/mapper/foo-test2", "state": "mounted" }, "name": "/opt/test2", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test2" } TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:163 Thursday 21 July 2022 15:24:59 +0000 (0:00:00.424) 0:02:27.830 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : retrieve facts for the /etc/crypttab file] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:171 Thursday 21 July 2022 15:24:59 +0000 (0:00:00.622) 0:02:28.453 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "stat": { "exists": false } } TASK [linux-system-roles.storage : manage /etc/crypttab to account for changes we just made] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:176 Thursday 21 July 2022 15:25:00 +0000 (0:00:00.371) 0:02:28.825 ********* TASK [linux-system-roles.storage : Update facts] ******************************* task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:198 Thursday 21 July 2022 15:25:00 +0000 (0:00:00.024) 0:02:28.849 ********* ok: [/cache/rhel-8-y.qcow2.snap] META: role_complete for /cache/rhel-8-y.qcow2.snap TASK [include_tasks] *********************************************************** task path: /tmp/tmpafnwmbol/tests/tests_lvm_percent_size.yml:160 Thursday 21 July 2022 15:25:01 +0000 (0:00:00.987) 0:02:29.836 ********* included: /tmp/tmpafnwmbol/tests/verify-role-results.yml for /cache/rhel-8-y.qcow2.snap TASK [Print out pool information] ********************************************** task path: /tmp/tmpafnwmbol/tests/verify-role-results.yml:1 Thursday 21 July 2022 15:25:01 +0000 (0:00:00.049) 0:02:29.885 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test2", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test2", "_raw_device": "/dev/mapper/foo-test2", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "50%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] } TASK [Print out volume information] ******************************************** task path: /tmp/tmpafnwmbol/tests/verify-role-results.yml:6 Thursday 21 July 2022 15:25:01 +0000 (0:00:00.052) 0:02:29.937 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => {} TASK [Collect info about the volumes.] ***************************************** task path: /tmp/tmpafnwmbol/tests/verify-role-results.yml:14 Thursday 21 July 2022 15:25:01 +0000 (0:00:00.038) 0:02:29.976 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "info": { "/dev/mapper/foo-test2": { "fstype": "ext4", "label": "", "name": "/dev/mapper/foo-test2", "size": "5G", "type": "lvm", "uuid": "ee7d1310-5176-4ec6-80b8-28181ffffed3" }, "/dev/nvme0n1": { "fstype": "", "label": "", "name": "/dev/nvme0n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme1n1": { "fstype": "", "label": "", "name": "/dev/nvme1n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme2n1": { "fstype": "", "label": "", "name": "/dev/nvme2n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sda": { "fstype": "LVM2_member", "label": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "irzZ3c-YIeb-eoxj-fRdj-JSvm-ls7D-iZNahz" }, "/dev/sdb": { "fstype": "", "label": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc": { "fstype": "", "label": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sr0": { "fstype": "iso9660", "label": "cidata", "name": "/dev/sr0", "size": "364K", "type": "rom", "uuid": "2022-07-21-15-22-18-00" }, "/dev/vda": { "fstype": "", "label": "", "name": "/dev/vda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vda1": { "fstype": "", "label": "", "name": "/dev/vda1", "size": "1M", "type": "partition", "uuid": "" }, "/dev/vda2": { "fstype": "vfat", "label": "", "name": "/dev/vda2", "size": "100M", "type": "partition", "uuid": "7B77-95E7" }, "/dev/vda3": { "fstype": "xfs", "label": "root", "name": "/dev/vda3", "size": "9.9G", "type": "partition", "uuid": "58dca2db-4ccd-4171-a017-dd53ad8c1c54" }, "/dev/vdb": { "fstype": "", "label": "", "name": "/dev/vdb", "size": "2G", "type": "disk", "uuid": "" }, "/dev/vdb1": { "fstype": "ext4", "label": "yumcache", "name": "/dev/vdb1", "size": "2G", "type": "partition", "uuid": "4d1ed9a1-5608-4ff8-ad89-21d95caa66fa" }, "/dev/vdc": { "fstype": "", "label": "", "name": "/dev/vdc", "size": "2G", "type": "disk", "uuid": "" }, "/dev/vdc1": { "fstype": "ext4", "label": "yumvarlib", "name": "/dev/vdc1", "size": "2G", "type": "partition", "uuid": "a7abcc97-edc5-476c-8ccd-694b4621fd28" }, "/dev/vdd": { "fstype": "", "label": "", "name": "/dev/vdd", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vde": { "fstype": "", "label": "", "name": "/dev/vde", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vdf": { "fstype": "", "label": "", "name": "/dev/vdf", "size": "10G", "type": "disk", "uuid": "" } } } TASK [Read the /etc/fstab file for volume existence] *************************** task path: /tmp/tmpafnwmbol/tests/verify-role-results.yml:19 Thursday 21 July 2022 15:25:01 +0000 (0:00:00.427) 0:02:30.404 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "cmd": [ "cat", "/etc/fstab" ], "delta": "0:00:00.003225", "end": "2022-07-21 11:25:02.219147", "rc": 0, "start": "2022-07-21 11:25:02.215922" } STDOUT: UUID=58dca2db-4ccd-4171-a017-dd53ad8c1c54 / xfs defaults 0 0 UUID=7B77-95E7 /boot/efi vfat defaults,uid=0,gid=0,umask=077,shortname=winnt 0 2 /dev/vdb1 /var/cache/dnf auto defaults,nofail,x-systemd.requires=cloud-init.service,_netdev,comment=cloudconfig 0 2 /dev/vdc1 /var/lib/dnf auto defaults,nofail,x-systemd.requires=cloud-init.service,_netdev,comment=cloudconfig 0 2 /dev/mapper/foo-test2 /opt/test2 ext4 defaults 0 0 TASK [Read the /etc/crypttab file] ********************************************* task path: /tmp/tmpafnwmbol/tests/verify-role-results.yml:24 Thursday 21 July 2022 15:25:02 +0000 (0:00:00.426) 0:02:30.831 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "cmd": [ "cat", "/etc/crypttab" ], "delta": "0:00:00.003220", "end": "2022-07-21 11:25:02.605508", "failed_when_result": false, "rc": 1, "start": "2022-07-21 11:25:02.602288" } STDERR: cat: /etc/crypttab: No such file or directory MSG: non-zero return code TASK [Verify the volumes listed in storage_pools were correctly managed] ******* task path: /tmp/tmpafnwmbol/tests/verify-role-results.yml:33 Thursday 21 July 2022 15:25:02 +0000 (0:00:00.385) 0:02:31.216 ********* included: /tmp/tmpafnwmbol/tests/test-verify-pool.yml for /cache/rhel-8-y.qcow2.snap => (item={'disks': ['sda'], 'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'name': 'foo', 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_metadata_version': None, 'state': 'present', 'type': 'lvm', 'volumes': [{'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'ext4', 'mount_options': 'defaults', 'mount_point': '/opt/test2', 'name': 'test2', 'raid_level': None, 'size': '50%', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test2', '_raw_device': '/dev/mapper/foo-test2', '_mount_id': '/dev/mapper/foo-test2', '_kernel_device': '/dev/dm-0', '_raw_kernel_device': '/dev/dm-0'}], 'raid_chunk_size': None}) TASK [Set _storage_pool_tests] ************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-pool.yml:5 Thursday 21 July 2022 15:25:02 +0000 (0:00:00.061) 0:02:31.278 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_pool_tests": [ "members", "volumes" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool.yml:18 Thursday 21 July 2022 15:25:02 +0000 (0:00:00.033) 0:02:31.312 ********* included: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml for /cache/rhel-8-y.qcow2.snap => (item=members) included: /tmp/tmpafnwmbol/tests/test-verify-pool-volumes.yml for /cache/rhel-8-y.qcow2.snap => (item=volumes) TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:1 Thursday 21 July 2022 15:25:02 +0000 (0:00:00.045) 0:02:31.358 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_pv_count": "1", "_storage_test_pool_pvs_lvm": [ "/dev/sda" ] }, "changed": false } TASK [Get the canonical device path for each member device] ******************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:10 Thursday 21 July 2022 15:25:02 +0000 (0:00:00.054) 0:02:31.412 ********* ok: [/cache/rhel-8-y.qcow2.snap] => (item=/dev/sda) => { "ansible_loop_var": "pv", "changed": false, "device": "/dev/sda", "pv": "/dev/sda" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:19 Thursday 21 July 2022 15:25:03 +0000 (0:00:00.406) 0:02:31.818 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "__pvs_lvm_len": "1" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:23 Thursday 21 July 2022 15:25:03 +0000 (0:00:00.049) 0:02:31.867 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_pool_pvs": [ "/dev/sda" ] }, "changed": false } TASK [Verify PV count] ********************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:27 Thursday 21 July 2022 15:25:03 +0000 (0:00:00.050) 0:02:31.917 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:34 Thursday 21 July 2022 15:25:03 +0000 (0:00:00.050) 0:02:31.968 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:38 Thursday 21 July 2022 15:25:03 +0000 (0:00:00.035) 0:02:32.003 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:42 Thursday 21 July 2022 15:25:03 +0000 (0:00:00.048) 0:02:32.052 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check the type of each PV] *********************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:46 Thursday 21 July 2022 15:25:03 +0000 (0:00:00.024) 0:02:32.077 ********* ok: [/cache/rhel-8-y.qcow2.snap] => (item=/dev/sda) => { "ansible_loop_var": "pv", "changed": false, "pv": "/dev/sda" } MSG: All assertions passed TASK [Check MD RAID] *********************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:56 Thursday 21 July 2022 15:25:03 +0000 (0:00:00.043) 0:02:32.120 ********* included: /tmp/tmpafnwmbol/tests/verify-pool-md.yml for /cache/rhel-8-y.qcow2.snap TASK [get information about RAID] ********************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-md.yml:6 Thursday 21 July 2022 15:25:03 +0000 (0:00:00.047) 0:02:32.167 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-md.yml:12 Thursday 21 July 2022 15:25:03 +0000 (0:00:00.030) 0:02:32.198 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-md.yml:16 Thursday 21 July 2022 15:25:03 +0000 (0:00:00.026) 0:02:32.224 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-md.yml:20 Thursday 21 July 2022 15:25:03 +0000 (0:00:00.026) 0:02:32.251 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-md.yml:24 Thursday 21 July 2022 15:25:03 +0000 (0:00:00.025) 0:02:32.276 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-md.yml:30 Thursday 21 July 2022 15:25:03 +0000 (0:00:00.026) 0:02:32.303 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmpafnwmbol/tests/verify-pool-md.yml:36 Thursday 21 July 2022 15:25:03 +0000 (0:00:00.025) 0:02:32.328 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-md.yml:44 Thursday 21 July 2022 15:25:03 +0000 (0:00:00.023) 0:02:32.352 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_md_active_devices_re": null, "storage_test_md_metadata_version_re": null, "storage_test_md_spare_devices_re": null }, "changed": false } TASK [Check LVM RAID] ********************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:59 Thursday 21 July 2022 15:25:03 +0000 (0:00:00.077) 0:02:32.430 ********* included: /tmp/tmpafnwmbol/tests/verify-pool-members-lvmraid.yml for /cache/rhel-8-y.qcow2.snap TASK [Validate pool member LVM RAID settings] ********************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-members-lvmraid.yml:1 Thursday 21 July 2022 15:25:03 +0000 (0:00:00.086) 0:02:32.516 ********* included: /tmp/tmpafnwmbol/tests/verify-pool-member-lvmraid.yml for /cache/rhel-8-y.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'ext4', 'mount_options': 'defaults', 'mount_point': '/opt/test2', 'name': 'test2', 'raid_level': None, 'size': '50%', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test2', '_raw_device': '/dev/mapper/foo-test2', '_mount_id': '/dev/mapper/foo-test2', '_kernel_device': '/dev/dm-0', '_raw_kernel_device': '/dev/dm-0'}) TASK [Get information about LVM RAID] ****************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-lvmraid.yml:3 Thursday 21 July 2022 15:25:04 +0000 (0:00:00.043) 0:02:32.560 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is LVM RAID] ******************************************* task path: /tmp/tmpafnwmbol/tests/verify-pool-member-lvmraid.yml:8 Thursday 21 July 2022 15:25:04 +0000 (0:00:00.027) 0:02:32.587 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-lvmraid.yml:12 Thursday 21 July 2022 15:25:04 +0000 (0:00:00.026) 0:02:32.614 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check Thin Pools] ******************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:62 Thursday 21 July 2022 15:25:04 +0000 (0:00:00.026) 0:02:32.640 ********* included: /tmp/tmpafnwmbol/tests/verify-pool-members-thin.yml for /cache/rhel-8-y.qcow2.snap TASK [Validate pool member thinpool settings] ********************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-members-thin.yml:1 Thursday 21 July 2022 15:25:04 +0000 (0:00:00.044) 0:02:32.684 ********* included: /tmp/tmpafnwmbol/tests/verify-pool-member-thin.yml for /cache/rhel-8-y.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'ext4', 'mount_options': 'defaults', 'mount_point': '/opt/test2', 'name': 'test2', 'raid_level': None, 'size': '50%', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test2', '_raw_device': '/dev/mapper/foo-test2', '_mount_id': '/dev/mapper/foo-test2', '_kernel_device': '/dev/dm-0', '_raw_kernel_device': '/dev/dm-0'}) TASK [Get information about thinpool] ****************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-thin.yml:3 Thursday 21 July 2022 15:25:04 +0000 (0:00:00.041) 0:02:32.726 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-thin.yml:8 Thursday 21 July 2022 15:25:04 +0000 (0:00:00.026) 0:02:32.752 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-thin.yml:13 Thursday 21 July 2022 15:25:04 +0000 (0:00:00.025) 0:02:32.778 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-thin.yml:17 Thursday 21 July 2022 15:25:04 +0000 (0:00:00.024) 0:02:32.803 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check member encryption] ************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:65 Thursday 21 July 2022 15:25:04 +0000 (0:00:00.022) 0:02:32.826 ********* included: /tmp/tmpafnwmbol/tests/verify-pool-members-encryption.yml for /cache/rhel-8-y.qcow2.snap TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-members-encryption.yml:4 Thursday 21 July 2022 15:25:04 +0000 (0:00:00.046) 0:02:32.872 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Validate pool member LUKS settings] ************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-members-encryption.yml:8 Thursday 21 July 2022 15:25:04 +0000 (0:00:00.049) 0:02:32.921 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => (item=/dev/sda) => { "_storage_test_pool_member_path": "/dev/sda", "ansible_loop_var": "_storage_test_pool_member_path", "changed": false, "skip_reason": "Conditional result was False" } TASK [Validate pool member crypttab entries] *********************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-members-encryption.yml:15 Thursday 21 July 2022 15:25:04 +0000 (0:00:00.029) 0:02:32.951 ********* included: /tmp/tmpafnwmbol/tests/verify-pool-member-crypttab.yml for /cache/rhel-8-y.qcow2.snap => (item=/dev/sda) TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-crypttab.yml:1 Thursday 21 July 2022 15:25:04 +0000 (0:00:00.046) 0:02:32.997 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": [] }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpafnwmbol/tests/verify-pool-member-crypttab.yml:6 Thursday 21 July 2022 15:25:04 +0000 (0:00:00.052) 0:02:33.049 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpafnwmbol/tests/verify-pool-member-crypttab.yml:11 Thursday 21 July 2022 15:25:04 +0000 (0:00:00.050) 0:02:33.100 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-crypttab.yml:17 Thursday 21 July 2022 15:25:04 +0000 (0:00:00.034) 0:02:33.135 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-crypttab.yml:23 Thursday 21 July 2022 15:25:04 +0000 (0:00:00.034) 0:02:33.169 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-crypttab.yml:29 Thursday 21 July 2022 15:25:04 +0000 (0:00:00.034) 0:02:33.203 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-members-encryption.yml:22 Thursday 21 July 2022 15:25:04 +0000 (0:00:00.035) 0:02:33.239 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_crypttab_key_file": null }, "changed": false } TASK [Check VDO] *************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:68 Thursday 21 July 2022 15:25:04 +0000 (0:00:00.031) 0:02:33.271 ********* included: /tmp/tmpafnwmbol/tests/verify-pool-members-vdo.yml for /cache/rhel-8-y.qcow2.snap TASK [Validate pool member VDO settings] *************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-members-vdo.yml:1 Thursday 21 July 2022 15:25:04 +0000 (0:00:00.046) 0:02:33.317 ********* included: /tmp/tmpafnwmbol/tests/verify-pool-member-vdo.yml for /cache/rhel-8-y.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'ext4', 'mount_options': 'defaults', 'mount_point': '/opt/test2', 'name': 'test2', 'raid_level': None, 'size': '50%', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test2', '_raw_device': '/dev/mapper/foo-test2', '_mount_id': '/dev/mapper/foo-test2', '_kernel_device': '/dev/dm-0', '_raw_kernel_device': '/dev/dm-0'}) TASK [get information about VDO deduplication] ********************************* task path: /tmp/tmpafnwmbol/tests/verify-pool-member-vdo.yml:3 Thursday 21 July 2022 15:25:04 +0000 (0:00:00.043) 0:02:33.361 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-vdo.yml:8 Thursday 21 July 2022 15:25:04 +0000 (0:00:00.022) 0:02:33.383 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-vdo.yml:11 Thursday 21 July 2022 15:25:04 +0000 (0:00:00.022) 0:02:33.405 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-vdo.yml:16 Thursday 21 July 2022 15:25:04 +0000 (0:00:00.022) 0:02:33.428 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-vdo.yml:21 Thursday 21 July 2022 15:25:04 +0000 (0:00:00.023) 0:02:33.451 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-vdo.yml:24 Thursday 21 July 2022 15:25:04 +0000 (0:00:00.023) 0:02:33.474 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-vdo.yml:29 Thursday 21 July 2022 15:25:04 +0000 (0:00:00.023) 0:02:33.498 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-member-vdo.yml:39 Thursday 21 July 2022 15:25:05 +0000 (0:00:00.025) 0:02:33.523 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:71 Thursday 21 July 2022 15:25:05 +0000 (0:00:00.035) 0:02:33.559 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "__pvs_lvm_len": null, "_storage_test_expected_pv_count": null, "_storage_test_expected_pv_type": null, "_storage_test_pool_pvs": [], "_storage_test_pool_pvs_lvm": [] }, "changed": false } TASK [verify the volumes] ****************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool-volumes.yml:3 Thursday 21 July 2022 15:25:05 +0000 (0:00:00.074) 0:02:33.633 ********* included: /tmp/tmpafnwmbol/tests/test-verify-volume.yml for /cache/rhel-8-y.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'ext4', 'mount_options': 'defaults', 'mount_point': '/opt/test2', 'name': 'test2', 'raid_level': None, 'size': '50%', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test2', '_raw_device': '/dev/mapper/foo-test2', '_mount_id': '/dev/mapper/foo-test2', '_kernel_device': '/dev/dm-0', '_raw_kernel_device': '/dev/dm-0'}) TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume.yml:2 Thursday 21 July 2022 15:25:05 +0000 (0:00:00.089) 0:02:33.723 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_volume_present": true, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume.yml:10 Thursday 21 July 2022 15:25:05 +0000 (0:00:00.052) 0:02:33.776 ********* included: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml for /cache/rhel-8-y.qcow2.snap => (item=mount) included: /tmp/tmpafnwmbol/tests/test-verify-volume-fstab.yml for /cache/rhel-8-y.qcow2.snap => (item=fstab) included: /tmp/tmpafnwmbol/tests/test-verify-volume-fs.yml for /cache/rhel-8-y.qcow2.snap => (item=fs) included: /tmp/tmpafnwmbol/tests/test-verify-volume-device.yml for /cache/rhel-8-y.qcow2.snap => (item=device) included: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml for /cache/rhel-8-y.qcow2.snap => (item=encryption) included: /tmp/tmpafnwmbol/tests/test-verify-volume-md.yml for /cache/rhel-8-y.qcow2.snap => (item=md) included: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml for /cache/rhel-8-y.qcow2.snap => (item=size) included: /tmp/tmpafnwmbol/tests/test-verify-volume-cache.yml for /cache/rhel-8-y.qcow2.snap => (item=cache) TASK [Get expected mount device based on device type] ************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml:6 Thursday 21 July 2022 15:25:05 +0000 (0:00:00.080) 0:02:33.857 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/foo-test2" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml:14 Thursday 21 July 2022 15:25:05 +0000 (0:00:00.042) 0:02:33.900 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_mount_device_matches": [ { "block_available": 1200036, "block_size": 4096, "block_total": 1269672, "block_used": 69636, "device": "/dev/mapper/foo-test2", "fstype": "ext4", "inode_available": 327669, "inode_total": 327680, "inode_used": 11, "mount": "/opt/test2", "options": "rw,seclabel,relatime", "size_available": 4915347456, "size_total": 5200576512, "uuid": "ee7d1310-5176-4ec6-80b8-28181ffffed3" } ], "storage_test_mount_expected_match_count": "1", "storage_test_mount_point_matches": [ { "block_available": 1200036, "block_size": 4096, "block_total": 1269672, "block_used": 69636, "device": "/dev/mapper/foo-test2", "fstype": "ext4", "inode_available": 327669, "inode_total": 327680, "inode_used": 11, "mount": "/opt/test2", "options": "rw,seclabel,relatime", "size_available": 4915347456, "size_total": 5200576512, "uuid": "ee7d1310-5176-4ec6-80b8-28181ffffed3" } ], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Verify the current mount state by device] ******************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml:28 Thursday 21 July 2022 15:25:05 +0000 (0:00:00.057) 0:02:33.957 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify the current mount state by mount point] *************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml:37 Thursday 21 July 2022 15:25:05 +0000 (0:00:00.053) 0:02:34.010 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify the mount fs type] ************************************************ task path: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml:45 Thursday 21 July 2022 15:25:05 +0000 (0:00:00.051) 0:02:34.062 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [command] ***************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml:54 Thursday 21 July 2022 15:25:05 +0000 (0:00:00.051) 0:02:34.113 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml:58 Thursday 21 July 2022 15:25:05 +0000 (0:00:00.023) 0:02:34.136 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml:63 Thursday 21 July 2022 15:25:05 +0000 (0:00:00.022) 0:02:34.158 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-mount.yml:75 Thursday 21 July 2022 15:25:05 +0000 (0:00:00.021) 0:02:34.180 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-fstab.yml:2 Thursday 21 July 2022 15:25:05 +0000 (0:00:00.035) 0:02:34.216 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "1", "storage_test_fstab_expected_mount_options_matches": "1", "storage_test_fstab_expected_mount_point_matches": "1", "storage_test_fstab_id_matches": [ "/dev/mapper/foo-test2 " ], "storage_test_fstab_mount_options_matches": [ " /opt/test2 ext4 defaults " ], "storage_test_fstab_mount_point_matches": [ " /opt/test2 " ] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-fstab.yml:25 Thursday 21 July 2022 15:25:05 +0000 (0:00:00.058) 0:02:34.274 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify the fstab mount point] ******************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-fstab.yml:32 Thursday 21 July 2022 15:25:05 +0000 (0:00:00.048) 0:02:34.322 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-fstab.yml:39 Thursday 21 July 2022 15:25:05 +0000 (0:00:00.055) 0:02:34.377 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-fstab.yml:49 Thursday 21 July 2022 15:25:05 +0000 (0:00:00.038) 0:02:34.415 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-fs.yml:4 Thursday 21 July 2022 15:25:05 +0000 (0:00:00.036) 0:02:34.452 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify fs label] ********************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-fs.yml:10 Thursday 21 July 2022 15:25:05 +0000 (0:00:00.040) 0:02:34.492 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [See whether the device node is present] ********************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-device.yml:4 Thursday 21 July 2022 15:25:06 +0000 (0:00:00.039) 0:02:34.532 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "stat": { "atime": 1658417097.7912602, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1658417097.7912602, "dev": 6, "device_type": 64768, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 62525, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/symlink", "mode": "0660", "mtime": 1658417097.7912602, "nlink": 1, "path": "/dev/mapper/foo-test2", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-device.yml:10 Thursday 21 July 2022 15:25:06 +0000 (0:00:00.379) 0:02:34.912 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about this volume] ********************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-device.yml:18 Thursday 21 July 2022 15:25:06 +0000 (0:00:00.040) 0:02:34.953 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [(1/2) Process volume type (set initial value)] *************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-device.yml:24 Thursday 21 July 2022 15:25:06 +0000 (0:00:00.042) 0:02:34.995 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [(2/2) Process volume type (get RAID value)] ****************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-device.yml:28 Thursday 21 July 2022 15:25:06 +0000 (0:00:00.036) 0:02:35.032 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-device.yml:33 Thursday 21 July 2022 15:25:06 +0000 (0:00:00.025) 0:02:35.057 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Stat the LUKS device, if encrypted] ************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:3 Thursday 21 July 2022 15:25:06 +0000 (0:00:00.044) 0:02:35.101 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:10 Thursday 21 July 2022 15:25:06 +0000 (0:00:00.025) 0:02:35.127 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [Collect LUKS info for this volume] *************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:15 Thursday 21 July 2022 15:25:07 +0000 (0:00:01.349) 0:02:36.477 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:21 Thursday 21 July 2022 15:25:07 +0000 (0:00:00.030) 0:02:36.507 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:30 Thursday 21 July 2022 15:25:08 +0000 (0:00:00.026) 0:02:36.534 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:38 Thursday 21 July 2022 15:25:08 +0000 (0:00:00.099) 0:02:36.634 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:44 Thursday 21 July 2022 15:25:08 +0000 (0:00:00.026) 0:02:36.660 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:49 Thursday 21 July 2022 15:25:08 +0000 (0:00:00.068) 0:02:36.728 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:55 Thursday 21 July 2022 15:25:08 +0000 (0:00:00.025) 0:02:36.753 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:61 Thursday 21 July 2022 15:25:08 +0000 (0:00:00.024) 0:02:36.777 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:67 Thursday 21 July 2022 15:25:08 +0000 (0:00:00.023) 0:02:36.801 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:74 Thursday 21 July 2022 15:25:08 +0000 (0:00:00.056) 0:02:36.858 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:79 Thursday 21 July 2022 15:25:08 +0000 (0:00:00.052) 0:02:36.911 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:85 Thursday 21 July 2022 15:25:08 +0000 (0:00:00.037) 0:02:36.948 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:91 Thursday 21 July 2022 15:25:08 +0000 (0:00:00.035) 0:02:36.984 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-encryption.yml:97 Thursday 21 July 2022 15:25:08 +0000 (0:00:00.040) 0:02:37.024 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [get information about RAID] ********************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-md.yml:7 Thursday 21 July 2022 15:25:08 +0000 (0:00:00.035) 0:02:37.059 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-md.yml:13 Thursday 21 July 2022 15:25:08 +0000 (0:00:00.040) 0:02:37.099 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-md.yml:17 Thursday 21 July 2022 15:25:08 +0000 (0:00:00.039) 0:02:37.139 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-md.yml:21 Thursday 21 July 2022 15:25:08 +0000 (0:00:00.042) 0:02:37.182 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-md.yml:25 Thursday 21 July 2022 15:25:08 +0000 (0:00:00.038) 0:02:37.220 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-md.yml:31 Thursday 21 July 2022 15:25:08 +0000 (0:00:00.038) 0:02:37.259 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-md.yml:37 Thursday 21 July 2022 15:25:08 +0000 (0:00:00.037) 0:02:37.297 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the actual size of the volume] ************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:3 Thursday 21 July 2022 15:25:08 +0000 (0:00:00.042) 0:02:37.339 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "bytes": 5368709120, "changed": false, "lvm": "5g", "parted": "5GiB", "size": "5 GiB" } TASK [parse the requested size of the volume] ********************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:9 Thursday 21 July 2022 15:25:09 +0000 (0:00:00.383) 0:02:37.722 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Establish base value for expected size] ********************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:15 Thursday 21 July 2022 15:25:09 +0000 (0:00:00.040) 0:02:37.763 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:20 Thursday 21 July 2022 15:25:09 +0000 (0:00:00.043) 0:02:37.806 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_expected_size": "2684354560.0" } TASK [debug] ******************************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:25 Thursday 21 July 2022 15:25:09 +0000 (0:00:00.038) 0:02:37.845 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_pool": { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test2", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test2", "_raw_device": "/dev/mapper/foo-test2", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "50%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } } TASK [debug] ******************************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:28 Thursday 21 July 2022 15:25:09 +0000 (0:00:00.055) 0:02:37.901 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_blkinfo": { "changed": false, "failed": false, "info": { "/dev/mapper/foo-test2": { "fstype": "ext4", "label": "", "name": "/dev/mapper/foo-test2", "size": "5G", "type": "lvm", "uuid": "ee7d1310-5176-4ec6-80b8-28181ffffed3" }, "/dev/nvme0n1": { "fstype": "", "label": "", "name": "/dev/nvme0n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme1n1": { "fstype": "", "label": "", "name": "/dev/nvme1n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme2n1": { "fstype": "", "label": "", "name": "/dev/nvme2n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sda": { "fstype": "LVM2_member", "label": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "irzZ3c-YIeb-eoxj-fRdj-JSvm-ls7D-iZNahz" }, "/dev/sdb": { "fstype": "", "label": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc": { "fstype": "", "label": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sr0": { "fstype": "iso9660", "label": "cidata", "name": "/dev/sr0", "size": "364K", "type": "rom", "uuid": "2022-07-21-15-22-18-00" }, "/dev/vda": { "fstype": "", "label": "", "name": "/dev/vda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vda1": { "fstype": "", "label": "", "name": "/dev/vda1", "size": "1M", "type": "partition", "uuid": "" }, "/dev/vda2": { "fstype": "vfat", "label": "", "name": "/dev/vda2", "size": "100M", "type": "partition", "uuid": "7B77-95E7" }, "/dev/vda3": { "fstype": "xfs", "label": "root", "name": "/dev/vda3", "size": "9.9G", "type": "partition", "uuid": "58dca2db-4ccd-4171-a017-dd53ad8c1c54" }, "/dev/vdb": { "fstype": "", "label": "", "name": "/dev/vdb", "size": "2G", "type": "disk", "uuid": "" }, "/dev/vdb1": { "fstype": "ext4", "label": "yumcache", "name": "/dev/vdb1", "size": "2G", "type": "partition", "uuid": "4d1ed9a1-5608-4ff8-ad89-21d95caa66fa" }, "/dev/vdc": { "fstype": "", "label": "", "name": "/dev/vdc", "size": "2G", "type": "disk", "uuid": "" }, "/dev/vdc1": { "fstype": "ext4", "label": "yumvarlib", "name": "/dev/vdc1", "size": "2G", "type": "partition", "uuid": "a7abcc97-edc5-476c-8ccd-694b4621fd28" }, "/dev/vdd": { "fstype": "", "label": "", "name": "/dev/vdd", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vde": { "fstype": "", "label": "", "name": "/dev/vde", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vdf": { "fstype": "", "label": "", "name": "/dev/vdf", "size": "10G", "type": "disk", "uuid": "" } } } } TASK [Get the size of parent/pool device] ************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:31 Thursday 21 July 2022 15:25:09 +0000 (0:00:00.059) 0:02:37.961 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "bytes": 10737418240, "changed": false, "lvm": "10g", "parted": "10GiB", "size": "10 GiB" } TASK [debug] ******************************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:36 Thursday 21 July 2022 15:25:09 +0000 (0:00:00.402) 0:02:38.363 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_pool_size": { "bytes": 10737418240, "changed": false, "failed": false, "lvm": "10g", "parted": "10GiB", "size": "10 GiB" } } TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:39 Thursday 21 July 2022 15:25:09 +0000 (0:00:00.053) 0:02:38.417 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_expected_size": "5368709120.0" }, "changed": false } TASK [debug] ******************************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:44 Thursday 21 July 2022 15:25:09 +0000 (0:00:00.054) 0:02:38.471 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_actual_size": { "bytes": 5368709120, "changed": false, "failed": false, "lvm": "5g", "parted": "5GiB", "size": "5 GiB" } } TASK [debug] ******************************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:47 Thursday 21 July 2022 15:25:09 +0000 (0:00:00.037) 0:02:38.509 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_expected_size": "5368709120.0" } TASK [assert] ****************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-size.yml:50 Thursday 21 July 2022 15:25:10 +0000 (0:00:00.038) 0:02:38.547 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Get information about the LV] ******************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-cache.yml:6 Thursday 21 July 2022 15:25:10 +0000 (0:00:00.054) 0:02:38.602 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "cmd": [ "lvs", "--noheadings", "--nameprefixes", "--units=b", "--nosuffix", "--unquoted", "-o", "name,attr,cache_total_blocks,chunk_size,segtype", "foo/test2" ], "delta": "0:00:00.034471", "end": "2022-07-21 11:25:10.425080", "rc": 0, "start": "2022-07-21 11:25:10.390609" } STDOUT: LVM2_LV_NAME=test2 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-cache.yml:14 Thursday 21 July 2022 15:25:10 +0000 (0:00:00.433) 0:02:39.036 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_lv_segtype": [ "linear" ] }, "changed": false } TASK [check segment type] ****************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-cache.yml:17 Thursday 21 July 2022 15:25:10 +0000 (0:00:00.093) 0:02:39.129 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-cache.yml:22 Thursday 21 July 2022 15:25:10 +0000 (0:00:00.053) 0:02:39.183 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the requested cache size] ****************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-cache.yml:26 Thursday 21 July 2022 15:25:10 +0000 (0:00:00.037) 0:02:39.220 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-cache.yml:32 Thursday 21 July 2022 15:25:10 +0000 (0:00:00.039) 0:02:39.260 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume-cache.yml:36 Thursday 21 July 2022 15:25:10 +0000 (0:00:00.038) 0:02:39.298 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-volume.yml:16 Thursday 21 July 2022 15:25:10 +0000 (0:00:00.037) 0:02:39.336 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [Verify the volumes with no pool were correctly managed] ****************** task path: /tmp/tmpafnwmbol/tests/verify-role-results.yml:43 Thursday 21 July 2022 15:25:10 +0000 (0:00:00.074) 0:02:39.410 ********* TASK [Clean up variable namespace] ********************************************* task path: /tmp/tmpafnwmbol/tests/verify-role-results.yml:53 Thursday 21 July 2022 15:25:10 +0000 (0:00:00.022) 0:02:39.433 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_blkinfo": null, "storage_test_crypttab": null, "storage_test_fstab": null }, "changed": false } TASK [Remove both of the LVM logical volumes in 'foo' created above] *********** task path: /tmp/tmpafnwmbol/tests/tests_lvm_percent_size.yml:162 Thursday 21 July 2022 15:25:10 +0000 (0:00:00.072) 0:02:39.506 ********* TASK [linux-system-roles.storage : set platform/version specific variables] **** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main.yml:2 Thursday 21 July 2022 15:25:11 +0000 (0:00:00.132) 0:02:39.638 ********* included: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for /cache/rhel-8-y.qcow2.snap TASK [linux-system-roles.storage : Ensure ansible_facts used by role] ********** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2 Thursday 21 July 2022 15:25:11 +0000 (0:00:00.037) 0:02:39.676 ********* ok: [/cache/rhel-8-y.qcow2.snap] TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8 Thursday 21 July 2022 15:25:11 +0000 (0:00:00.538) 0:02:40.214 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [/cache/rhel-8-y.qcow2.snap] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [/cache/rhel-8-y.qcow2.snap] => (item=RedHat_8.yml) => { "ansible_facts": { "blivet_package_list": [ "python3-blivet", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap", "vdo", "kmod-kvdo", "xfsprogs" ] }, "ansible_included_var_files": [ "/tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/vars/RedHat_8.yml" ], "ansible_loop_var": "item", "changed": false, "item": "RedHat_8.yml" } skipping: [/cache/rhel-8-y.qcow2.snap] => (item=RedHat_8.7.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat_8.7.yml", "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : define an empty list of pools to be used in testing] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main.yml:5 Thursday 21 July 2022 15:25:11 +0000 (0:00:00.068) 0:02:40.283 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : define an empty list of volumes to be used in testing] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main.yml:9 Thursday 21 July 2022 15:25:11 +0000 (0:00:00.036) 0:02:40.319 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : include the appropriate provider tasks] ***** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main.yml:13 Thursday 21 July 2022 15:25:11 +0000 (0:00:00.035) 0:02:40.355 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount included: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for /cache/rhel-8-y.qcow2.snap TASK [linux-system-roles.storage : get a list of rpm packages installed on host machine] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2 Thursday 21 July 2022 15:25:11 +0000 (0:00:00.049) 0:02:40.404 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : make sure blivet is available] ************** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:7 Thursday 21 July 2022 15:25:11 +0000 (0:00:00.023) 0:02:40.428 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [linux-system-roles.storage : show storage_pools] ************************* task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:14 Thursday 21 July 2022 15:25:13 +0000 (0:00:01.339) 0:02:41.767 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_pools": [ { "disks": [ "sda" ], "name": "foo", "state": "absent" } ] } TASK [linux-system-roles.storage : show storage_volumes] *********************** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:19 Thursday 21 July 2022 15:25:13 +0000 (0:00:00.038) 0:02:41.805 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [linux-system-roles.storage : get required packages] ********************** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:24 Thursday 21 July 2022 15:25:13 +0000 (0:00:00.038) 0:02:41.844 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [], "pools": [], "volumes": [] } TASK [linux-system-roles.storage : enable copr repositories if needed] ********* task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:37 Thursday 21 July 2022 15:25:14 +0000 (0:00:01.506) 0:02:43.350 ********* included: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml for /cache/rhel-8-y.qcow2.snap TASK [linux-system-roles.storage : check if the COPR support packages should be installed] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:2 Thursday 21 July 2022 15:25:14 +0000 (0:00:00.051) 0:02:43.402 ********* TASK [linux-system-roles.storage : make sure COPR support packages are present] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:13 Thursday 21 July 2022 15:25:14 +0000 (0:00:00.037) 0:02:43.440 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : enable COPRs] ******************************* task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:18 Thursday 21 July 2022 15:25:14 +0000 (0:00:00.040) 0:02:43.480 ********* TASK [linux-system-roles.storage : make sure required packages are installed] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:44 Thursday 21 July 2022 15:25:14 +0000 (0:00:00.036) 0:02:43.517 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [linux-system-roles.storage : get service facts] ************************** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:51 Thursday 21 July 2022 15:25:16 +0000 (0:00:01.302) 0:02:44.820 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "services": { "NetworkManager-dispatcher.service": { "name": "NetworkManager-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "NetworkManager-wait-online.service": { "name": "NetworkManager-wait-online.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "NetworkManager.service": { "name": "NetworkManager.service", "source": "systemd", "state": "running", "status": "enabled" }, "arp-ethers.service": { "name": "arp-ethers.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "auditd.service": { "name": "auditd.service", "source": "systemd", "state": "running", "status": "enabled" }, "auth-rpcgss-module.service": { "name": "auth-rpcgss-module.service", "source": "systemd", "state": "stopped", "status": "static" }, "autovt@.service": { "name": "autovt@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "blivet.service": { "name": "blivet.service", "source": "systemd", "state": "inactive", "status": "static" }, "blk-availability.service": { "name": "blk-availability.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "chrony-dnssrv@.service": { "name": "chrony-dnssrv@.service", "source": "systemd", "state": "unknown", "status": "static" }, "chrony-wait.service": { "name": "chrony-wait.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd.service": { "name": "chronyd.service", "source": "systemd", "state": "running", "status": "enabled" }, "cloud-config.service": { "name": "cloud-config.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-final.service": { "name": "cloud-final.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init-hotplugd.service": { "name": "cloud-init-hotplugd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cloud-init-local.service": { "name": "cloud-init-local.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init.service": { "name": "cloud-init.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cockpit-motd.service": { "name": "cockpit-motd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cockpit-wsinstance-http.service": { "name": "cockpit-wsinstance-http.service", "source": "systemd", "state": "inactive", "status": "static" }, "cockpit-wsinstance-https-factory@.service": { "name": "cockpit-wsinstance-https-factory@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cockpit-wsinstance-https@.service": { "name": "cockpit-wsinstance-https@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cockpit.service": { "name": "cockpit.service", "source": "systemd", "state": "inactive", "status": "static" }, "console-getty.service": { "name": "console-getty.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "container-getty@.service": { "name": "container-getty@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cpupower.service": { "name": "cpupower.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "crond.service": { "name": "crond.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-org.freedesktop.hostname1.service": { "name": "dbus-org.freedesktop.hostname1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.locale1.service": { "name": "dbus-org.freedesktop.locale1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.login1.service": { "name": "dbus-org.freedesktop.login1.service", "source": "systemd", "state": "active", "status": "static" }, "dbus-org.freedesktop.nm-dispatcher.service": { "name": "dbus-org.freedesktop.nm-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "dbus-org.freedesktop.portable1.service": { "name": "dbus-org.freedesktop.portable1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.timedate1.service": { "name": "dbus-org.freedesktop.timedate1.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "dbus.service": { "name": "dbus.service", "source": "systemd", "state": "running", "status": "static" }, "dbxtool.service": { "name": "dbxtool.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "debug-shell.service": { "name": "debug-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "display-manager.service": { "name": "display-manager.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "dm-event.service": { "name": "dm-event.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-makecache.service": { "name": "dnf-makecache.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-cmdline.service": { "name": "dracut-cmdline.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-initqueue.service": { "name": "dracut-initqueue.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-mount.service": { "name": "dracut-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-mount.service": { "name": "dracut-pre-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-pivot.service": { "name": "dracut-pre-pivot.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-trigger.service": { "name": "dracut-pre-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-udev.service": { "name": "dracut-pre-udev.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown.service": { "name": "dracut-shutdown.service", "source": "systemd", "state": "stopped", "status": "static" }, "emergency.service": { "name": "emergency.service", "source": "systemd", "state": "stopped", "status": "static" }, "fcoe.service": { "name": "fcoe.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "fstrim.service": { "name": "fstrim.service", "source": "systemd", "state": "inactive", "status": "static" }, "getty@.service": { "name": "getty@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "getty@tty1.service": { "name": "getty@tty1.service", "source": "systemd", "state": "running", "status": "active" }, "grub-boot-indeterminate.service": { "name": "grub-boot-indeterminate.service", "source": "systemd", "state": "inactive", "status": "static" }, "gssproxy.service": { "name": "gssproxy.service", "source": "systemd", "state": "running", "status": "disabled" }, "halt-local.service": { "name": "halt-local.service", "source": "systemd", "state": "inactive", "status": "static" }, "hv_kvp_daemon.service": { "name": "hv_kvp_daemon.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "import-state.service": { "name": "import-state.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "initrd-cleanup.service": { "name": "initrd-cleanup.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-parse-etc.service": { "name": "initrd-parse-etc.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-switch-root.service": { "name": "initrd-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-udevadm-cleanup-db.service": { "name": "initrd-udevadm-cleanup-db.service", "source": "systemd", "state": "stopped", "status": "static" }, "insights-client-boot.service": { "name": "insights-client-boot.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "insights-client-results.service": { "name": "insights-client-results.service", "source": "systemd", "state": "inactive", "status": "static" }, "insights-client.service": { "name": "insights-client.service", "source": "systemd", "state": "inactive", "status": "static" }, "irqbalance.service": { "name": "irqbalance.service", "source": "systemd", "state": "running", "status": "enabled" }, "iscsi-shutdown.service": { "name": "iscsi-shutdown.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "iscsi.service": { "name": "iscsi.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "iscsid.service": { "name": "iscsid.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "kdump.service": { "name": "kdump.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "kmod-static-nodes.service": { "name": "kmod-static-nodes.service", "source": "systemd", "state": "stopped", "status": "static" }, "kvm_stat.service": { "name": "kvm_stat.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "ldconfig.service": { "name": "ldconfig.service", "source": "systemd", "state": "stopped", "status": "static" }, "loadmodules.service": { "name": "loadmodules.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "lvm2-activation-early.service": { "name": "lvm2-activation-early.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "lvm2-activation.service": { "name": "lvm2-activation.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "lvm2-lvmpolld.service": { "name": "lvm2-lvmpolld.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-monitor.service": { "name": "lvm2-monitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "lvm2-pvscan@.service": { "name": "lvm2-pvscan@.service", "source": "systemd", "state": "unknown", "status": "static" }, "lvm2-pvscan@8:0.service": { "name": "lvm2-pvscan@8:0.service", "source": "systemd", "state": "stopped", "status": "active" }, "man-db-cache-update.service": { "name": "man-db-cache-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "man-db-restart-cache-update.service": { "name": "man-db-restart-cache-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "mdadm-grow-continue@.service": { "name": "mdadm-grow-continue@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdadm-last-resort@.service": { "name": "mdadm-last-resort@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdcheck_continue.service": { "name": "mdcheck_continue.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdcheck_start.service": { "name": "mdcheck_start.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmon@.service": { "name": "mdmon@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdmonitor-oneshot.service": { "name": "mdmonitor-oneshot.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmonitor.service": { "name": "mdmonitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "messagebus.service": { "name": "messagebus.service", "source": "systemd", "state": "active", "status": "static" }, "microcode.service": { "name": "microcode.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "multipathd.service": { "name": "multipathd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "ndctl-monitor.service": { "name": "ndctl-monitor.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "network.service": { "name": "network.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "nfs-blkmap.service": { "name": "nfs-blkmap.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-convert.service": { "name": "nfs-convert.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-idmapd.service": { "name": "nfs-idmapd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-mountd.service": { "name": "nfs-mountd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-server.service": { "name": "nfs-server.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-utils.service": { "name": "nfs-utils.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfsdcld.service": { "name": "nfsdcld.service", "source": "systemd", "state": "stopped", "status": "static" }, "nis-domainname.service": { "name": "nis-domainname.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "ntpd.service": { "name": "ntpd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ntpdate.service": { "name": "ntpdate.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "oddjobd.service": { "name": "oddjobd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "packagekit-offline-update.service": { "name": "packagekit-offline-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "packagekit.service": { "name": "packagekit.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-quit-wait.service": { "name": "plymouth-quit-wait.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "plymouth-start.service": { "name": "plymouth-start.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "polkit.service": { "name": "polkit.service", "source": "systemd", "state": "running", "status": "static" }, "qemu-guest-agent.service": { "name": "qemu-guest-agent.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "quotaon.service": { "name": "quotaon.service", "source": "systemd", "state": "inactive", "status": "static" }, "rbdmap.service": { "name": "rbdmap.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "rc-local.service": { "name": "rc-local.service", "source": "systemd", "state": "stopped", "status": "static" }, "rdisc.service": { "name": "rdisc.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rescue.service": { "name": "rescue.service", "source": "systemd", "state": "stopped", "status": "static" }, "rhcd.service": { "name": "rhcd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhsm-facts.service": { "name": "rhsm-facts.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhsm.service": { "name": "rhsm.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhsmcertd.service": { "name": "rhsmcertd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpc-gssd.service": { "name": "rpc-gssd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd-notify.service": { "name": "rpc-statd-notify.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd.service": { "name": "rpc-statd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpcbind.service": { "name": "rpcbind.service", "source": "systemd", "state": "running", "status": "enabled" }, "rsyslog.service": { "name": "rsyslog.service", "source": "systemd", "state": "running", "status": "enabled" }, "selinux-autorelabel-mark.service": { "name": "selinux-autorelabel-mark.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "selinux-autorelabel.service": { "name": "selinux-autorelabel.service", "source": "systemd", "state": "inactive", "status": "static" }, "serial-getty@.service": { "name": "serial-getty@.service", "source": "systemd", "state": "unknown", "status": "indirect" }, "serial-getty@ttyS0.service": { "name": "serial-getty@ttyS0.service", "source": "systemd", "state": "running", "status": "active" }, "snapd.seeded.service": { "name": "snapd.seeded.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sntp.service": { "name": "sntp.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sshd-keygen.service": { "name": "sshd-keygen.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sshd-keygen@.service": { "name": "sshd-keygen@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@ecdsa.service": { "name": "sshd-keygen@ecdsa.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd-keygen@ed25519.service": { "name": "sshd-keygen@ed25519.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd-keygen@rsa.service": { "name": "sshd-keygen@rsa.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd.service": { "name": "sshd.service", "source": "systemd", "state": "running", "status": "enabled" }, "sshd@.service": { "name": "sshd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "sssd-autofs.service": { "name": "sssd-autofs.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-kcm.service": { "name": "sssd-kcm.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "sssd-nss.service": { "name": "sssd-nss.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pac.service": { "name": "sssd-pac.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pam.service": { "name": "sssd-pam.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-ssh.service": { "name": "sssd-ssh.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-sudo.service": { "name": "sssd-sudo.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd.service": { "name": "sssd.service", "source": "systemd", "state": "running", "status": "enabled" }, "syslog.service": { "name": "syslog.service", "source": "systemd", "state": "active", "status": "enabled" }, "system-update-cleanup.service": { "name": "system-update-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-ask-password-console.service": { "name": "systemd-ask-password-console.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-wall.service": { "name": "systemd-ask-password-wall.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-backlight@.service": { "name": "systemd-backlight@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-binfmt.service": { "name": "systemd-binfmt.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-coredump@.service": { "name": "systemd-coredump@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-exit.service": { "name": "systemd-exit.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-firstboot.service": { "name": "systemd-firstboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck-root.service": { "name": "systemd-fsck-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck@.service": { "name": "systemd-fsck@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-fsck@dev-disk-by\\x2duuid-7B77\\x2d95E7.service": { "name": "systemd-fsck@dev-disk-by\\x2duuid-7B77\\x2d95E7.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-fsck@dev-vdb1.service": { "name": "systemd-fsck@dev-vdb1.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-fsck@dev-vdc1.service": { "name": "systemd-fsck@dev-vdc1.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-halt.service": { "name": "systemd-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hibernate-resume@.service": { "name": "systemd-hibernate-resume@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-hibernate.service": { "name": "systemd-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hostnamed.service": { "name": "systemd-hostnamed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hwdb-update.service": { "name": "systemd-hwdb-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-hybrid-sleep.service": { "name": "systemd-hybrid-sleep.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-initctl.service": { "name": "systemd-initctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-catalog-update.service": { "name": "systemd-journal-catalog-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-flush.service": { "name": "systemd-journal-flush.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journald.service": { "name": "systemd-journald.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-kexec.service": { "name": "systemd-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-localed.service": { "name": "systemd-localed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-logind.service": { "name": "systemd-logind.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-machine-id-commit.service": { "name": "systemd-machine-id-commit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-modules-load.service": { "name": "systemd-modules-load.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-networkd-wait-online.service": { "name": "systemd-networkd-wait-online.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "systemd-portabled.service": { "name": "systemd-portabled.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-poweroff.service": { "name": "systemd-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-quotacheck.service": { "name": "systemd-quotacheck.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-random-seed.service": { "name": "systemd-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-reboot.service": { "name": "systemd-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-remount-fs.service": { "name": "systemd-remount-fs.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-resolved.service": { "name": "systemd-resolved.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-rfkill.service": { "name": "systemd-rfkill.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend-then-hibernate.service": { "name": "systemd-suspend-then-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend.service": { "name": "systemd-suspend.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-sysctl.service": { "name": "systemd-sysctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-sysusers.service": { "name": "systemd-sysusers.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-timedated.service": { "name": "systemd-timedated.service", "source": "systemd", "state": "inactive", "status": "masked" }, "systemd-timesyncd.service": { "name": "systemd-timesyncd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "systemd-tmpfiles-clean.service": { "name": "systemd-tmpfiles-clean.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev.service": { "name": "systemd-tmpfiles-setup-dev.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup.service": { "name": "systemd-tmpfiles-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-settle.service": { "name": "systemd-udev-settle.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-trigger.service": { "name": "systemd-udev-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udevd.service": { "name": "systemd-udevd.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-update-done.service": { "name": "systemd-update-done.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp-runlevel.service": { "name": "systemd-update-utmp-runlevel.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp.service": { "name": "systemd-update-utmp.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-user-sessions.service": { "name": "systemd-user-sessions.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-vconsole-setup.service": { "name": "systemd-vconsole-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-volatile-root.service": { "name": "systemd-volatile-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "tcsd.service": { "name": "tcsd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "teamd@.service": { "name": "teamd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "timedatex.service": { "name": "timedatex.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "tuned.service": { "name": "tuned.service", "source": "systemd", "state": "running", "status": "enabled" }, "unbound-anchor.service": { "name": "unbound-anchor.service", "source": "systemd", "state": "stopped", "status": "static" }, "user-runtime-dir@.service": { "name": "user-runtime-dir@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user-runtime-dir@0.service": { "name": "user-runtime-dir@0.service", "source": "systemd", "state": "stopped", "status": "active" }, "user@.service": { "name": "user@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user@0.service": { "name": "user@0.service", "source": "systemd", "state": "running", "status": "active" }, "vdo-start-by-dev@.service": { "name": "vdo-start-by-dev@.service", "source": "systemd", "state": "unknown", "status": "static" }, "vdo.service": { "name": "vdo.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "ypbind.service": { "name": "ypbind.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "yppasswdd.service": { "name": "yppasswdd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ypserv.service": { "name": "ypserv.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ypxfrd.service": { "name": "ypxfrd.service", "source": "systemd", "state": "stopped", "status": "not-found" } } }, "changed": false } TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************ task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:58 Thursday 21 July 2022 15:25:18 +0000 (0:00:01.750) 0:02:46.570 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] ******* task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:71 Thursday 21 July 2022 15:25:18 +0000 (0:00:00.112) 0:02:46.683 ********* TASK [linux-system-roles.storage : manage the pools and volumes to match the specified state] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77 Thursday 21 July 2022 15:25:18 +0000 (0:00:00.065) 0:02:46.748 ********* changed: [/cache/rhel-8-y.qcow2.snap] => { "actions": [ { "action": "destroy format", "device": "/dev/mapper/foo-test2", "fs_type": "ext4" }, { "action": "destroy device", "device": "/dev/mapper/foo-test2", "fs_type": null }, { "action": "destroy device", "device": "/dev/foo", "fs_type": null }, { "action": "destroy format", "device": "/dev/sda", "fs_type": "lvmpv" } ], "changed": true, "crypts": [], "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/vda2", "/dev/vda3", "/dev/sda", "/dev/sdb", "/dev/sdc", "/dev/nvme0n1", "/dev/nvme1n1", "/dev/nvme2n1", "/dev/vdb1", "/dev/vdc1", "/dev/vdd", "/dev/vde", "/dev/vdf" ], "mounts": [ { "fstype": "ext4", "path": "/opt/test2", "src": "/dev/mapper/foo-test2", "state": "absent" } ], "packages": [ "e2fsprogs", "dosfstools", "xfsprogs" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "absent", "type": "lvm", "volumes": [] } ], "volumes": [] } TASK [linux-system-roles.storage : Workaround for udev issue on some platforms] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:91 Thursday 21 July 2022 15:25:20 +0000 (0:00:02.143) 0:02:48.891 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] ***** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:103 Thursday 21 July 2022 15:25:20 +0000 (0:00:00.036) 0:02:48.928 ********* TASK [linux-system-roles.storage : show blivet_output] ************************* task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:109 Thursday 21 July 2022 15:25:20 +0000 (0:00:00.021) 0:02:48.949 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "blivet_output": { "actions": [ { "action": "destroy format", "device": "/dev/mapper/foo-test2", "fs_type": "ext4" }, { "action": "destroy device", "device": "/dev/mapper/foo-test2", "fs_type": null }, { "action": "destroy device", "device": "/dev/foo", "fs_type": null }, { "action": "destroy format", "device": "/dev/sda", "fs_type": "lvmpv" } ], "changed": true, "crypts": [], "failed": false, "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/vda2", "/dev/vda3", "/dev/sda", "/dev/sdb", "/dev/sdc", "/dev/nvme0n1", "/dev/nvme1n1", "/dev/nvme2n1", "/dev/vdb1", "/dev/vdc1", "/dev/vdd", "/dev/vde", "/dev/vdf" ], "mounts": [ { "fstype": "ext4", "path": "/opt/test2", "src": "/dev/mapper/foo-test2", "state": "absent" } ], "packages": [ "e2fsprogs", "dosfstools", "xfsprogs" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "absent", "type": "lvm", "volumes": [] } ], "volumes": [] } } TASK [linux-system-roles.storage : set the list of pools for test verification] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:114 Thursday 21 July 2022 15:25:20 +0000 (0:00:00.038) 0:02:48.988 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "absent", "type": "lvm", "volumes": [] } ] }, "changed": false } TASK [linux-system-roles.storage : set the list of volumes for test verification] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:118 Thursday 21 July 2022 15:25:20 +0000 (0:00:00.036) 0:02:49.024 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : remove obsolete mounts] ********************* task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:134 Thursday 21 July 2022 15:25:20 +0000 (0:00:00.038) 0:02:49.063 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount changed: [/cache/rhel-8-y.qcow2.snap] => (item={'src': '/dev/mapper/foo-test2', 'path': '/opt/test2', 'state': 'absent', 'fstype': 'ext4'}) => { "ansible_loop_var": "mount_info", "backup_file": "", "boot": "yes", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "ext4", "mount_info": { "fstype": "ext4", "path": "/opt/test2", "src": "/dev/mapper/foo-test2", "state": "absent" }, "name": "/opt/test2", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test2" } TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:146 Thursday 21 July 2022 15:25:20 +0000 (0:00:00.408) 0:02:49.472 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : set up new/current mounts] ****************** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:151 Thursday 21 July 2022 15:25:21 +0000 (0:00:00.627) 0:02:50.099 ********* TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:163 Thursday 21 July 2022 15:25:21 +0000 (0:00:00.040) 0:02:50.139 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : retrieve facts for the /etc/crypttab file] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:171 Thursday 21 July 2022 15:25:22 +0000 (0:00:00.632) 0:02:50.772 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "stat": { "exists": false } } TASK [linux-system-roles.storage : manage /etc/crypttab to account for changes we just made] *** task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:176 Thursday 21 July 2022 15:25:22 +0000 (0:00:00.384) 0:02:51.156 ********* TASK [linux-system-roles.storage : Update facts] ******************************* task path: /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:198 Thursday 21 July 2022 15:25:22 +0000 (0:00:00.023) 0:02:51.180 ********* ok: [/cache/rhel-8-y.qcow2.snap] META: role_complete for /cache/rhel-8-y.qcow2.snap TASK [include_tasks] *********************************************************** task path: /tmp/tmpafnwmbol/tests/tests_lvm_percent_size.yml:171 Thursday 21 July 2022 15:25:23 +0000 (0:00:00.993) 0:02:52.174 ********* included: /tmp/tmpafnwmbol/tests/verify-role-results.yml for /cache/rhel-8-y.qcow2.snap TASK [Print out pool information] ********************************************** task path: /tmp/tmpafnwmbol/tests/verify-role-results.yml:1 Thursday 21 July 2022 15:25:23 +0000 (0:00:00.051) 0:02:52.226 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "absent", "type": "lvm", "volumes": [] } ] } TASK [Print out volume information] ******************************************** task path: /tmp/tmpafnwmbol/tests/verify-role-results.yml:6 Thursday 21 July 2022 15:25:23 +0000 (0:00:00.120) 0:02:52.346 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => {} TASK [Collect info about the volumes.] ***************************************** task path: /tmp/tmpafnwmbol/tests/verify-role-results.yml:14 Thursday 21 July 2022 15:25:23 +0000 (0:00:00.037) 0:02:52.383 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "info": { "/dev/nvme0n1": { "fstype": "", "label": "", "name": "/dev/nvme0n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme1n1": { "fstype": "", "label": "", "name": "/dev/nvme1n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme2n1": { "fstype": "", "label": "", "name": "/dev/nvme2n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sda": { "fstype": "", "label": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdb": { "fstype": "", "label": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc": { "fstype": "", "label": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sr0": { "fstype": "iso9660", "label": "cidata", "name": "/dev/sr0", "size": "364K", "type": "rom", "uuid": "2022-07-21-15-22-18-00" }, "/dev/vda": { "fstype": "", "label": "", "name": "/dev/vda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vda1": { "fstype": "", "label": "", "name": "/dev/vda1", "size": "1M", "type": "partition", "uuid": "" }, "/dev/vda2": { "fstype": "vfat", "label": "", "name": "/dev/vda2", "size": "100M", "type": "partition", "uuid": "7B77-95E7" }, "/dev/vda3": { "fstype": "xfs", "label": "root", "name": "/dev/vda3", "size": "9.9G", "type": "partition", "uuid": "58dca2db-4ccd-4171-a017-dd53ad8c1c54" }, "/dev/vdb": { "fstype": "", "label": "", "name": "/dev/vdb", "size": "2G", "type": "disk", "uuid": "" }, "/dev/vdb1": { "fstype": "ext4", "label": "yumcache", "name": "/dev/vdb1", "size": "2G", "type": "partition", "uuid": "4d1ed9a1-5608-4ff8-ad89-21d95caa66fa" }, "/dev/vdc": { "fstype": "", "label": "", "name": "/dev/vdc", "size": "2G", "type": "disk", "uuid": "" }, "/dev/vdc1": { "fstype": "ext4", "label": "yumvarlib", "name": "/dev/vdc1", "size": "2G", "type": "partition", "uuid": "a7abcc97-edc5-476c-8ccd-694b4621fd28" }, "/dev/vdd": { "fstype": "", "label": "", "name": "/dev/vdd", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vde": { "fstype": "", "label": "", "name": "/dev/vde", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vdf": { "fstype": "", "label": "", "name": "/dev/vdf", "size": "10G", "type": "disk", "uuid": "" } } } TASK [Read the /etc/fstab file for volume existence] *************************** task path: /tmp/tmpafnwmbol/tests/verify-role-results.yml:19 Thursday 21 July 2022 15:25:24 +0000 (0:00:00.386) 0:02:52.770 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "cmd": [ "cat", "/etc/fstab" ], "delta": "0:00:01.004092", "end": "2022-07-21 11:25:25.536034", "rc": 0, "start": "2022-07-21 11:25:24.531942" } STDOUT: UUID=58dca2db-4ccd-4171-a017-dd53ad8c1c54 / xfs defaults 0 0 UUID=7B77-95E7 /boot/efi vfat defaults,uid=0,gid=0,umask=077,shortname=winnt 0 2 /dev/vdb1 /var/cache/dnf auto defaults,nofail,x-systemd.requires=cloud-init.service,_netdev,comment=cloudconfig 0 2 /dev/vdc1 /var/lib/dnf auto defaults,nofail,x-systemd.requires=cloud-init.service,_netdev,comment=cloudconfig 0 2 TASK [Read the /etc/crypttab file] ********************************************* task path: /tmp/tmpafnwmbol/tests/verify-role-results.yml:24 Thursday 21 July 2022 15:25:25 +0000 (0:00:01.379) 0:02:54.149 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "cmd": [ "cat", "/etc/crypttab" ], "delta": "0:00:00.003218", "end": "2022-07-21 11:25:25.923197", "failed_when_result": false, "rc": 1, "start": "2022-07-21 11:25:25.919979" } STDERR: cat: /etc/crypttab: No such file or directory MSG: non-zero return code TASK [Verify the volumes listed in storage_pools were correctly managed] ******* task path: /tmp/tmpafnwmbol/tests/verify-role-results.yml:33 Thursday 21 July 2022 15:25:26 +0000 (0:00:00.387) 0:02:54.536 ********* included: /tmp/tmpafnwmbol/tests/test-verify-pool.yml for /cache/rhel-8-y.qcow2.snap => (item={'disks': ['sda'], 'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'name': 'foo', 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_metadata_version': None, 'state': 'absent', 'type': 'lvm', 'volumes': [], 'raid_chunk_size': None}) TASK [Set _storage_pool_tests] ************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-pool.yml:5 Thursday 21 July 2022 15:25:26 +0000 (0:00:00.059) 0:02:54.596 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_pool_tests": [ "members", "volumes" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool.yml:18 Thursday 21 July 2022 15:25:26 +0000 (0:00:00.033) 0:02:54.629 ********* included: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml for /cache/rhel-8-y.qcow2.snap => (item=members) included: /tmp/tmpafnwmbol/tests/test-verify-pool-volumes.yml for /cache/rhel-8-y.qcow2.snap => (item=volumes) TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:1 Thursday 21 July 2022 15:25:26 +0000 (0:00:00.045) 0:02:54.675 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_pv_count": "0", "_storage_test_pool_pvs_lvm": [] }, "changed": false } TASK [Get the canonical device path for each member device] ******************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:10 Thursday 21 July 2022 15:25:26 +0000 (0:00:00.050) 0:02:54.725 ********* TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:19 Thursday 21 July 2022 15:25:26 +0000 (0:00:00.030) 0:02:54.756 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "__pvs_lvm_len": "0" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:23 Thursday 21 July 2022 15:25:26 +0000 (0:00:00.055) 0:02:54.812 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_pool_pvs": [] }, "changed": false } TASK [Verify PV count] ********************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:27 Thursday 21 July 2022 15:25:26 +0000 (0:00:00.049) 0:02:54.861 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:34 Thursday 21 July 2022 15:25:26 +0000 (0:00:00.050) 0:02:54.912 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:38 Thursday 21 July 2022 15:25:26 +0000 (0:00:00.043) 0:02:54.955 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:42 Thursday 21 July 2022 15:25:26 +0000 (0:00:00.052) 0:02:55.007 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check the type of each PV] *********************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:46 Thursday 21 July 2022 15:25:26 +0000 (0:00:00.026) 0:02:55.034 ********* TASK [Check MD RAID] *********************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:56 Thursday 21 July 2022 15:25:26 +0000 (0:00:00.021) 0:02:55.055 ********* included: /tmp/tmpafnwmbol/tests/verify-pool-md.yml for /cache/rhel-8-y.qcow2.snap TASK [get information about RAID] ********************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-md.yml:6 Thursday 21 July 2022 15:25:26 +0000 (0:00:00.041) 0:02:55.097 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-md.yml:12 Thursday 21 July 2022 15:25:26 +0000 (0:00:00.023) 0:02:55.121 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-md.yml:16 Thursday 21 July 2022 15:25:26 +0000 (0:00:00.024) 0:02:55.145 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-md.yml:20 Thursday 21 July 2022 15:25:26 +0000 (0:00:00.026) 0:02:55.171 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-md.yml:24 Thursday 21 July 2022 15:25:26 +0000 (0:00:00.025) 0:02:55.197 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-md.yml:30 Thursday 21 July 2022 15:25:26 +0000 (0:00:00.023) 0:02:55.220 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmpafnwmbol/tests/verify-pool-md.yml:36 Thursday 21 July 2022 15:25:26 +0000 (0:00:00.023) 0:02:55.244 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-md.yml:44 Thursday 21 July 2022 15:25:26 +0000 (0:00:00.024) 0:02:55.269 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_md_active_devices_re": null, "storage_test_md_metadata_version_re": null, "storage_test_md_spare_devices_re": null }, "changed": false } TASK [Check LVM RAID] ********************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:59 Thursday 21 July 2022 15:25:26 +0000 (0:00:00.033) 0:02:55.302 ********* included: /tmp/tmpafnwmbol/tests/verify-pool-members-lvmraid.yml for /cache/rhel-8-y.qcow2.snap TASK [Validate pool member LVM RAID settings] ********************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-members-lvmraid.yml:1 Thursday 21 July 2022 15:25:26 +0000 (0:00:00.045) 0:02:55.348 ********* TASK [Check Thin Pools] ******************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:62 Thursday 21 July 2022 15:25:26 +0000 (0:00:00.064) 0:02:55.413 ********* included: /tmp/tmpafnwmbol/tests/verify-pool-members-thin.yml for /cache/rhel-8-y.qcow2.snap TASK [Validate pool member thinpool settings] ********************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-members-thin.yml:1 Thursday 21 July 2022 15:25:26 +0000 (0:00:00.056) 0:02:55.470 ********* TASK [Check member encryption] ************************************************* task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:65 Thursday 21 July 2022 15:25:26 +0000 (0:00:00.027) 0:02:55.498 ********* included: /tmp/tmpafnwmbol/tests/verify-pool-members-encryption.yml for /cache/rhel-8-y.qcow2.snap TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-members-encryption.yml:4 Thursday 21 July 2022 15:25:27 +0000 (0:00:00.053) 0:02:55.551 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Validate pool member LUKS settings] ************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-members-encryption.yml:8 Thursday 21 July 2022 15:25:27 +0000 (0:00:00.051) 0:02:55.603 ********* TASK [Validate pool member crypttab entries] *********************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-members-encryption.yml:15 Thursday 21 July 2022 15:25:27 +0000 (0:00:00.024) 0:02:55.628 ********* TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-members-encryption.yml:22 Thursday 21 July 2022 15:25:27 +0000 (0:00:00.022) 0:02:55.650 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_crypttab_key_file": null }, "changed": false } TASK [Check VDO] *************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:68 Thursday 21 July 2022 15:25:27 +0000 (0:00:00.034) 0:02:55.685 ********* included: /tmp/tmpafnwmbol/tests/verify-pool-members-vdo.yml for /cache/rhel-8-y.qcow2.snap TASK [Validate pool member VDO settings] *************************************** task path: /tmp/tmpafnwmbol/tests/verify-pool-members-vdo.yml:1 Thursday 21 July 2022 15:25:27 +0000 (0:00:00.047) 0:02:55.732 ********* TASK [set_fact] **************************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool-members.yml:71 Thursday 21 July 2022 15:25:27 +0000 (0:00:00.021) 0:02:55.754 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "__pvs_lvm_len": null, "_storage_test_expected_pv_count": null, "_storage_test_expected_pv_type": null, "_storage_test_pool_pvs": [], "_storage_test_pool_pvs_lvm": [] }, "changed": false } TASK [verify the volumes] ****************************************************** task path: /tmp/tmpafnwmbol/tests/test-verify-pool-volumes.yml:3 Thursday 21 July 2022 15:25:27 +0000 (0:00:00.035) 0:02:55.789 ********* TASK [Verify the volumes with no pool were correctly managed] ****************** task path: /tmp/tmpafnwmbol/tests/verify-role-results.yml:43 Thursday 21 July 2022 15:25:27 +0000 (0:00:00.021) 0:02:55.811 ********* TASK [Clean up variable namespace] ********************************************* task path: /tmp/tmpafnwmbol/tests/verify-role-results.yml:53 Thursday 21 July 2022 15:25:27 +0000 (0:00:00.023) 0:02:55.834 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_blkinfo": null, "storage_test_crypttab": null, "storage_test_fstab": null }, "changed": false } META: ran handlers META: ran handlers PLAY RECAP ********************************************************************* /cache/rhel-8-y.qcow2.snap : ok=795 changed=10 unreachable=0 failed=1 skipped=565 rescued=1 ignored=0 Thursday 21 July 2022 15:25:27 +0000 (0:00:00.058) 0:02:55.893 ********* =============================================================================== linux-system-roles.storage : get service facts -------------------------- 2.75s /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:51 linux-system-roles.storage : manage the pools and volumes to match the specified state --- 2.29s /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77 linux-system-roles.storage : manage the pools and volumes to match the specified state --- 2.29s /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77 linux-system-roles.storage : manage the pools and volumes to match the specified state --- 2.27s /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77 linux-system-roles.storage : manage the pools and volumes to match the specified state --- 2.14s /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77 linux-system-roles.storage : manage the pools and volumes to match the specified state --- 2.09s /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77 linux-system-roles.storage : get service facts -------------------------- 1.97s /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:51 linux-system-roles.storage : make sure blivet is available -------------- 1.81s /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:7 linux-system-roles.storage : get service facts -------------------------- 1.75s /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:51 linux-system-roles.storage : manage the pools and volumes to match the specified state --- 1.75s /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77 linux-system-roles.storage : get service facts -------------------------- 1.75s /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:51 linux-system-roles.storage : get service facts -------------------------- 1.74s /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:51 linux-system-roles.storage : get service facts -------------------------- 1.72s /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:51 linux-system-roles.storage : get service facts -------------------------- 1.72s /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:51 linux-system-roles.storage : get service facts -------------------------- 1.70s /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:51 linux-system-roles.storage : get required packages ---------------------- 1.61s /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:24 linux-system-roles.storage : get required packages ---------------------- 1.57s /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:24 linux-system-roles.storage : get required packages ---------------------- 1.57s /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:24 linux-system-roles.storage : get required packages ---------------------- 1.52s /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:24 linux-system-roles.storage : get required packages ---------------------- 1.51s /tmp/tmpafnwmbol/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:24 ansible-playbook [core 2.12.6] config file = /etc/ansible/ansible.cfg configured module search path = ['/root/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules'] ansible python module location = /usr/lib/python3.9/site-packages/ansible ansible collection location = /tmp/tmpp_vyhrdp executable location = /usr/bin/ansible-playbook python version = 3.9.13 (main, May 18 2022, 00:00:00) [GCC 11.3.1 20220421 (Red Hat 11.3.1-2)] jinja version = 2.11.3 libyaml = True Using /etc/ansible/ansible.cfg as config file Skipping callback 'debug', as we already have a stdout callback. Skipping callback 'default', as we already have a stdout callback. Skipping callback 'minimal', as we already have a stdout callback. Skipping callback 'oneline', as we already have a stdout callback. PLAYBOOK: tests_lvm_percent_size.yml ******************************************* 1 plays in /tmp/tmpmnqeja9w/tests/storage/tests_lvm_percent_size.yml PLAY [all] ********************************************************************* TASK [Gathering Facts] ********************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/tests_lvm_percent_size.yml:2 Thursday 21 July 2022 19:04:42 +0000 (0:00:00.013) 0:00:00.013 ********* ok: [/cache/rhel-8-y.qcow2.snap] META: ran handlers TASK [include_role : fedora.linux_system_roles.storage] ************************ task path: /tmp/tmpmnqeja9w/tests/storage/tests_lvm_percent_size.yml:17 Thursday 21 July 2022 19:04:43 +0000 (0:00:01.384) 0:00:01.398 ********* TASK [fedora.linux_system_roles.storage : set platform/version specific variables] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2 Thursday 21 July 2022 19:04:43 +0000 (0:00:00.034) 0:00:01.433 ********* included: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for /cache/rhel-8-y.qcow2.snap TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2 Thursday 21 July 2022 19:04:43 +0000 (0:00:00.033) 0:00:01.466 ********* ok: [/cache/rhel-8-y.qcow2.snap] TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7 Thursday 21 July 2022 19:04:44 +0000 (0:00:00.523) 0:00:01.989 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [/cache/rhel-8-y.qcow2.snap] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [/cache/rhel-8-y.qcow2.snap] => (item=RedHat_8.yml) => { "ansible_facts": { "blivet_package_list": [ "python3-blivet", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap", "vdo", "kmod-kvdo", "xfsprogs" ] }, "ansible_included_var_files": [ "/tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/vars/RedHat_8.yml" ], "ansible_loop_var": "item", "changed": false, "item": "RedHat_8.yml" } skipping: [/cache/rhel-8-y.qcow2.snap] => (item=RedHat_8.7.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat_8.7.yml", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : define an empty list of pools to be used in testing] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5 Thursday 21 July 2022 19:04:44 +0000 (0:00:00.062) 0:00:02.052 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : define an empty list of volumes to be used in testing] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9 Thursday 21 July 2022 19:04:44 +0000 (0:00:00.030) 0:00:02.082 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : include the appropriate provider tasks] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13 Thursday 21 July 2022 19:04:44 +0000 (0:00:00.030) 0:00:02.112 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount included: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for /cache/rhel-8-y.qcow2.snap TASK [fedora.linux_system_roles.storage : get a list of rpm packages installed on host machine] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2 Thursday 21 July 2022 19:04:44 +0000 (0:00:00.053) 0:00:02.166 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : make sure blivet is available] ******* task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:7 Thursday 21 July 2022 19:04:44 +0000 (0:00:00.017) 0:00:02.184 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [fedora.linux_system_roles.storage : show storage_pools] ****************** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:13 Thursday 21 July 2022 19:04:46 +0000 (0:00:01.764) 0:00:03.949 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_pools": "VARIABLE IS NOT DEFINED!: 'storage_pools' is undefined" } TASK [fedora.linux_system_roles.storage : show storage_volumes] **************** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:18 Thursday 21 July 2022 19:04:46 +0000 (0:00:00.033) 0:00:03.983 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [fedora.linux_system_roles.storage : get required packages] *************** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:23 Thursday 21 July 2022 19:04:46 +0000 (0:00:00.033) 0:00:04.016 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [], "pools": [], "volumes": [] } TASK [fedora.linux_system_roles.storage : enable copr repositories if needed] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:35 Thursday 21 July 2022 19:04:47 +0000 (0:00:00.711) 0:00:04.727 ********* included: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml for /cache/rhel-8-y.qcow2.snap TASK [fedora.linux_system_roles.storage : check if the COPR support packages should be installed] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:2 Thursday 21 July 2022 19:04:47 +0000 (0:00:00.042) 0:00:04.770 ********* TASK [fedora.linux_system_roles.storage : make sure COPR support packages are present] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:13 Thursday 21 July 2022 19:04:47 +0000 (0:00:00.031) 0:00:04.801 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : enable COPRs] ************************ task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:18 Thursday 21 July 2022 19:04:47 +0000 (0:00:00.036) 0:00:04.837 ********* TASK [fedora.linux_system_roles.storage : make sure required packages are installed] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:41 Thursday 21 July 2022 19:04:47 +0000 (0:00:00.031) 0:00:04.868 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [fedora.linux_system_roles.storage : get service facts] ******************* task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:47 Thursday 21 July 2022 19:04:48 +0000 (0:00:01.319) 0:00:06.188 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "services": { "NetworkManager-dispatcher.service": { "name": "NetworkManager-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "NetworkManager-wait-online.service": { "name": "NetworkManager-wait-online.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "NetworkManager.service": { "name": "NetworkManager.service", "source": "systemd", "state": "running", "status": "enabled" }, "arp-ethers.service": { "name": "arp-ethers.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "auditd.service": { "name": "auditd.service", "source": "systemd", "state": "running", "status": "enabled" }, "auth-rpcgss-module.service": { "name": "auth-rpcgss-module.service", "source": "systemd", "state": "stopped", "status": "static" }, "autovt@.service": { "name": "autovt@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "blivet.service": { "name": "blivet.service", "source": "systemd", "state": "inactive", "status": "static" }, "blk-availability.service": { "name": "blk-availability.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "chrony-dnssrv@.service": { "name": "chrony-dnssrv@.service", "source": "systemd", "state": "unknown", "status": "static" }, "chrony-wait.service": { "name": "chrony-wait.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd.service": { "name": "chronyd.service", "source": "systemd", "state": "running", "status": "enabled" }, "cloud-config.service": { "name": "cloud-config.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-final.service": { "name": "cloud-final.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init-hotplugd.service": { "name": "cloud-init-hotplugd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cloud-init-local.service": { "name": "cloud-init-local.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init.service": { "name": "cloud-init.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cockpit-motd.service": { "name": "cockpit-motd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cockpit-wsinstance-http.service": { "name": "cockpit-wsinstance-http.service", "source": "systemd", "state": "inactive", "status": "static" }, "cockpit-wsinstance-https-factory@.service": { "name": "cockpit-wsinstance-https-factory@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cockpit-wsinstance-https@.service": { "name": "cockpit-wsinstance-https@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cockpit.service": { "name": "cockpit.service", "source": "systemd", "state": "inactive", "status": "static" }, "console-getty.service": { "name": "console-getty.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "container-getty@.service": { "name": "container-getty@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cpupower.service": { "name": "cpupower.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "crond.service": { "name": "crond.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-org.freedesktop.hostname1.service": { "name": "dbus-org.freedesktop.hostname1.service", "source": "systemd", "state": "active", "status": "static" }, "dbus-org.freedesktop.locale1.service": { "name": "dbus-org.freedesktop.locale1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.login1.service": { "name": "dbus-org.freedesktop.login1.service", "source": "systemd", "state": "active", "status": "static" }, "dbus-org.freedesktop.nm-dispatcher.service": { "name": "dbus-org.freedesktop.nm-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "dbus-org.freedesktop.portable1.service": { "name": "dbus-org.freedesktop.portable1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.timedate1.service": { "name": "dbus-org.freedesktop.timedate1.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "dbus.service": { "name": "dbus.service", "source": "systemd", "state": "running", "status": "static" }, "dbxtool.service": { "name": "dbxtool.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "debug-shell.service": { "name": "debug-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "display-manager.service": { "name": "display-manager.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "dm-event.service": { "name": "dm-event.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-makecache.service": { "name": "dnf-makecache.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-cmdline.service": { "name": "dracut-cmdline.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-initqueue.service": { "name": "dracut-initqueue.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-mount.service": { "name": "dracut-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-mount.service": { "name": "dracut-pre-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-pivot.service": { "name": "dracut-pre-pivot.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-trigger.service": { "name": "dracut-pre-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-udev.service": { "name": "dracut-pre-udev.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown.service": { "name": "dracut-shutdown.service", "source": "systemd", "state": "stopped", "status": "static" }, "emergency.service": { "name": "emergency.service", "source": "systemd", "state": "stopped", "status": "static" }, "fcoe.service": { "name": "fcoe.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "fstrim.service": { "name": "fstrim.service", "source": "systemd", "state": "inactive", "status": "static" }, "getty@.service": { "name": "getty@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "getty@tty1.service": { "name": "getty@tty1.service", "source": "systemd", "state": "running", "status": "active" }, "grub-boot-indeterminate.service": { "name": "grub-boot-indeterminate.service", "source": "systemd", "state": "inactive", "status": "static" }, "gssproxy.service": { "name": "gssproxy.service", "source": "systemd", "state": "running", "status": "disabled" }, "halt-local.service": { "name": "halt-local.service", "source": "systemd", "state": "inactive", "status": "static" }, "hv_kvp_daemon.service": { "name": "hv_kvp_daemon.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "import-state.service": { "name": "import-state.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "initrd-cleanup.service": { "name": "initrd-cleanup.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-parse-etc.service": { "name": "initrd-parse-etc.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-switch-root.service": { "name": "initrd-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-udevadm-cleanup-db.service": { "name": "initrd-udevadm-cleanup-db.service", "source": "systemd", "state": "stopped", "status": "static" }, "insights-client-boot.service": { "name": "insights-client-boot.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "insights-client-results.service": { "name": "insights-client-results.service", "source": "systemd", "state": "inactive", "status": "static" }, "insights-client.service": { "name": "insights-client.service", "source": "systemd", "state": "inactive", "status": "static" }, "irqbalance.service": { "name": "irqbalance.service", "source": "systemd", "state": "running", "status": "enabled" }, "iscsi-shutdown.service": { "name": "iscsi-shutdown.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "iscsi.service": { "name": "iscsi.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "iscsid.service": { "name": "iscsid.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "kdump.service": { "name": "kdump.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "kmod-static-nodes.service": { "name": "kmod-static-nodes.service", "source": "systemd", "state": "stopped", "status": "static" }, "kvm_stat.service": { "name": "kvm_stat.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "ldconfig.service": { "name": "ldconfig.service", "source": "systemd", "state": "stopped", "status": "static" }, "loadmodules.service": { "name": "loadmodules.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "lvm2-activation-early.service": { "name": "lvm2-activation-early.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "lvm2-activation.service": { "name": "lvm2-activation.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "lvm2-lvmpolld.service": { "name": "lvm2-lvmpolld.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-monitor.service": { "name": "lvm2-monitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "lvm2-pvscan@.service": { "name": "lvm2-pvscan@.service", "source": "systemd", "state": "unknown", "status": "static" }, "man-db-cache-update.service": { "name": "man-db-cache-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "man-db-restart-cache-update.service": { "name": "man-db-restart-cache-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "mdadm-grow-continue@.service": { "name": "mdadm-grow-continue@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdadm-last-resort@.service": { "name": "mdadm-last-resort@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdcheck_continue.service": { "name": "mdcheck_continue.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdcheck_start.service": { "name": "mdcheck_start.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmon@.service": { "name": "mdmon@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdmonitor-oneshot.service": { "name": "mdmonitor-oneshot.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmonitor.service": { "name": "mdmonitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "messagebus.service": { "name": "messagebus.service", "source": "systemd", "state": "active", "status": "static" }, "microcode.service": { "name": "microcode.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "multipathd.service": { "name": "multipathd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "ndctl-monitor.service": { "name": "ndctl-monitor.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "network.service": { "name": "network.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "nfs-blkmap.service": { "name": "nfs-blkmap.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-convert.service": { "name": "nfs-convert.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-idmapd.service": { "name": "nfs-idmapd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-mountd.service": { "name": "nfs-mountd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-server.service": { "name": "nfs-server.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-utils.service": { "name": "nfs-utils.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfsdcld.service": { "name": "nfsdcld.service", "source": "systemd", "state": "stopped", "status": "static" }, "nis-domainname.service": { "name": "nis-domainname.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "ntpd.service": { "name": "ntpd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ntpdate.service": { "name": "ntpdate.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "oddjobd.service": { "name": "oddjobd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "packagekit-offline-update.service": { "name": "packagekit-offline-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "packagekit.service": { "name": "packagekit.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-quit-wait.service": { "name": "plymouth-quit-wait.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "plymouth-start.service": { "name": "plymouth-start.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "polkit.service": { "name": "polkit.service", "source": "systemd", "state": "running", "status": "static" }, "qemu-guest-agent.service": { "name": "qemu-guest-agent.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "quotaon.service": { "name": "quotaon.service", "source": "systemd", "state": "inactive", "status": "static" }, "rbdmap.service": { "name": "rbdmap.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "rc-local.service": { "name": "rc-local.service", "source": "systemd", "state": "stopped", "status": "static" }, "rdisc.service": { "name": "rdisc.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rescue.service": { "name": "rescue.service", "source": "systemd", "state": "stopped", "status": "static" }, "rhcd.service": { "name": "rhcd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhsm-facts.service": { "name": "rhsm-facts.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhsm.service": { "name": "rhsm.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhsmcertd.service": { "name": "rhsmcertd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpc-gssd.service": { "name": "rpc-gssd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd-notify.service": { "name": "rpc-statd-notify.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd.service": { "name": "rpc-statd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpcbind.service": { "name": "rpcbind.service", "source": "systemd", "state": "running", "status": "enabled" }, "rsyslog.service": { "name": "rsyslog.service", "source": "systemd", "state": "running", "status": "enabled" }, "selinux-autorelabel-mark.service": { "name": "selinux-autorelabel-mark.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "selinux-autorelabel.service": { "name": "selinux-autorelabel.service", "source": "systemd", "state": "inactive", "status": "static" }, "serial-getty@.service": { "name": "serial-getty@.service", "source": "systemd", "state": "unknown", "status": "indirect" }, "serial-getty@ttyS0.service": { "name": "serial-getty@ttyS0.service", "source": "systemd", "state": "running", "status": "active" }, "snapd.seeded.service": { "name": "snapd.seeded.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sntp.service": { "name": "sntp.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sshd-keygen.service": { "name": "sshd-keygen.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sshd-keygen@.service": { "name": "sshd-keygen@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@ecdsa.service": { "name": "sshd-keygen@ecdsa.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd-keygen@ed25519.service": { "name": "sshd-keygen@ed25519.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd-keygen@rsa.service": { "name": "sshd-keygen@rsa.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd.service": { "name": "sshd.service", "source": "systemd", "state": "running", "status": "enabled" }, "sshd@.service": { "name": "sshd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "sssd-autofs.service": { "name": "sssd-autofs.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-kcm.service": { "name": "sssd-kcm.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "sssd-nss.service": { "name": "sssd-nss.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pac.service": { "name": "sssd-pac.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pam.service": { "name": "sssd-pam.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-ssh.service": { "name": "sssd-ssh.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-sudo.service": { "name": "sssd-sudo.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd.service": { "name": "sssd.service", "source": "systemd", "state": "running", "status": "enabled" }, "syslog.service": { "name": "syslog.service", "source": "systemd", "state": "active", "status": "enabled" }, "system-update-cleanup.service": { "name": "system-update-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-ask-password-console.service": { "name": "systemd-ask-password-console.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-wall.service": { "name": "systemd-ask-password-wall.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-backlight@.service": { "name": "systemd-backlight@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-binfmt.service": { "name": "systemd-binfmt.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-coredump@.service": { "name": "systemd-coredump@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-exit.service": { "name": "systemd-exit.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-firstboot.service": { "name": "systemd-firstboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck-root.service": { "name": "systemd-fsck-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck@.service": { "name": "systemd-fsck@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-fsck@dev-disk-by\\x2duuid-7B77\\x2d95E7.service": { "name": "systemd-fsck@dev-disk-by\\x2duuid-7B77\\x2d95E7.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-fsck@dev-vdb1.service": { "name": "systemd-fsck@dev-vdb1.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-fsck@dev-vdc1.service": { "name": "systemd-fsck@dev-vdc1.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-halt.service": { "name": "systemd-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hibernate-resume@.service": { "name": "systemd-hibernate-resume@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-hibernate.service": { "name": "systemd-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hostnamed.service": { "name": "systemd-hostnamed.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-hwdb-update.service": { "name": "systemd-hwdb-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-hybrid-sleep.service": { "name": "systemd-hybrid-sleep.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-initctl.service": { "name": "systemd-initctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-catalog-update.service": { "name": "systemd-journal-catalog-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-flush.service": { "name": "systemd-journal-flush.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journald.service": { "name": "systemd-journald.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-kexec.service": { "name": "systemd-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-localed.service": { "name": "systemd-localed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-logind.service": { "name": "systemd-logind.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-machine-id-commit.service": { "name": "systemd-machine-id-commit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-modules-load.service": { "name": "systemd-modules-load.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-networkd-wait-online.service": { "name": "systemd-networkd-wait-online.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "systemd-portabled.service": { "name": "systemd-portabled.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-poweroff.service": { "name": "systemd-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-quotacheck.service": { "name": "systemd-quotacheck.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-random-seed.service": { "name": "systemd-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-reboot.service": { "name": "systemd-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-remount-fs.service": { "name": "systemd-remount-fs.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-resolved.service": { "name": "systemd-resolved.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-rfkill.service": { "name": "systemd-rfkill.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend-then-hibernate.service": { "name": "systemd-suspend-then-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend.service": { "name": "systemd-suspend.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-sysctl.service": { "name": "systemd-sysctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-sysusers.service": { "name": "systemd-sysusers.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-timedated.service": { "name": "systemd-timedated.service", "source": "systemd", "state": "inactive", "status": "masked" }, "systemd-timesyncd.service": { "name": "systemd-timesyncd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "systemd-tmpfiles-clean.service": { "name": "systemd-tmpfiles-clean.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev.service": { "name": "systemd-tmpfiles-setup-dev.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup.service": { "name": "systemd-tmpfiles-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-settle.service": { "name": "systemd-udev-settle.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-trigger.service": { "name": "systemd-udev-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udevd.service": { "name": "systemd-udevd.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-update-done.service": { "name": "systemd-update-done.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp-runlevel.service": { "name": "systemd-update-utmp-runlevel.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp.service": { "name": "systemd-update-utmp.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-user-sessions.service": { "name": "systemd-user-sessions.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-vconsole-setup.service": { "name": "systemd-vconsole-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-volatile-root.service": { "name": "systemd-volatile-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "tcsd.service": { "name": "tcsd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "teamd@.service": { "name": "teamd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "timedatex.service": { "name": "timedatex.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "tuned.service": { "name": "tuned.service", "source": "systemd", "state": "running", "status": "enabled" }, "unbound-anchor.service": { "name": "unbound-anchor.service", "source": "systemd", "state": "stopped", "status": "static" }, "user-runtime-dir@.service": { "name": "user-runtime-dir@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user-runtime-dir@0.service": { "name": "user-runtime-dir@0.service", "source": "systemd", "state": "stopped", "status": "active" }, "user@.service": { "name": "user@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user@0.service": { "name": "user@0.service", "source": "systemd", "state": "running", "status": "active" }, "vdo-start-by-dev@.service": { "name": "vdo-start-by-dev@.service", "source": "systemd", "state": "unknown", "status": "static" }, "vdo.service": { "name": "vdo.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "ypbind.service": { "name": "ypbind.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "yppasswdd.service": { "name": "yppasswdd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ypserv.service": { "name": "ypserv.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ypxfrd.service": { "name": "ypxfrd.service", "source": "systemd", "state": "stopped", "status": "not-found" } } }, "changed": false } TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] ***** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:53 Thursday 21 July 2022 19:04:50 +0000 (0:00:01.908) 0:00:08.097 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:58 Thursday 21 July 2022 19:04:50 +0000 (0:00:00.054) 0:00:08.152 ********* TASK [fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64 Thursday 21 July 2022 19:04:50 +0000 (0:00:00.020) 0:00:08.173 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [], "pools": [], "volumes": [] } TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:78 Thursday 21 July 2022 19:04:51 +0000 (0:00:00.553) 0:00:08.726 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90 Thursday 21 July 2022 19:04:51 +0000 (0:00:00.034) 0:00:08.760 ********* TASK [fedora.linux_system_roles.storage : show blivet_output] ****************** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:96 Thursday 21 July 2022 19:04:51 +0000 (0:00:00.021) 0:00:08.781 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "blivet_output": { "actions": [], "changed": false, "crypts": [], "failed": false, "leaves": [], "mounts": [], "packages": [], "pools": [], "volumes": [] } } TASK [fedora.linux_system_roles.storage : set the list of pools for test verification] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:101 Thursday 21 July 2022 19:04:51 +0000 (0:00:00.035) 0:00:08.817 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : set the list of volumes for test verification] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:105 Thursday 21 July 2022 19:04:51 +0000 (0:00:00.038) 0:00:08.856 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : remove obsolete mounts] ************** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:121 Thursday 21 July 2022 19:04:51 +0000 (0:00:00.038) 0:00:08.895 ********* TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132 Thursday 21 July 2022 19:04:51 +0000 (0:00:00.046) 0:00:08.941 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : set up new/current mounts] *********** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:137 Thursday 21 July 2022 19:04:51 +0000 (0:00:00.024) 0:00:08.966 ********* TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148 Thursday 21 July 2022 19:04:51 +0000 (0:00:00.035) 0:00:09.001 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : retrieve facts for the /etc/crypttab file] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:156 Thursday 21 July 2022 19:04:51 +0000 (0:00:00.024) 0:00:09.025 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "stat": { "exists": false } } TASK [fedora.linux_system_roles.storage : manage /etc/crypttab to account for changes we just made] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:161 Thursday 21 July 2022 19:04:51 +0000 (0:00:00.501) 0:00:09.526 ********* TASK [fedora.linux_system_roles.storage : Update facts] ************************ task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:183 Thursday 21 July 2022 19:04:51 +0000 (0:00:00.023) 0:00:09.550 ********* ok: [/cache/rhel-8-y.qcow2.snap] META: role_complete for /cache/rhel-8-y.qcow2.snap TASK [include_tasks] *********************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/tests_lvm_percent_size.yml:20 Thursday 21 July 2022 19:04:52 +0000 (0:00:00.977) 0:00:10.528 ********* included: /tmp/tmpmnqeja9w/tests/storage/get_unused_disk.yml for /cache/rhel-8-y.qcow2.snap TASK [Find unused disks in the system] ***************************************** task path: /tmp/tmpmnqeja9w/tests/storage/get_unused_disk.yml:2 Thursday 21 July 2022 19:04:53 +0000 (0:00:00.035) 0:00:10.564 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "disks": [ "sda" ] } TASK [Set unused_disks if necessary] ******************************************* task path: /tmp/tmpmnqeja9w/tests/storage/get_unused_disk.yml:9 Thursday 21 July 2022 19:04:53 +0000 (0:00:00.524) 0:00:11.088 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "unused_disks": [ "sda" ] }, "changed": false } TASK [Exit playbook when there's not enough unused disks in the system] ******** task path: /tmp/tmpmnqeja9w/tests/storage/get_unused_disk.yml:14 Thursday 21 July 2022 19:04:53 +0000 (0:00:00.035) 0:00:11.124 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Print unused disks] ****************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/get_unused_disk.yml:19 Thursday 21 July 2022 19:04:53 +0000 (0:00:00.040) 0:00:11.164 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "unused_disks": [ "sda" ] } TASK [Try to create LVM with an invalid size specification.] ******************* task path: /tmp/tmpmnqeja9w/tests/storage/tests_lvm_percent_size.yml:27 Thursday 21 July 2022 19:04:53 +0000 (0:00:00.066) 0:00:11.231 ********* TASK [fedora.linux_system_roles.storage : set platform/version specific variables] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2 Thursday 21 July 2022 19:04:53 +0000 (0:00:00.041) 0:00:11.272 ********* included: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for /cache/rhel-8-y.qcow2.snap TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2 Thursday 21 July 2022 19:04:53 +0000 (0:00:00.037) 0:00:11.309 ********* ok: [/cache/rhel-8-y.qcow2.snap] TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7 Thursday 21 July 2022 19:04:54 +0000 (0:00:00.556) 0:00:11.866 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [/cache/rhel-8-y.qcow2.snap] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [/cache/rhel-8-y.qcow2.snap] => (item=RedHat_8.yml) => { "ansible_facts": { "blivet_package_list": [ "python3-blivet", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap", "vdo", "kmod-kvdo", "xfsprogs" ] }, "ansible_included_var_files": [ "/tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/vars/RedHat_8.yml" ], "ansible_loop_var": "item", "changed": false, "item": "RedHat_8.yml" } skipping: [/cache/rhel-8-y.qcow2.snap] => (item=RedHat_8.7.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat_8.7.yml", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : define an empty list of pools to be used in testing] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5 Thursday 21 July 2022 19:04:54 +0000 (0:00:00.061) 0:00:11.927 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : define an empty list of volumes to be used in testing] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9 Thursday 21 July 2022 19:04:54 +0000 (0:00:00.035) 0:00:11.962 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : include the appropriate provider tasks] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13 Thursday 21 July 2022 19:04:54 +0000 (0:00:00.036) 0:00:11.999 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount included: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for /cache/rhel-8-y.qcow2.snap TASK [fedora.linux_system_roles.storage : get a list of rpm packages installed on host machine] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2 Thursday 21 July 2022 19:04:54 +0000 (0:00:00.058) 0:00:12.058 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : make sure blivet is available] ******* task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:7 Thursday 21 July 2022 19:04:54 +0000 (0:00:00.022) 0:00:12.081 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [fedora.linux_system_roles.storage : show storage_pools] ****************** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:13 Thursday 21 July 2022 19:04:55 +0000 (0:00:01.333) 0:00:13.415 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_pools": [ { "disks": [ "sda" ], "name": "foo", "volumes": [ { "mount_point": "/opt/test1", "name": "test1", "size": "2x%" } ] } ] } TASK [fedora.linux_system_roles.storage : show storage_volumes] **************** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:18 Thursday 21 July 2022 19:04:55 +0000 (0:00:00.038) 0:00:13.454 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [fedora.linux_system_roles.storage : get required packages] *************** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:23 Thursday 21 July 2022 19:04:55 +0000 (0:00:00.038) 0:00:13.492 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [ "lvm2" ], "pools": [], "volumes": [] } TASK [fedora.linux_system_roles.storage : enable copr repositories if needed] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:35 Thursday 21 July 2022 19:05:07 +0000 (0:00:11.632) 0:00:25.125 ********* included: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml for /cache/rhel-8-y.qcow2.snap TASK [fedora.linux_system_roles.storage : check if the COPR support packages should be installed] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:2 Thursday 21 July 2022 19:05:07 +0000 (0:00:00.046) 0:00:25.171 ********* TASK [fedora.linux_system_roles.storage : make sure COPR support packages are present] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:13 Thursday 21 July 2022 19:05:07 +0000 (0:00:00.035) 0:00:25.207 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : enable COPRs] ************************ task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:18 Thursday 21 July 2022 19:05:07 +0000 (0:00:00.039) 0:00:25.246 ********* TASK [fedora.linux_system_roles.storage : make sure required packages are installed] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:41 Thursday 21 July 2022 19:05:07 +0000 (0:00:00.034) 0:00:25.281 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [fedora.linux_system_roles.storage : get service facts] ******************* task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:47 Thursday 21 July 2022 19:05:09 +0000 (0:00:01.376) 0:00:26.657 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "services": { "NetworkManager-dispatcher.service": { "name": "NetworkManager-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "NetworkManager-wait-online.service": { "name": "NetworkManager-wait-online.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "NetworkManager.service": { "name": "NetworkManager.service", "source": "systemd", "state": "running", "status": "enabled" }, "arp-ethers.service": { "name": "arp-ethers.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "auditd.service": { "name": "auditd.service", "source": "systemd", "state": "running", "status": "enabled" }, "auth-rpcgss-module.service": { "name": "auth-rpcgss-module.service", "source": "systemd", "state": "stopped", "status": "static" }, "autovt@.service": { "name": "autovt@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "blivet.service": { "name": "blivet.service", "source": "systemd", "state": "inactive", "status": "static" }, "blk-availability.service": { "name": "blk-availability.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "chrony-dnssrv@.service": { "name": "chrony-dnssrv@.service", "source": "systemd", "state": "unknown", "status": "static" }, "chrony-wait.service": { "name": "chrony-wait.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd.service": { "name": "chronyd.service", "source": "systemd", "state": "running", "status": "enabled" }, "cloud-config.service": { "name": "cloud-config.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-final.service": { "name": "cloud-final.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init-hotplugd.service": { "name": "cloud-init-hotplugd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cloud-init-local.service": { "name": "cloud-init-local.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init.service": { "name": "cloud-init.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cockpit-motd.service": { "name": "cockpit-motd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cockpit-wsinstance-http.service": { "name": "cockpit-wsinstance-http.service", "source": "systemd", "state": "inactive", "status": "static" }, "cockpit-wsinstance-https-factory@.service": { "name": "cockpit-wsinstance-https-factory@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cockpit-wsinstance-https@.service": { "name": "cockpit-wsinstance-https@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cockpit.service": { "name": "cockpit.service", "source": "systemd", "state": "inactive", "status": "static" }, "console-getty.service": { "name": "console-getty.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "container-getty@.service": { "name": "container-getty@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cpupower.service": { "name": "cpupower.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "crond.service": { "name": "crond.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-org.freedesktop.hostname1.service": { "name": "dbus-org.freedesktop.hostname1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.locale1.service": { "name": "dbus-org.freedesktop.locale1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.login1.service": { "name": "dbus-org.freedesktop.login1.service", "source": "systemd", "state": "active", "status": "static" }, "dbus-org.freedesktop.nm-dispatcher.service": { "name": "dbus-org.freedesktop.nm-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "dbus-org.freedesktop.portable1.service": { "name": "dbus-org.freedesktop.portable1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.timedate1.service": { "name": "dbus-org.freedesktop.timedate1.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "dbus.service": { "name": "dbus.service", "source": "systemd", "state": "running", "status": "static" }, "dbxtool.service": { "name": "dbxtool.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "debug-shell.service": { "name": "debug-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "display-manager.service": { "name": "display-manager.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "dm-event.service": { "name": "dm-event.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-makecache.service": { "name": "dnf-makecache.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-cmdline.service": { "name": "dracut-cmdline.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-initqueue.service": { "name": "dracut-initqueue.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-mount.service": { "name": "dracut-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-mount.service": { "name": "dracut-pre-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-pivot.service": { "name": "dracut-pre-pivot.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-trigger.service": { "name": "dracut-pre-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-udev.service": { "name": "dracut-pre-udev.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown.service": { "name": "dracut-shutdown.service", "source": "systemd", "state": "stopped", "status": "static" }, "emergency.service": { "name": "emergency.service", "source": "systemd", "state": "stopped", "status": "static" }, "fcoe.service": { "name": "fcoe.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "fstrim.service": { "name": "fstrim.service", "source": "systemd", "state": "inactive", "status": "static" }, "getty@.service": { "name": "getty@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "getty@tty1.service": { "name": "getty@tty1.service", "source": "systemd", "state": "running", "status": "active" }, "grub-boot-indeterminate.service": { "name": "grub-boot-indeterminate.service", "source": "systemd", "state": "inactive", "status": "static" }, "gssproxy.service": { "name": "gssproxy.service", "source": "systemd", "state": "running", "status": "disabled" }, "halt-local.service": { "name": "halt-local.service", "source": "systemd", "state": "inactive", "status": "static" }, "hv_kvp_daemon.service": { "name": "hv_kvp_daemon.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "import-state.service": { "name": "import-state.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "initrd-cleanup.service": { "name": "initrd-cleanup.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-parse-etc.service": { "name": "initrd-parse-etc.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-switch-root.service": { "name": "initrd-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-udevadm-cleanup-db.service": { "name": "initrd-udevadm-cleanup-db.service", "source": "systemd", "state": "stopped", "status": "static" }, "insights-client-boot.service": { "name": "insights-client-boot.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "insights-client-results.service": { "name": "insights-client-results.service", "source": "systemd", "state": "inactive", "status": "static" }, "insights-client.service": { "name": "insights-client.service", "source": "systemd", "state": "inactive", "status": "static" }, "irqbalance.service": { "name": "irqbalance.service", "source": "systemd", "state": "running", "status": "enabled" }, "iscsi-shutdown.service": { "name": "iscsi-shutdown.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "iscsi.service": { "name": "iscsi.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "iscsid.service": { "name": "iscsid.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "kdump.service": { "name": "kdump.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "kmod-static-nodes.service": { "name": "kmod-static-nodes.service", "source": "systemd", "state": "stopped", "status": "static" }, "kvm_stat.service": { "name": "kvm_stat.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "ldconfig.service": { "name": "ldconfig.service", "source": "systemd", "state": "stopped", "status": "static" }, "loadmodules.service": { "name": "loadmodules.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "lvm2-activation-early.service": { "name": "lvm2-activation-early.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "lvm2-activation.service": { "name": "lvm2-activation.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "lvm2-lvmpolld.service": { "name": "lvm2-lvmpolld.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-monitor.service": { "name": "lvm2-monitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "lvm2-pvscan@.service": { "name": "lvm2-pvscan@.service", "source": "systemd", "state": "unknown", "status": "static" }, "man-db-cache-update.service": { "name": "man-db-cache-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "man-db-restart-cache-update.service": { "name": "man-db-restart-cache-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "mdadm-grow-continue@.service": { "name": "mdadm-grow-continue@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdadm-last-resort@.service": { "name": "mdadm-last-resort@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdcheck_continue.service": { "name": "mdcheck_continue.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdcheck_start.service": { "name": "mdcheck_start.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmon@.service": { "name": "mdmon@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdmonitor-oneshot.service": { "name": "mdmonitor-oneshot.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmonitor.service": { "name": "mdmonitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "messagebus.service": { "name": "messagebus.service", "source": "systemd", "state": "active", "status": "static" }, "microcode.service": { "name": "microcode.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "multipathd.service": { "name": "multipathd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "ndctl-monitor.service": { "name": "ndctl-monitor.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "network.service": { "name": "network.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "nfs-blkmap.service": { "name": "nfs-blkmap.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-convert.service": { "name": "nfs-convert.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-idmapd.service": { "name": "nfs-idmapd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-mountd.service": { "name": "nfs-mountd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-server.service": { "name": "nfs-server.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-utils.service": { "name": "nfs-utils.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfsdcld.service": { "name": "nfsdcld.service", "source": "systemd", "state": "stopped", "status": "static" }, "nis-domainname.service": { "name": "nis-domainname.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "ntpd.service": { "name": "ntpd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ntpdate.service": { "name": "ntpdate.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "oddjobd.service": { "name": "oddjobd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "packagekit-offline-update.service": { "name": "packagekit-offline-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "packagekit.service": { "name": "packagekit.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-quit-wait.service": { "name": "plymouth-quit-wait.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "plymouth-start.service": { "name": "plymouth-start.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "polkit.service": { "name": "polkit.service", "source": "systemd", "state": "running", "status": "static" }, "qemu-guest-agent.service": { "name": "qemu-guest-agent.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "quotaon.service": { "name": "quotaon.service", "source": "systemd", "state": "inactive", "status": "static" }, "rbdmap.service": { "name": "rbdmap.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "rc-local.service": { "name": "rc-local.service", "source": "systemd", "state": "stopped", "status": "static" }, "rdisc.service": { "name": "rdisc.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rescue.service": { "name": "rescue.service", "source": "systemd", "state": "stopped", "status": "static" }, "rhcd.service": { "name": "rhcd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhsm-facts.service": { "name": "rhsm-facts.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhsm.service": { "name": "rhsm.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhsmcertd.service": { "name": "rhsmcertd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpc-gssd.service": { "name": "rpc-gssd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd-notify.service": { "name": "rpc-statd-notify.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd.service": { "name": "rpc-statd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpcbind.service": { "name": "rpcbind.service", "source": "systemd", "state": "running", "status": "enabled" }, "rsyslog.service": { "name": "rsyslog.service", "source": "systemd", "state": "running", "status": "enabled" }, "selinux-autorelabel-mark.service": { "name": "selinux-autorelabel-mark.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "selinux-autorelabel.service": { "name": "selinux-autorelabel.service", "source": "systemd", "state": "inactive", "status": "static" }, "serial-getty@.service": { "name": "serial-getty@.service", "source": "systemd", "state": "unknown", "status": "indirect" }, "serial-getty@ttyS0.service": { "name": "serial-getty@ttyS0.service", "source": "systemd", "state": "running", "status": "active" }, "snapd.seeded.service": { "name": "snapd.seeded.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sntp.service": { "name": "sntp.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sshd-keygen.service": { "name": "sshd-keygen.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sshd-keygen@.service": { "name": "sshd-keygen@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@ecdsa.service": { "name": "sshd-keygen@ecdsa.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd-keygen@ed25519.service": { "name": "sshd-keygen@ed25519.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd-keygen@rsa.service": { "name": "sshd-keygen@rsa.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd.service": { "name": "sshd.service", "source": "systemd", "state": "running", "status": "enabled" }, "sshd@.service": { "name": "sshd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "sssd-autofs.service": { "name": "sssd-autofs.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-kcm.service": { "name": "sssd-kcm.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "sssd-nss.service": { "name": "sssd-nss.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pac.service": { "name": "sssd-pac.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pam.service": { "name": "sssd-pam.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-ssh.service": { "name": "sssd-ssh.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-sudo.service": { "name": "sssd-sudo.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd.service": { "name": "sssd.service", "source": "systemd", "state": "running", "status": "enabled" }, "syslog.service": { "name": "syslog.service", "source": "systemd", "state": "active", "status": "enabled" }, "system-update-cleanup.service": { "name": "system-update-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-ask-password-console.service": { "name": "systemd-ask-password-console.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-wall.service": { "name": "systemd-ask-password-wall.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-backlight@.service": { "name": "systemd-backlight@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-binfmt.service": { "name": "systemd-binfmt.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-coredump@.service": { "name": "systemd-coredump@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-exit.service": { "name": "systemd-exit.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-firstboot.service": { "name": "systemd-firstboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck-root.service": { "name": "systemd-fsck-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck@.service": { "name": "systemd-fsck@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-fsck@dev-disk-by\\x2duuid-7B77\\x2d95E7.service": { "name": "systemd-fsck@dev-disk-by\\x2duuid-7B77\\x2d95E7.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-fsck@dev-vdb1.service": { "name": "systemd-fsck@dev-vdb1.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-fsck@dev-vdc1.service": { "name": "systemd-fsck@dev-vdc1.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-halt.service": { "name": "systemd-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hibernate-resume@.service": { "name": "systemd-hibernate-resume@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-hibernate.service": { "name": "systemd-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hostnamed.service": { "name": "systemd-hostnamed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hwdb-update.service": { "name": "systemd-hwdb-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-hybrid-sleep.service": { "name": "systemd-hybrid-sleep.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-initctl.service": { "name": "systemd-initctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-catalog-update.service": { "name": "systemd-journal-catalog-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-flush.service": { "name": "systemd-journal-flush.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journald.service": { "name": "systemd-journald.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-kexec.service": { "name": "systemd-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-localed.service": { "name": "systemd-localed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-logind.service": { "name": "systemd-logind.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-machine-id-commit.service": { "name": "systemd-machine-id-commit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-modules-load.service": { "name": "systemd-modules-load.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-networkd-wait-online.service": { "name": "systemd-networkd-wait-online.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "systemd-portabled.service": { "name": "systemd-portabled.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-poweroff.service": { "name": "systemd-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-quotacheck.service": { "name": "systemd-quotacheck.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-random-seed.service": { "name": "systemd-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-reboot.service": { "name": "systemd-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-remount-fs.service": { "name": "systemd-remount-fs.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-resolved.service": { "name": "systemd-resolved.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-rfkill.service": { "name": "systemd-rfkill.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend-then-hibernate.service": { "name": "systemd-suspend-then-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend.service": { "name": "systemd-suspend.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-sysctl.service": { "name": "systemd-sysctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-sysusers.service": { "name": "systemd-sysusers.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-timedated.service": { "name": "systemd-timedated.service", "source": "systemd", "state": "inactive", "status": "masked" }, "systemd-timesyncd.service": { "name": "systemd-timesyncd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "systemd-tmpfiles-clean.service": { "name": "systemd-tmpfiles-clean.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev.service": { "name": "systemd-tmpfiles-setup-dev.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup.service": { "name": "systemd-tmpfiles-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-settle.service": { "name": "systemd-udev-settle.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-trigger.service": { "name": "systemd-udev-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udevd.service": { "name": "systemd-udevd.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-update-done.service": { "name": "systemd-update-done.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp-runlevel.service": { "name": "systemd-update-utmp-runlevel.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp.service": { "name": "systemd-update-utmp.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-user-sessions.service": { "name": "systemd-user-sessions.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-vconsole-setup.service": { "name": "systemd-vconsole-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-volatile-root.service": { "name": "systemd-volatile-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "tcsd.service": { "name": "tcsd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "teamd@.service": { "name": "teamd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "timedatex.service": { "name": "timedatex.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "tuned.service": { "name": "tuned.service", "source": "systemd", "state": "running", "status": "enabled" }, "unbound-anchor.service": { "name": "unbound-anchor.service", "source": "systemd", "state": "stopped", "status": "static" }, "user-runtime-dir@.service": { "name": "user-runtime-dir@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user-runtime-dir@0.service": { "name": "user-runtime-dir@0.service", "source": "systemd", "state": "stopped", "status": "active" }, "user@.service": { "name": "user@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user@0.service": { "name": "user@0.service", "source": "systemd", "state": "running", "status": "active" }, "vdo-start-by-dev@.service": { "name": "vdo-start-by-dev@.service", "source": "systemd", "state": "unknown", "status": "static" }, "vdo.service": { "name": "vdo.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "ypbind.service": { "name": "ypbind.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "yppasswdd.service": { "name": "yppasswdd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ypserv.service": { "name": "ypserv.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ypxfrd.service": { "name": "ypxfrd.service", "source": "systemd", "state": "stopped", "status": "not-found" } } }, "changed": false } TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] ***** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:53 Thursday 21 July 2022 19:05:10 +0000 (0:00:01.746) 0:00:28.403 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:58 Thursday 21 July 2022 19:05:10 +0000 (0:00:00.121) 0:00:28.524 ********* TASK [fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64 Thursday 21 July 2022 19:05:10 +0000 (0:00:00.022) 0:00:28.547 ********* fatal: [/cache/rhel-8-y.qcow2.snap]: FAILED! => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [], "pools": [], "volumes": [] } MSG: invalid percentage '2x%' size specified for volume in pool 'foo' TASK [fedora.linux_system_roles.storage : failed message] ********************** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:86 Thursday 21 July 2022 19:05:12 +0000 (0:00:01.323) 0:00:29.870 ********* fatal: [/cache/rhel-8-y.qcow2.snap]: FAILED! => { "changed": false } MSG: {'changed': False, 'actions': [], 'leaves': [], 'mounts': [], 'crypts': [], 'pools': [], 'volumes': [], 'packages': [], 'failed': True, 'msg': "invalid percentage '2x%' size specified for volume in pool 'foo'", 'invocation': {'module_args': {'pools': [{'disks': ['sda'], 'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'name': 'foo', 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_metadata_version': None, 'state': 'present', 'type': 'lvm', 'volumes': [{'encryption': None, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': None, 'fs_label': '', 'fs_type': None, 'mount_options': None, 'mount_point': '/opt/test1', 'name': 'test1', 'raid_level': None, 'size': '2x%', 'state': 'present', 'type': None, 'cached': None, 'cache_devices': [], 'cache_mode': None, 'cache_size': None, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None}], 'raid_chunk_size': None}], 'volumes': [], 'use_partitions': None, 'disklabel_type': None, 'pool_defaults': {'state': 'present', 'type': 'lvm', 'disks': [], 'volumes': [], 'encryption': False, 'encryption_password': None, 'encryption_key': None, 'encryption_cipher': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None}, 'volume_defaults': {'state': 'present', 'type': 'lvm', 'size': 0, 'disks': [], 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_overwrite_existing': True, 'mount_point': '', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, 'encryption': False, 'encryption_password': None, 'encryption_key': None, 'encryption_cipher': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'compression': None, 'deduplication': None, 'vdo_pool_size': None, 'thin': None, 'thin_pool_name': None, 'thin_pool_size': None, 'cached': False, 'cache_size': 0, 'cache_mode': None, 'cache_devices': []}, 'safe_mode': False, 'packages_only': False, 'diskvolume_mkfs_option_map': {}}}, '_ansible_no_log': False} TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90 Thursday 21 July 2022 19:05:12 +0000 (0:00:00.040) 0:00:29.911 ********* TASK [Check that we failed in the role] **************************************** task path: /tmp/tmpmnqeja9w/tests/storage/tests_lvm_percent_size.yml:44 Thursday 21 July 2022 19:05:12 +0000 (0:00:00.022) 0:00:29.933 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Check for the expected error message] ************************************ task path: /tmp/tmpmnqeja9w/tests/storage/tests_lvm_percent_size.yml:50 Thursday 21 July 2022 19:05:12 +0000 (0:00:00.036) 0:00:29.970 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Create two LVM logical volumes under volume group 'foo' using percentage sizes] *** task path: /tmp/tmpmnqeja9w/tests/storage/tests_lvm_percent_size.yml:63 Thursday 21 July 2022 19:05:12 +0000 (0:00:00.036) 0:00:30.006 ********* TASK [fedora.linux_system_roles.storage : set platform/version specific variables] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2 Thursday 21 July 2022 19:05:12 +0000 (0:00:00.042) 0:00:30.049 ********* included: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for /cache/rhel-8-y.qcow2.snap TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2 Thursday 21 July 2022 19:05:12 +0000 (0:00:00.034) 0:00:30.083 ********* ok: [/cache/rhel-8-y.qcow2.snap] TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7 Thursday 21 July 2022 19:05:13 +0000 (0:00:00.518) 0:00:30.602 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [/cache/rhel-8-y.qcow2.snap] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [/cache/rhel-8-y.qcow2.snap] => (item=RedHat_8.yml) => { "ansible_facts": { "blivet_package_list": [ "python3-blivet", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap", "vdo", "kmod-kvdo", "xfsprogs" ] }, "ansible_included_var_files": [ "/tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/vars/RedHat_8.yml" ], "ansible_loop_var": "item", "changed": false, "item": "RedHat_8.yml" } skipping: [/cache/rhel-8-y.qcow2.snap] => (item=RedHat_8.7.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat_8.7.yml", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : define an empty list of pools to be used in testing] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5 Thursday 21 July 2022 19:05:13 +0000 (0:00:00.062) 0:00:30.665 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : define an empty list of volumes to be used in testing] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9 Thursday 21 July 2022 19:05:13 +0000 (0:00:00.035) 0:00:30.700 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : include the appropriate provider tasks] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13 Thursday 21 July 2022 19:05:13 +0000 (0:00:00.033) 0:00:30.733 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount included: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for /cache/rhel-8-y.qcow2.snap TASK [fedora.linux_system_roles.storage : get a list of rpm packages installed on host machine] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2 Thursday 21 July 2022 19:05:13 +0000 (0:00:00.054) 0:00:30.788 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : make sure blivet is available] ******* task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:7 Thursday 21 July 2022 19:05:13 +0000 (0:00:00.021) 0:00:30.810 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [fedora.linux_system_roles.storage : show storage_pools] ****************** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:13 Thursday 21 July 2022 19:05:14 +0000 (0:00:01.339) 0:00:32.149 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_pools": [ { "disks": [ "sda" ], "name": "foo", "volumes": [ { "mount_point": "/opt/test1", "name": "test1", "size": "60%" }, { "fs_type": "ext4", "mount_point": "/opt/test2", "name": "test2", "size": "40%" } ] } ] } TASK [fedora.linux_system_roles.storage : show storage_volumes] **************** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:18 Thursday 21 July 2022 19:05:14 +0000 (0:00:00.041) 0:00:32.190 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [fedora.linux_system_roles.storage : get required packages] *************** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:23 Thursday 21 July 2022 19:05:14 +0000 (0:00:00.036) 0:00:32.227 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [ "e2fsprogs", "lvm2" ], "pools": [], "volumes": [] } TASK [fedora.linux_system_roles.storage : enable copr repositories if needed] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:35 Thursday 21 July 2022 19:05:16 +0000 (0:00:01.329) 0:00:33.556 ********* included: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml for /cache/rhel-8-y.qcow2.snap TASK [fedora.linux_system_roles.storage : check if the COPR support packages should be installed] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:2 Thursday 21 July 2022 19:05:16 +0000 (0:00:00.080) 0:00:33.637 ********* TASK [fedora.linux_system_roles.storage : make sure COPR support packages are present] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:13 Thursday 21 July 2022 19:05:16 +0000 (0:00:00.036) 0:00:33.674 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : enable COPRs] ************************ task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:18 Thursday 21 July 2022 19:05:16 +0000 (0:00:00.037) 0:00:33.711 ********* TASK [fedora.linux_system_roles.storage : make sure required packages are installed] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:41 Thursday 21 July 2022 19:05:16 +0000 (0:00:00.034) 0:00:33.745 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [fedora.linux_system_roles.storage : get service facts] ******************* task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:47 Thursday 21 July 2022 19:05:17 +0000 (0:00:01.304) 0:00:35.050 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "services": { "NetworkManager-dispatcher.service": { "name": "NetworkManager-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "NetworkManager-wait-online.service": { "name": "NetworkManager-wait-online.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "NetworkManager.service": { "name": "NetworkManager.service", "source": "systemd", "state": "running", "status": "enabled" }, "arp-ethers.service": { "name": "arp-ethers.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "auditd.service": { "name": "auditd.service", "source": "systemd", "state": "running", "status": "enabled" }, "auth-rpcgss-module.service": { "name": "auth-rpcgss-module.service", "source": "systemd", "state": "stopped", "status": "static" }, "autovt@.service": { "name": "autovt@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "blivet.service": { "name": "blivet.service", "source": "systemd", "state": "inactive", "status": "static" }, "blk-availability.service": { "name": "blk-availability.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "chrony-dnssrv@.service": { "name": "chrony-dnssrv@.service", "source": "systemd", "state": "unknown", "status": "static" }, "chrony-wait.service": { "name": "chrony-wait.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd.service": { "name": "chronyd.service", "source": "systemd", "state": "running", "status": "enabled" }, "cloud-config.service": { "name": "cloud-config.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-final.service": { "name": "cloud-final.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init-hotplugd.service": { "name": "cloud-init-hotplugd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cloud-init-local.service": { "name": "cloud-init-local.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init.service": { "name": "cloud-init.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cockpit-motd.service": { "name": "cockpit-motd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cockpit-wsinstance-http.service": { "name": "cockpit-wsinstance-http.service", "source": "systemd", "state": "inactive", "status": "static" }, "cockpit-wsinstance-https-factory@.service": { "name": "cockpit-wsinstance-https-factory@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cockpit-wsinstance-https@.service": { "name": "cockpit-wsinstance-https@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cockpit.service": { "name": "cockpit.service", "source": "systemd", "state": "inactive", "status": "static" }, "console-getty.service": { "name": "console-getty.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "container-getty@.service": { "name": "container-getty@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cpupower.service": { "name": "cpupower.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "crond.service": { "name": "crond.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-org.freedesktop.hostname1.service": { "name": "dbus-org.freedesktop.hostname1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.locale1.service": { "name": "dbus-org.freedesktop.locale1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.login1.service": { "name": "dbus-org.freedesktop.login1.service", "source": "systemd", "state": "active", "status": "static" }, "dbus-org.freedesktop.nm-dispatcher.service": { "name": "dbus-org.freedesktop.nm-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "dbus-org.freedesktop.portable1.service": { "name": "dbus-org.freedesktop.portable1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.timedate1.service": { "name": "dbus-org.freedesktop.timedate1.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "dbus.service": { "name": "dbus.service", "source": "systemd", "state": "running", "status": "static" }, "dbxtool.service": { "name": "dbxtool.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "debug-shell.service": { "name": "debug-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "display-manager.service": { "name": "display-manager.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "dm-event.service": { "name": "dm-event.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-makecache.service": { "name": "dnf-makecache.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-cmdline.service": { "name": "dracut-cmdline.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-initqueue.service": { "name": "dracut-initqueue.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-mount.service": { "name": "dracut-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-mount.service": { "name": "dracut-pre-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-pivot.service": { "name": "dracut-pre-pivot.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-trigger.service": { "name": "dracut-pre-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-udev.service": { "name": "dracut-pre-udev.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown.service": { "name": "dracut-shutdown.service", "source": "systemd", "state": "stopped", "status": "static" }, "emergency.service": { "name": "emergency.service", "source": "systemd", "state": "stopped", "status": "static" }, "fcoe.service": { "name": "fcoe.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "fstrim.service": { "name": "fstrim.service", "source": "systemd", "state": "inactive", "status": "static" }, "getty@.service": { "name": "getty@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "getty@tty1.service": { "name": "getty@tty1.service", "source": "systemd", "state": "running", "status": "active" }, "grub-boot-indeterminate.service": { "name": "grub-boot-indeterminate.service", "source": "systemd", "state": "inactive", "status": "static" }, "gssproxy.service": { "name": "gssproxy.service", "source": "systemd", "state": "running", "status": "disabled" }, "halt-local.service": { "name": "halt-local.service", "source": "systemd", "state": "inactive", "status": "static" }, "hv_kvp_daemon.service": { "name": "hv_kvp_daemon.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "import-state.service": { "name": "import-state.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "initrd-cleanup.service": { "name": "initrd-cleanup.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-parse-etc.service": { "name": "initrd-parse-etc.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-switch-root.service": { "name": "initrd-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-udevadm-cleanup-db.service": { "name": "initrd-udevadm-cleanup-db.service", "source": "systemd", "state": "stopped", "status": "static" }, "insights-client-boot.service": { "name": "insights-client-boot.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "insights-client-results.service": { "name": "insights-client-results.service", "source": "systemd", "state": "inactive", "status": "static" }, "insights-client.service": { "name": "insights-client.service", "source": "systemd", "state": "inactive", "status": "static" }, "irqbalance.service": { "name": "irqbalance.service", "source": "systemd", "state": "running", "status": "enabled" }, "iscsi-shutdown.service": { "name": "iscsi-shutdown.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "iscsi.service": { "name": "iscsi.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "iscsid.service": { "name": "iscsid.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "kdump.service": { "name": "kdump.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "kmod-static-nodes.service": { "name": "kmod-static-nodes.service", "source": "systemd", "state": "stopped", "status": "static" }, "kvm_stat.service": { "name": "kvm_stat.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "ldconfig.service": { "name": "ldconfig.service", "source": "systemd", "state": "stopped", "status": "static" }, "loadmodules.service": { "name": "loadmodules.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "lvm2-activation-early.service": { "name": "lvm2-activation-early.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "lvm2-activation.service": { "name": "lvm2-activation.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "lvm2-lvmpolld.service": { "name": "lvm2-lvmpolld.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-monitor.service": { "name": "lvm2-monitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "lvm2-pvscan@.service": { "name": "lvm2-pvscan@.service", "source": "systemd", "state": "unknown", "status": "static" }, "man-db-cache-update.service": { "name": "man-db-cache-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "man-db-restart-cache-update.service": { "name": "man-db-restart-cache-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "mdadm-grow-continue@.service": { "name": "mdadm-grow-continue@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdadm-last-resort@.service": { "name": "mdadm-last-resort@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdcheck_continue.service": { "name": "mdcheck_continue.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdcheck_start.service": { "name": "mdcheck_start.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmon@.service": { "name": "mdmon@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdmonitor-oneshot.service": { "name": "mdmonitor-oneshot.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmonitor.service": { "name": "mdmonitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "messagebus.service": { "name": "messagebus.service", "source": "systemd", "state": "active", "status": "static" }, "microcode.service": { "name": "microcode.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "multipathd.service": { "name": "multipathd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "ndctl-monitor.service": { "name": "ndctl-monitor.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "network.service": { "name": "network.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "nfs-blkmap.service": { "name": "nfs-blkmap.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-convert.service": { "name": "nfs-convert.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-idmapd.service": { "name": "nfs-idmapd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-mountd.service": { "name": "nfs-mountd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-server.service": { "name": "nfs-server.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-utils.service": { "name": "nfs-utils.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfsdcld.service": { "name": "nfsdcld.service", "source": "systemd", "state": "stopped", "status": "static" }, "nis-domainname.service": { "name": "nis-domainname.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "ntpd.service": { "name": "ntpd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ntpdate.service": { "name": "ntpdate.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "oddjobd.service": { "name": "oddjobd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "packagekit-offline-update.service": { "name": "packagekit-offline-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "packagekit.service": { "name": "packagekit.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-quit-wait.service": { "name": "plymouth-quit-wait.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "plymouth-start.service": { "name": "plymouth-start.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "polkit.service": { "name": "polkit.service", "source": "systemd", "state": "running", "status": "static" }, "qemu-guest-agent.service": { "name": "qemu-guest-agent.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "quotaon.service": { "name": "quotaon.service", "source": "systemd", "state": "inactive", "status": "static" }, "rbdmap.service": { "name": "rbdmap.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "rc-local.service": { "name": "rc-local.service", "source": "systemd", "state": "stopped", "status": "static" }, "rdisc.service": { "name": "rdisc.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rescue.service": { "name": "rescue.service", "source": "systemd", "state": "stopped", "status": "static" }, "rhcd.service": { "name": "rhcd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhsm-facts.service": { "name": "rhsm-facts.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhsm.service": { "name": "rhsm.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhsmcertd.service": { "name": "rhsmcertd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpc-gssd.service": { "name": "rpc-gssd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd-notify.service": { "name": "rpc-statd-notify.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd.service": { "name": "rpc-statd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpcbind.service": { "name": "rpcbind.service", "source": "systemd", "state": "running", "status": "enabled" }, "rsyslog.service": { "name": "rsyslog.service", "source": "systemd", "state": "running", "status": "enabled" }, "selinux-autorelabel-mark.service": { "name": "selinux-autorelabel-mark.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "selinux-autorelabel.service": { "name": "selinux-autorelabel.service", "source": "systemd", "state": "inactive", "status": "static" }, "serial-getty@.service": { "name": "serial-getty@.service", "source": "systemd", "state": "unknown", "status": "indirect" }, "serial-getty@ttyS0.service": { "name": "serial-getty@ttyS0.service", "source": "systemd", "state": "running", "status": "active" }, "snapd.seeded.service": { "name": "snapd.seeded.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sntp.service": { "name": "sntp.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sshd-keygen.service": { "name": "sshd-keygen.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sshd-keygen@.service": { "name": "sshd-keygen@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@ecdsa.service": { "name": "sshd-keygen@ecdsa.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd-keygen@ed25519.service": { "name": "sshd-keygen@ed25519.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd-keygen@rsa.service": { "name": "sshd-keygen@rsa.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd.service": { "name": "sshd.service", "source": "systemd", "state": "running", "status": "enabled" }, "sshd@.service": { "name": "sshd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "sssd-autofs.service": { "name": "sssd-autofs.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-kcm.service": { "name": "sssd-kcm.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "sssd-nss.service": { "name": "sssd-nss.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pac.service": { "name": "sssd-pac.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pam.service": { "name": "sssd-pam.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-ssh.service": { "name": "sssd-ssh.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-sudo.service": { "name": "sssd-sudo.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd.service": { "name": "sssd.service", "source": "systemd", "state": "running", "status": "enabled" }, "syslog.service": { "name": "syslog.service", "source": "systemd", "state": "active", "status": "enabled" }, "system-update-cleanup.service": { "name": "system-update-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-ask-password-console.service": { "name": "systemd-ask-password-console.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-wall.service": { "name": "systemd-ask-password-wall.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-backlight@.service": { "name": "systemd-backlight@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-binfmt.service": { "name": "systemd-binfmt.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-coredump@.service": { "name": "systemd-coredump@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-exit.service": { "name": "systemd-exit.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-firstboot.service": { "name": "systemd-firstboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck-root.service": { "name": "systemd-fsck-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck@.service": { "name": "systemd-fsck@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-fsck@dev-disk-by\\x2duuid-7B77\\x2d95E7.service": { "name": "systemd-fsck@dev-disk-by\\x2duuid-7B77\\x2d95E7.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-fsck@dev-vdb1.service": { "name": "systemd-fsck@dev-vdb1.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-fsck@dev-vdc1.service": { "name": "systemd-fsck@dev-vdc1.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-halt.service": { "name": "systemd-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hibernate-resume@.service": { "name": "systemd-hibernate-resume@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-hibernate.service": { "name": "systemd-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hostnamed.service": { "name": "systemd-hostnamed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hwdb-update.service": { "name": "systemd-hwdb-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-hybrid-sleep.service": { "name": "systemd-hybrid-sleep.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-initctl.service": { "name": "systemd-initctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-catalog-update.service": { "name": "systemd-journal-catalog-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-flush.service": { "name": "systemd-journal-flush.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journald.service": { "name": "systemd-journald.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-kexec.service": { "name": "systemd-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-localed.service": { "name": "systemd-localed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-logind.service": { "name": "systemd-logind.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-machine-id-commit.service": { "name": "systemd-machine-id-commit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-modules-load.service": { "name": "systemd-modules-load.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-networkd-wait-online.service": { "name": "systemd-networkd-wait-online.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "systemd-portabled.service": { "name": "systemd-portabled.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-poweroff.service": { "name": "systemd-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-quotacheck.service": { "name": "systemd-quotacheck.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-random-seed.service": { "name": "systemd-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-reboot.service": { "name": "systemd-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-remount-fs.service": { "name": "systemd-remount-fs.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-resolved.service": { "name": "systemd-resolved.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-rfkill.service": { "name": "systemd-rfkill.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend-then-hibernate.service": { "name": "systemd-suspend-then-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend.service": { "name": "systemd-suspend.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-sysctl.service": { "name": "systemd-sysctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-sysusers.service": { "name": "systemd-sysusers.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-timedated.service": { "name": "systemd-timedated.service", "source": "systemd", "state": "inactive", "status": "masked" }, "systemd-timesyncd.service": { "name": "systemd-timesyncd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "systemd-tmpfiles-clean.service": { "name": "systemd-tmpfiles-clean.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev.service": { "name": "systemd-tmpfiles-setup-dev.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup.service": { "name": "systemd-tmpfiles-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-settle.service": { "name": "systemd-udev-settle.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-trigger.service": { "name": "systemd-udev-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udevd.service": { "name": "systemd-udevd.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-update-done.service": { "name": "systemd-update-done.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp-runlevel.service": { "name": "systemd-update-utmp-runlevel.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp.service": { "name": "systemd-update-utmp.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-user-sessions.service": { "name": "systemd-user-sessions.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-vconsole-setup.service": { "name": "systemd-vconsole-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-volatile-root.service": { "name": "systemd-volatile-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "tcsd.service": { "name": "tcsd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "teamd@.service": { "name": "teamd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "timedatex.service": { "name": "timedatex.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "tuned.service": { "name": "tuned.service", "source": "systemd", "state": "running", "status": "enabled" }, "unbound-anchor.service": { "name": "unbound-anchor.service", "source": "systemd", "state": "stopped", "status": "static" }, "user-runtime-dir@.service": { "name": "user-runtime-dir@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user-runtime-dir@0.service": { "name": "user-runtime-dir@0.service", "source": "systemd", "state": "stopped", "status": "active" }, "user@.service": { "name": "user@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user@0.service": { "name": "user@0.service", "source": "systemd", "state": "running", "status": "active" }, "vdo-start-by-dev@.service": { "name": "vdo-start-by-dev@.service", "source": "systemd", "state": "unknown", "status": "static" }, "vdo.service": { "name": "vdo.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "ypbind.service": { "name": "ypbind.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "yppasswdd.service": { "name": "yppasswdd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ypserv.service": { "name": "ypserv.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ypxfrd.service": { "name": "ypxfrd.service", "source": "systemd", "state": "stopped", "status": "not-found" } } }, "changed": false } TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] ***** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:53 Thursday 21 July 2022 19:05:20 +0000 (0:00:02.727) 0:00:37.777 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:58 Thursday 21 July 2022 19:05:20 +0000 (0:00:00.056) 0:00:37.834 ********* TASK [fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64 Thursday 21 July 2022 19:05:20 +0000 (0:00:00.021) 0:00:37.855 ********* changed: [/cache/rhel-8-y.qcow2.snap] => { "actions": [ { "action": "create format", "device": "/dev/sda", "fs_type": "lvmpv" }, { "action": "create device", "device": "/dev/foo", "fs_type": null }, { "action": "create device", "device": "/dev/mapper/foo-test2", "fs_type": null }, { "action": "create format", "device": "/dev/mapper/foo-test2", "fs_type": "ext4" }, { "action": "create device", "device": "/dev/mapper/foo-test1", "fs_type": null }, { "action": "create format", "device": "/dev/mapper/foo-test1", "fs_type": "xfs" } ], "changed": true, "crypts": [], "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/vda2", "/dev/vda3", "/dev/sdb", "/dev/sdc", "/dev/nvme0n1", "/dev/nvme1n1", "/dev/nvme2n1", "/dev/vdb1", "/dev/vdc1", "/dev/vdd", "/dev/vde", "/dev/vdf", "/dev/mapper/foo-test1", "/dev/mapper/foo-test2" ], "mounts": [ { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" }, { "dump": 0, "fstype": "ext4", "opts": "defaults", "passno": 0, "path": "/opt/test2", "src": "/dev/mapper/foo-test2", "state": "mounted" } ], "packages": [ "xfsprogs", "dosfstools", "lvm2", "e2fsprogs" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "60%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test2", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test2", "_raw_device": "/dev/mapper/foo-test2", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "40%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:78 Thursday 21 July 2022 19:05:22 +0000 (0:00:02.302) 0:00:40.158 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90 Thursday 21 July 2022 19:05:22 +0000 (0:00:00.036) 0:00:40.194 ********* TASK [fedora.linux_system_roles.storage : show blivet_output] ****************** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:96 Thursday 21 July 2022 19:05:22 +0000 (0:00:00.021) 0:00:40.216 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "blivet_output": { "actions": [ { "action": "create format", "device": "/dev/sda", "fs_type": "lvmpv" }, { "action": "create device", "device": "/dev/foo", "fs_type": null }, { "action": "create device", "device": "/dev/mapper/foo-test2", "fs_type": null }, { "action": "create format", "device": "/dev/mapper/foo-test2", "fs_type": "ext4" }, { "action": "create device", "device": "/dev/mapper/foo-test1", "fs_type": null }, { "action": "create format", "device": "/dev/mapper/foo-test1", "fs_type": "xfs" } ], "changed": true, "crypts": [], "failed": false, "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/vda2", "/dev/vda3", "/dev/sdb", "/dev/sdc", "/dev/nvme0n1", "/dev/nvme1n1", "/dev/nvme2n1", "/dev/vdb1", "/dev/vdc1", "/dev/vdd", "/dev/vde", "/dev/vdf", "/dev/mapper/foo-test1", "/dev/mapper/foo-test2" ], "mounts": [ { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" }, { "dump": 0, "fstype": "ext4", "opts": "defaults", "passno": 0, "path": "/opt/test2", "src": "/dev/mapper/foo-test2", "state": "mounted" } ], "packages": [ "xfsprogs", "dosfstools", "lvm2", "e2fsprogs" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "60%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test2", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test2", "_raw_device": "/dev/mapper/foo-test2", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "40%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } } TASK [fedora.linux_system_roles.storage : set the list of pools for test verification] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:101 Thursday 21 July 2022 19:05:22 +0000 (0:00:00.041) 0:00:40.257 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "60%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test2", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test2", "_raw_device": "/dev/mapper/foo-test2", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "40%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] }, "changed": false } TASK [fedora.linux_system_roles.storage : set the list of volumes for test verification] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:105 Thursday 21 July 2022 19:05:22 +0000 (0:00:00.040) 0:00:40.298 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : remove obsolete mounts] ************** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:121 Thursday 21 July 2022 19:05:22 +0000 (0:00:00.035) 0:00:40.333 ********* TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132 Thursday 21 July 2022 19:05:22 +0000 (0:00:00.035) 0:00:40.369 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.storage : set up new/current mounts] *********** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:137 Thursday 21 July 2022 19:05:23 +0000 (0:00:00.878) 0:00:41.247 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount changed: [/cache/rhel-8-y.qcow2.snap] => (item={'src': '/dev/mapper/foo-test1', 'path': '/opt/test1', 'fstype': 'xfs', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted'}) => { "ansible_loop_var": "mount_info", "backup_file": "", "boot": "yes", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" }, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test1" } redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount changed: [/cache/rhel-8-y.qcow2.snap] => (item={'src': '/dev/mapper/foo-test2', 'path': '/opt/test2', 'fstype': 'ext4', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted'}) => { "ansible_loop_var": "mount_info", "backup_file": "", "boot": "yes", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "ext4", "mount_info": { "dump": 0, "fstype": "ext4", "opts": "defaults", "passno": 0, "path": "/opt/test2", "src": "/dev/mapper/foo-test2", "state": "mounted" }, "name": "/opt/test2", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test2" } TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148 Thursday 21 July 2022 19:05:24 +0000 (0:00:00.927) 0:00:42.175 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.storage : retrieve facts for the /etc/crypttab file] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:156 Thursday 21 July 2022 19:05:25 +0000 (0:00:00.674) 0:00:42.850 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "stat": { "exists": false } } TASK [fedora.linux_system_roles.storage : manage /etc/crypttab to account for changes we just made] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:161 Thursday 21 July 2022 19:05:25 +0000 (0:00:00.376) 0:00:43.227 ********* TASK [fedora.linux_system_roles.storage : Update facts] ************************ task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:183 Thursday 21 July 2022 19:05:25 +0000 (0:00:00.024) 0:00:43.251 ********* ok: [/cache/rhel-8-y.qcow2.snap] META: role_complete for /cache/rhel-8-y.qcow2.snap TASK [include_tasks] *********************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/tests_lvm_percent_size.yml:79 Thursday 21 July 2022 19:05:26 +0000 (0:00:00.997) 0:00:44.248 ********* included: /tmp/tmpmnqeja9w/tests/storage/verify-role-results.yml for /cache/rhel-8-y.qcow2.snap TASK [Print out pool information] ********************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-role-results.yml:1 Thursday 21 July 2022 19:05:26 +0000 (0:00:00.042) 0:00:44.291 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "60%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test2", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test2", "_raw_device": "/dev/mapper/foo-test2", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "40%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] } TASK [Print out volume information] ******************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-role-results.yml:6 Thursday 21 July 2022 19:05:26 +0000 (0:00:00.049) 0:00:44.340 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => {} TASK [Collect info about the volumes.] ***************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-role-results.yml:14 Thursday 21 July 2022 19:05:26 +0000 (0:00:00.034) 0:00:44.374 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "info": { "/dev/mapper/foo-test1": { "fstype": "xfs", "label": "", "name": "/dev/mapper/foo-test1", "size": "6G", "type": "lvm", "uuid": "d739e41d-c1a1-4003-b357-25bf62753af5" }, "/dev/mapper/foo-test2": { "fstype": "ext4", "label": "", "name": "/dev/mapper/foo-test2", "size": "4G", "type": "lvm", "uuid": "64454f89-f2cd-4a8b-958f-4c900f72cf79" }, "/dev/nvme0n1": { "fstype": "", "label": "", "name": "/dev/nvme0n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme1n1": { "fstype": "", "label": "", "name": "/dev/nvme1n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme2n1": { "fstype": "", "label": "", "name": "/dev/nvme2n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sda": { "fstype": "LVM2_member", "label": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "HA17fj-P3dB-xyZl-9Zru-ZzXr-ULh6-93Sgui" }, "/dev/sdb": { "fstype": "", "label": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc": { "fstype": "", "label": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sr0": { "fstype": "iso9660", "label": "cidata", "name": "/dev/sr0", "size": "364K", "type": "rom", "uuid": "2022-07-21-19-04-29-00" }, "/dev/vda": { "fstype": "", "label": "", "name": "/dev/vda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vda1": { "fstype": "", "label": "", "name": "/dev/vda1", "size": "1M", "type": "partition", "uuid": "" }, "/dev/vda2": { "fstype": "vfat", "label": "", "name": "/dev/vda2", "size": "100M", "type": "partition", "uuid": "7B77-95E7" }, "/dev/vda3": { "fstype": "xfs", "label": "root", "name": "/dev/vda3", "size": "9.9G", "type": "partition", "uuid": "58dca2db-4ccd-4171-a017-dd53ad8c1c54" }, "/dev/vdb": { "fstype": "", "label": "", "name": "/dev/vdb", "size": "2G", "type": "disk", "uuid": "" }, "/dev/vdb1": { "fstype": "ext4", "label": "yumcache", "name": "/dev/vdb1", "size": "2G", "type": "partition", "uuid": "4d1ed9a1-5608-4ff8-ad89-21d95caa66fa" }, "/dev/vdc": { "fstype": "", "label": "", "name": "/dev/vdc", "size": "2G", "type": "disk", "uuid": "" }, "/dev/vdc1": { "fstype": "ext4", "label": "yumvarlib", "name": "/dev/vdc1", "size": "2G", "type": "partition", "uuid": "a7abcc97-edc5-476c-8ccd-694b4621fd28" }, "/dev/vdd": { "fstype": "", "label": "", "name": "/dev/vdd", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vde": { "fstype": "", "label": "", "name": "/dev/vde", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vdf": { "fstype": "", "label": "", "name": "/dev/vdf", "size": "10G", "type": "disk", "uuid": "" } } } TASK [Read the /etc/fstab file for volume existence] *************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-role-results.yml:19 Thursday 21 July 2022 19:05:27 +0000 (0:00:00.501) 0:00:44.876 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "cmd": [ "cat", "/etc/fstab" ], "delta": "0:00:00.003362", "end": "2022-07-21 15:05:27.900003", "rc": 0, "start": "2022-07-21 15:05:27.896641" } STDOUT: UUID=58dca2db-4ccd-4171-a017-dd53ad8c1c54 / xfs defaults 0 0 UUID=7B77-95E7 /boot/efi vfat defaults,uid=0,gid=0,umask=077,shortname=winnt 0 2 /dev/vdb1 /var/cache/dnf auto defaults,nofail,x-systemd.requires=cloud-init.service,_netdev,comment=cloudconfig 0 2 /dev/vdc1 /var/lib/dnf auto defaults,nofail,x-systemd.requires=cloud-init.service,_netdev,comment=cloudconfig 0 2 /dev/mapper/foo-test1 /opt/test1 xfs defaults 0 0 /dev/mapper/foo-test2 /opt/test2 ext4 defaults 0 0 TASK [Read the /etc/crypttab file] ********************************************* task path: /tmp/tmpmnqeja9w/tests/storage/verify-role-results.yml:24 Thursday 21 July 2022 19:05:27 +0000 (0:00:00.504) 0:00:45.380 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "cmd": [ "cat", "/etc/crypttab" ], "delta": "0:00:00.002799", "end": "2022-07-21 15:05:28.268403", "failed_when_result": false, "rc": 1, "start": "2022-07-21 15:05:28.265604" } STDERR: cat: /etc/crypttab: No such file or directory MSG: non-zero return code TASK [Verify the volumes listed in storage_pools were correctly managed] ******* task path: /tmp/tmpmnqeja9w/tests/storage/verify-role-results.yml:33 Thursday 21 July 2022 19:05:28 +0000 (0:00:00.369) 0:00:45.749 ********* included: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool.yml for /cache/rhel-8-y.qcow2.snap => (item={'disks': ['sda'], 'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'name': 'foo', 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_metadata_version': None, 'state': 'present', 'type': 'lvm', 'volumes': [{'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'name': 'test1', 'raid_level': None, 'size': '60%', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': [], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test1', '_raw_device': '/dev/mapper/foo-test1', '_mount_id': '/dev/mapper/foo-test1', '_kernel_device': '/dev/dm-1', '_raw_kernel_device': '/dev/dm-1'}, {'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'ext4', 'mount_options': 'defaults', 'mount_point': '/opt/test2', 'name': 'test2', 'raid_level': None, 'size': '40%', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': [], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test2', '_raw_device': '/dev/mapper/foo-test2', '_mount_id': '/dev/mapper/foo-test2', '_kernel_device': '/dev/dm-0', '_raw_kernel_device': '/dev/dm-0'}], 'raid_chunk_size': None}) TASK [Set _storage_pool_tests] ************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool.yml:5 Thursday 21 July 2022 19:05:28 +0000 (0:00:00.061) 0:00:45.811 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_pool_tests": [ "members", "volumes" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool.yml:18 Thursday 21 July 2022 19:05:28 +0000 (0:00:00.033) 0:00:45.844 ********* included: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml for /cache/rhel-8-y.qcow2.snap => (item=members) included: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-volumes.yml for /cache/rhel-8-y.qcow2.snap => (item=volumes) TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:1 Thursday 21 July 2022 19:05:28 +0000 (0:00:00.051) 0:00:45.896 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_pv_count": "1", "_storage_test_pool_pvs_lvm": [ "/dev/sda" ] }, "changed": false } TASK [Get the canonical device path for each member device] ******************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:6 Thursday 21 July 2022 19:05:28 +0000 (0:00:00.085) 0:00:45.982 ********* ok: [/cache/rhel-8-y.qcow2.snap] => (item=/dev/sda) => { "ansible_loop_var": "pv", "changed": false, "device": "/dev/sda", "pv": "/dev/sda" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:15 Thursday 21 July 2022 19:05:28 +0000 (0:00:00.511) 0:00:46.494 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "__pvs_lvm_len": "1" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:19 Thursday 21 July 2022 19:05:29 +0000 (0:00:00.080) 0:00:46.574 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_pool_pvs": [ "/dev/sda" ] }, "changed": false } TASK [Verify PV count] ********************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:23 Thursday 21 July 2022 19:05:29 +0000 (0:00:00.116) 0:00:46.691 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:29 Thursday 21 July 2022 19:05:29 +0000 (0:00:00.053) 0:00:46.744 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:33 Thursday 21 July 2022 19:05:29 +0000 (0:00:00.040) 0:00:46.785 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:37 Thursday 21 July 2022 19:05:29 +0000 (0:00:00.049) 0:00:46.834 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check the type of each PV] *********************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:41 Thursday 21 July 2022 19:05:29 +0000 (0:00:00.022) 0:00:46.857 ********* ok: [/cache/rhel-8-y.qcow2.snap] => (item=/dev/sda) => { "ansible_loop_var": "pv", "changed": false, "pv": "/dev/sda" } MSG: All assertions passed TASK [Check MD RAID] *********************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:50 Thursday 21 July 2022 19:05:29 +0000 (0:00:00.040) 0:00:46.898 ********* included: /tmp/tmpmnqeja9w/tests/storage/verify-pool-md.yml for /cache/rhel-8-y.qcow2.snap TASK [get information about RAID] ********************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-md.yml:6 Thursday 21 July 2022 19:05:29 +0000 (0:00:00.045) 0:00:46.943 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-md.yml:12 Thursday 21 July 2022 19:05:29 +0000 (0:00:00.026) 0:00:46.969 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-md.yml:16 Thursday 21 July 2022 19:05:29 +0000 (0:00:00.025) 0:00:46.995 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-md.yml:20 Thursday 21 July 2022 19:05:29 +0000 (0:00:00.025) 0:00:47.020 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-md.yml:24 Thursday 21 July 2022 19:05:29 +0000 (0:00:00.024) 0:00:47.045 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-md.yml:30 Thursday 21 July 2022 19:05:29 +0000 (0:00:00.023) 0:00:47.068 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-md.yml:36 Thursday 21 July 2022 19:05:29 +0000 (0:00:00.025) 0:00:47.094 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-md.yml:44 Thursday 21 July 2022 19:05:29 +0000 (0:00:00.025) 0:00:47.120 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_md_active_devices_re": null, "storage_test_md_metadata_version_re": null, "storage_test_md_spare_devices_re": null }, "changed": false } TASK [Check LVM RAID] ********************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:53 Thursday 21 July 2022 19:05:29 +0000 (0:00:00.039) 0:00:47.159 ********* included: /tmp/tmpmnqeja9w/tests/storage/verify-pool-members-lvmraid.yml for /cache/rhel-8-y.qcow2.snap TASK [Validate pool member LVM RAID settings] ********************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-members-lvmraid.yml:1 Thursday 21 July 2022 19:05:29 +0000 (0:00:00.045) 0:00:47.204 ********* included: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-lvmraid.yml for /cache/rhel-8-y.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'name': 'test1', 'raid_level': None, 'size': '60%', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': [], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test1', '_raw_device': '/dev/mapper/foo-test1', '_mount_id': '/dev/mapper/foo-test1', '_kernel_device': '/dev/dm-1', '_raw_kernel_device': '/dev/dm-1'}) included: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-lvmraid.yml for /cache/rhel-8-y.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'ext4', 'mount_options': 'defaults', 'mount_point': '/opt/test2', 'name': 'test2', 'raid_level': None, 'size': '40%', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': [], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test2', '_raw_device': '/dev/mapper/foo-test2', '_mount_id': '/dev/mapper/foo-test2', '_kernel_device': '/dev/dm-0', '_raw_kernel_device': '/dev/dm-0'}) TASK [Get information about LVM RAID] ****************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-lvmraid.yml:3 Thursday 21 July 2022 19:05:29 +0000 (0:00:00.048) 0:00:47.253 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is LVM RAID] ******************************************* task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-lvmraid.yml:8 Thursday 21 July 2022 19:05:29 +0000 (0:00:00.026) 0:00:47.280 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-lvmraid.yml:12 Thursday 21 July 2022 19:05:29 +0000 (0:00:00.026) 0:00:47.306 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about LVM RAID] ****************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-lvmraid.yml:3 Thursday 21 July 2022 19:05:29 +0000 (0:00:00.030) 0:00:47.336 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is LVM RAID] ******************************************* task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-lvmraid.yml:8 Thursday 21 July 2022 19:05:29 +0000 (0:00:00.029) 0:00:47.366 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-lvmraid.yml:12 Thursday 21 July 2022 19:05:29 +0000 (0:00:00.029) 0:00:47.395 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check Thin Pools] ******************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:56 Thursday 21 July 2022 19:05:29 +0000 (0:00:00.028) 0:00:47.423 ********* included: /tmp/tmpmnqeja9w/tests/storage/verify-pool-members-thin.yml for /cache/rhel-8-y.qcow2.snap TASK [Validate pool member thinpool settings] ********************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-members-thin.yml:1 Thursday 21 July 2022 19:05:29 +0000 (0:00:00.047) 0:00:47.470 ********* included: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-thin.yml for /cache/rhel-8-y.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'name': 'test1', 'raid_level': None, 'size': '60%', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': [], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test1', '_raw_device': '/dev/mapper/foo-test1', '_mount_id': '/dev/mapper/foo-test1', '_kernel_device': '/dev/dm-1', '_raw_kernel_device': '/dev/dm-1'}) included: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-thin.yml for /cache/rhel-8-y.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'ext4', 'mount_options': 'defaults', 'mount_point': '/opt/test2', 'name': 'test2', 'raid_level': None, 'size': '40%', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': [], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test2', '_raw_device': '/dev/mapper/foo-test2', '_mount_id': '/dev/mapper/foo-test2', '_kernel_device': '/dev/dm-0', '_raw_kernel_device': '/dev/dm-0'}) TASK [Get information about thinpool] ****************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-thin.yml:3 Thursday 21 July 2022 19:05:29 +0000 (0:00:00.055) 0:00:47.525 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-thin.yml:8 Thursday 21 July 2022 19:05:30 +0000 (0:00:00.026) 0:00:47.552 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-thin.yml:13 Thursday 21 July 2022 19:05:30 +0000 (0:00:00.023) 0:00:47.576 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-thin.yml:17 Thursday 21 July 2022 19:05:30 +0000 (0:00:00.023) 0:00:47.599 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about thinpool] ****************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-thin.yml:3 Thursday 21 July 2022 19:05:30 +0000 (0:00:00.024) 0:00:47.623 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-thin.yml:8 Thursday 21 July 2022 19:05:30 +0000 (0:00:00.022) 0:00:47.646 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-thin.yml:13 Thursday 21 July 2022 19:05:30 +0000 (0:00:00.022) 0:00:47.668 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-thin.yml:17 Thursday 21 July 2022 19:05:30 +0000 (0:00:00.024) 0:00:47.693 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check member encryption] ************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:59 Thursday 21 July 2022 19:05:30 +0000 (0:00:00.024) 0:00:47.717 ********* included: /tmp/tmpmnqeja9w/tests/storage/verify-pool-members-encryption.yml for /cache/rhel-8-y.qcow2.snap TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-members-encryption.yml:4 Thursday 21 July 2022 19:05:30 +0000 (0:00:00.046) 0:00:47.763 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Validate pool member LUKS settings] ************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-members-encryption.yml:8 Thursday 21 July 2022 19:05:30 +0000 (0:00:00.110) 0:00:47.874 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => (item=/dev/sda) => { "_storage_test_pool_member_path": "/dev/sda", "ansible_loop_var": "_storage_test_pool_member_path", "changed": false, "skip_reason": "Conditional result was False" } TASK [Validate pool member crypttab entries] *********************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-members-encryption.yml:15 Thursday 21 July 2022 19:05:30 +0000 (0:00:00.028) 0:00:47.903 ********* included: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-crypttab.yml for /cache/rhel-8-y.qcow2.snap => (item=/dev/sda) TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-crypttab.yml:1 Thursday 21 July 2022 19:05:30 +0000 (0:00:00.046) 0:00:47.950 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": [] }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-crypttab.yml:4 Thursday 21 July 2022 19:05:30 +0000 (0:00:00.053) 0:00:48.003 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-crypttab.yml:9 Thursday 21 July 2022 19:05:30 +0000 (0:00:00.051) 0:00:48.055 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-crypttab.yml:15 Thursday 21 July 2022 19:05:30 +0000 (0:00:00.038) 0:00:48.094 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-crypttab.yml:21 Thursday 21 July 2022 19:05:30 +0000 (0:00:00.037) 0:00:48.131 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-crypttab.yml:27 Thursday 21 July 2022 19:05:30 +0000 (0:00:00.040) 0:00:48.172 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-members-encryption.yml:22 Thursday 21 July 2022 19:05:30 +0000 (0:00:00.035) 0:00:48.207 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_crypttab_key_file": null }, "changed": false } TASK [Check VDO] *************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:62 Thursday 21 July 2022 19:05:30 +0000 (0:00:00.036) 0:00:48.244 ********* included: /tmp/tmpmnqeja9w/tests/storage/verify-pool-members-vdo.yml for /cache/rhel-8-y.qcow2.snap TASK [Validate pool member VDO settings] *************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-members-vdo.yml:1 Thursday 21 July 2022 19:05:30 +0000 (0:00:00.052) 0:00:48.296 ********* included: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-vdo.yml for /cache/rhel-8-y.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'name': 'test1', 'raid_level': None, 'size': '60%', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': [], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test1', '_raw_device': '/dev/mapper/foo-test1', '_mount_id': '/dev/mapper/foo-test1', '_kernel_device': '/dev/dm-1', '_raw_kernel_device': '/dev/dm-1'}) included: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-vdo.yml for /cache/rhel-8-y.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'ext4', 'mount_options': 'defaults', 'mount_point': '/opt/test2', 'name': 'test2', 'raid_level': None, 'size': '40%', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': [], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test2', '_raw_device': '/dev/mapper/foo-test2', '_mount_id': '/dev/mapper/foo-test2', '_kernel_device': '/dev/dm-0', '_raw_kernel_device': '/dev/dm-0'}) TASK [get information about VDO deduplication] ********************************* task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-vdo.yml:3 Thursday 21 July 2022 19:05:30 +0000 (0:00:00.054) 0:00:48.351 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-vdo.yml:8 Thursday 21 July 2022 19:05:30 +0000 (0:00:00.026) 0:00:48.377 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-vdo.yml:11 Thursday 21 July 2022 19:05:30 +0000 (0:00:00.026) 0:00:48.403 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-vdo.yml:16 Thursday 21 July 2022 19:05:30 +0000 (0:00:00.026) 0:00:48.430 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-vdo.yml:21 Thursday 21 July 2022 19:05:30 +0000 (0:00:00.025) 0:00:48.456 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-vdo.yml:24 Thursday 21 July 2022 19:05:30 +0000 (0:00:00.025) 0:00:48.481 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-vdo.yml:29 Thursday 21 July 2022 19:05:30 +0000 (0:00:00.025) 0:00:48.507 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-vdo.yml:39 Thursday 21 July 2022 19:05:30 +0000 (0:00:00.023) 0:00:48.530 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [get information about VDO deduplication] ********************************* task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-vdo.yml:3 Thursday 21 July 2022 19:05:31 +0000 (0:00:00.035) 0:00:48.566 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-vdo.yml:8 Thursday 21 July 2022 19:05:31 +0000 (0:00:00.024) 0:00:48.590 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-vdo.yml:11 Thursday 21 July 2022 19:05:31 +0000 (0:00:00.025) 0:00:48.615 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-vdo.yml:16 Thursday 21 July 2022 19:05:31 +0000 (0:00:00.025) 0:00:48.641 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-vdo.yml:21 Thursday 21 July 2022 19:05:31 +0000 (0:00:00.027) 0:00:48.668 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-vdo.yml:24 Thursday 21 July 2022 19:05:31 +0000 (0:00:00.023) 0:00:48.692 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-vdo.yml:29 Thursday 21 July 2022 19:05:31 +0000 (0:00:00.025) 0:00:48.717 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-vdo.yml:39 Thursday 21 July 2022 19:05:31 +0000 (0:00:00.023) 0:00:48.741 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:65 Thursday 21 July 2022 19:05:31 +0000 (0:00:00.038) 0:00:48.779 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "__pvs_lvm_len": null, "_storage_test_expected_pv_count": null, "_storage_test_expected_pv_type": null, "_storage_test_pool_pvs": [], "_storage_test_pool_pvs_lvm": [] }, "changed": false } TASK [verify the volumes] ****************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-volumes.yml:3 Thursday 21 July 2022 19:05:31 +0000 (0:00:00.033) 0:00:48.812 ********* included: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume.yml for /cache/rhel-8-y.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'name': 'test1', 'raid_level': None, 'size': '60%', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': [], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test1', '_raw_device': '/dev/mapper/foo-test1', '_mount_id': '/dev/mapper/foo-test1', '_kernel_device': '/dev/dm-1', '_raw_kernel_device': '/dev/dm-1'}) included: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume.yml for /cache/rhel-8-y.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'ext4', 'mount_options': 'defaults', 'mount_point': '/opt/test2', 'name': 'test2', 'raid_level': None, 'size': '40%', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': [], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test2', '_raw_device': '/dev/mapper/foo-test2', '_mount_id': '/dev/mapper/foo-test2', '_kernel_device': '/dev/dm-0', '_raw_kernel_device': '/dev/dm-0'}) TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume.yml:2 Thursday 21 July 2022 19:05:31 +0000 (0:00:00.056) 0:00:48.869 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_volume_present": true, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume.yml:10 Thursday 21 July 2022 19:05:31 +0000 (0:00:00.065) 0:00:48.935 ********* included: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml for /cache/rhel-8-y.qcow2.snap => (item=mount) included: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-fstab.yml for /cache/rhel-8-y.qcow2.snap => (item=fstab) included: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-fs.yml for /cache/rhel-8-y.qcow2.snap => (item=fs) included: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-device.yml for /cache/rhel-8-y.qcow2.snap => (item=device) included: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml for /cache/rhel-8-y.qcow2.snap => (item=encryption) included: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-md.yml for /cache/rhel-8-y.qcow2.snap => (item=md) included: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml for /cache/rhel-8-y.qcow2.snap => (item=size) included: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-cache.yml for /cache/rhel-8-y.qcow2.snap => (item=cache) TASK [Get expected mount device based on device type] ************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml:6 Thursday 21 July 2022 19:05:31 +0000 (0:00:00.086) 0:00:49.022 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/foo-test1" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml:10 Thursday 21 July 2022 19:05:31 +0000 (0:00:00.142) 0:00:49.165 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_mount_device_matches": [ { "block_available": 1551090, "block_size": 4096, "block_total": 1570304, "block_used": 19214, "device": "/dev/mapper/foo-test1", "fstype": "xfs", "inode_available": 3145725, "inode_total": 3145728, "inode_used": 3, "mount": "/opt/test1", "options": "rw,seclabel,relatime,attr2,inode64,logbufs=8,logbsize=32k,noquota", "size_available": 6353264640, "size_total": 6431965184, "uuid": "d739e41d-c1a1-4003-b357-25bf62753af5" } ], "storage_test_mount_expected_match_count": "1", "storage_test_mount_point_matches": [ { "block_available": 1551090, "block_size": 4096, "block_total": 1570304, "block_used": 19214, "device": "/dev/mapper/foo-test1", "fstype": "xfs", "inode_available": 3145725, "inode_total": 3145728, "inode_used": 3, "mount": "/opt/test1", "options": "rw,seclabel,relatime,attr2,inode64,logbufs=8,logbsize=32k,noquota", "size_available": 6353264640, "size_total": 6431965184, "uuid": "d739e41d-c1a1-4003-b357-25bf62753af5" } ], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Verify the current mount state by device] ******************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml:20 Thursday 21 July 2022 19:05:31 +0000 (0:00:00.056) 0:00:49.221 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify the current mount state by mount point] *************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml:29 Thursday 21 July 2022 19:05:31 +0000 (0:00:00.051) 0:00:49.272 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify the mount fs type] ************************************************ task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml:37 Thursday 21 July 2022 19:05:31 +0000 (0:00:00.048) 0:00:49.320 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [command] ***************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml:46 Thursday 21 July 2022 19:05:31 +0000 (0:00:00.047) 0:00:49.368 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml:50 Thursday 21 July 2022 19:05:31 +0000 (0:00:00.023) 0:00:49.392 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml:55 Thursday 21 July 2022 19:05:31 +0000 (0:00:00.023) 0:00:49.415 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml:65 Thursday 21 July 2022 19:05:31 +0000 (0:00:00.024) 0:00:49.439 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-fstab.yml:2 Thursday 21 July 2022 19:05:31 +0000 (0:00:00.032) 0:00:49.472 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "1", "storage_test_fstab_expected_mount_options_matches": "1", "storage_test_fstab_expected_mount_point_matches": "1", "storage_test_fstab_id_matches": [ "/dev/mapper/foo-test1 " ], "storage_test_fstab_mount_options_matches": [ " /opt/test1 xfs defaults " ], "storage_test_fstab_mount_point_matches": [ " /opt/test1 " ] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-fstab.yml:12 Thursday 21 July 2022 19:05:31 +0000 (0:00:00.060) 0:00:49.533 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify the fstab mount point] ******************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-fstab.yml:19 Thursday 21 July 2022 19:05:32 +0000 (0:00:00.051) 0:00:49.584 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-fstab.yml:25 Thursday 21 July 2022 19:05:32 +0000 (0:00:00.052) 0:00:49.636 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-fstab.yml:34 Thursday 21 July 2022 19:05:32 +0000 (0:00:00.036) 0:00:49.673 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-fs.yml:4 Thursday 21 July 2022 19:05:32 +0000 (0:00:00.037) 0:00:49.710 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify fs label] ********************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-fs.yml:10 Thursday 21 July 2022 19:05:32 +0000 (0:00:00.042) 0:00:49.752 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [See whether the device node is present] ********************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-device.yml:4 Thursday 21 July 2022 19:05:32 +0000 (0:00:00.038) 0:00:49.791 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "stat": { "atime": 1658430322.6091166, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1658430322.6091166, "dev": 6, "device_type": 64769, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 60295, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/symlink", "mode": "0660", "mtime": 1658430322.6091166, "nlink": 1, "path": "/dev/mapper/foo-test1", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-device.yml:10 Thursday 21 July 2022 19:05:32 +0000 (0:00:00.381) 0:00:50.173 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about this volume] ********************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-device.yml:15 Thursday 21 July 2022 19:05:32 +0000 (0:00:00.041) 0:00:50.215 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [(1/2) Process volume type (set initial value)] *************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-device.yml:21 Thursday 21 July 2022 19:05:32 +0000 (0:00:00.039) 0:00:50.254 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [(2/2) Process volume type (get RAID value)] ****************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-device.yml:25 Thursday 21 July 2022 19:05:32 +0000 (0:00:00.036) 0:00:50.291 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-device.yml:30 Thursday 21 July 2022 19:05:32 +0000 (0:00:00.025) 0:00:50.316 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Stat the LUKS device, if encrypted] ************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:3 Thursday 21 July 2022 19:05:32 +0000 (0:00:00.041) 0:00:50.357 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:10 Thursday 21 July 2022 19:05:32 +0000 (0:00:00.026) 0:00:50.383 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [Collect LUKS info for this volume] *************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:15 Thursday 21 July 2022 19:05:34 +0000 (0:00:01.383) 0:00:51.767 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:21 Thursday 21 July 2022 19:05:34 +0000 (0:00:00.053) 0:00:51.820 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:27 Thursday 21 July 2022 19:05:34 +0000 (0:00:00.024) 0:00:51.845 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:33 Thursday 21 July 2022 19:05:34 +0000 (0:00:00.050) 0:00:51.896 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:39 Thursday 21 July 2022 19:05:34 +0000 (0:00:00.025) 0:00:51.921 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:44 Thursday 21 July 2022 19:05:34 +0000 (0:00:00.025) 0:00:51.947 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:50 Thursday 21 July 2022 19:05:34 +0000 (0:00:00.025) 0:00:51.972 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:56 Thursday 21 July 2022 19:05:34 +0000 (0:00:00.025) 0:00:51.998 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:62 Thursday 21 July 2022 19:05:34 +0000 (0:00:00.025) 0:00:52.024 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:67 Thursday 21 July 2022 19:05:34 +0000 (0:00:00.053) 0:00:52.077 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:72 Thursday 21 July 2022 19:05:34 +0000 (0:00:00.050) 0:00:52.128 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:78 Thursday 21 July 2022 19:05:34 +0000 (0:00:00.038) 0:00:52.166 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:84 Thursday 21 July 2022 19:05:34 +0000 (0:00:00.042) 0:00:52.209 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:90 Thursday 21 July 2022 19:05:34 +0000 (0:00:00.038) 0:00:52.247 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [get information about RAID] ********************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-md.yml:7 Thursday 21 July 2022 19:05:34 +0000 (0:00:00.036) 0:00:52.284 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-md.yml:13 Thursday 21 July 2022 19:05:34 +0000 (0:00:00.037) 0:00:52.321 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-md.yml:17 Thursday 21 July 2022 19:05:34 +0000 (0:00:00.042) 0:00:52.364 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-md.yml:21 Thursday 21 July 2022 19:05:34 +0000 (0:00:00.038) 0:00:52.402 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-md.yml:25 Thursday 21 July 2022 19:05:34 +0000 (0:00:00.038) 0:00:52.440 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-md.yml:31 Thursday 21 July 2022 19:05:34 +0000 (0:00:00.044) 0:00:52.484 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-md.yml:37 Thursday 21 July 2022 19:05:34 +0000 (0:00:00.046) 0:00:52.531 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the actual size of the volume] ************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:3 Thursday 21 July 2022 19:05:35 +0000 (0:00:00.043) 0:00:52.574 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "bytes": 6442450944, "changed": false, "lvm": "6g", "parted": "6GiB", "size": "6 GiB" } TASK [parse the requested size of the volume] ********************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:9 Thursday 21 July 2022 19:05:35 +0000 (0:00:00.520) 0:00:53.094 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Establish base value for expected size] ********************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:15 Thursday 21 July 2022 19:05:35 +0000 (0:00:00.042) 0:00:53.137 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:20 Thursday 21 July 2022 19:05:35 +0000 (0:00:00.037) 0:00:53.174 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_expected_size": "VARIABLE IS NOT DEFINED!: 'storage_test_expected_size' is undefined" } TASK [debug] ******************************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:25 Thursday 21 July 2022 19:05:35 +0000 (0:00:00.036) 0:00:53.211 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_pool": { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "60%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test2", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test2", "_raw_device": "/dev/mapper/foo-test2", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "40%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } } TASK [debug] ******************************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:28 Thursday 21 July 2022 19:05:35 +0000 (0:00:00.051) 0:00:53.262 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_blkinfo": { "changed": false, "failed": false, "info": { "/dev/mapper/foo-test1": { "fstype": "xfs", "label": "", "name": "/dev/mapper/foo-test1", "size": "6G", "type": "lvm", "uuid": "d739e41d-c1a1-4003-b357-25bf62753af5" }, "/dev/mapper/foo-test2": { "fstype": "ext4", "label": "", "name": "/dev/mapper/foo-test2", "size": "4G", "type": "lvm", "uuid": "64454f89-f2cd-4a8b-958f-4c900f72cf79" }, "/dev/nvme0n1": { "fstype": "", "label": "", "name": "/dev/nvme0n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme1n1": { "fstype": "", "label": "", "name": "/dev/nvme1n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme2n1": { "fstype": "", "label": "", "name": "/dev/nvme2n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sda": { "fstype": "LVM2_member", "label": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "HA17fj-P3dB-xyZl-9Zru-ZzXr-ULh6-93Sgui" }, "/dev/sdb": { "fstype": "", "label": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc": { "fstype": "", "label": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sr0": { "fstype": "iso9660", "label": "cidata", "name": "/dev/sr0", "size": "364K", "type": "rom", "uuid": "2022-07-21-19-04-29-00" }, "/dev/vda": { "fstype": "", "label": "", "name": "/dev/vda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vda1": { "fstype": "", "label": "", "name": "/dev/vda1", "size": "1M", "type": "partition", "uuid": "" }, "/dev/vda2": { "fstype": "vfat", "label": "", "name": "/dev/vda2", "size": "100M", "type": "partition", "uuid": "7B77-95E7" }, "/dev/vda3": { "fstype": "xfs", "label": "root", "name": "/dev/vda3", "size": "9.9G", "type": "partition", "uuid": "58dca2db-4ccd-4171-a017-dd53ad8c1c54" }, "/dev/vdb": { "fstype": "", "label": "", "name": "/dev/vdb", "size": "2G", "type": "disk", "uuid": "" }, "/dev/vdb1": { "fstype": "ext4", "label": "yumcache", "name": "/dev/vdb1", "size": "2G", "type": "partition", "uuid": "4d1ed9a1-5608-4ff8-ad89-21d95caa66fa" }, "/dev/vdc": { "fstype": "", "label": "", "name": "/dev/vdc", "size": "2G", "type": "disk", "uuid": "" }, "/dev/vdc1": { "fstype": "ext4", "label": "yumvarlib", "name": "/dev/vdc1", "size": "2G", "type": "partition", "uuid": "a7abcc97-edc5-476c-8ccd-694b4621fd28" }, "/dev/vdd": { "fstype": "", "label": "", "name": "/dev/vdd", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vde": { "fstype": "", "label": "", "name": "/dev/vde", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vdf": { "fstype": "", "label": "", "name": "/dev/vdf", "size": "10G", "type": "disk", "uuid": "" } } } } TASK [Get the size of parent/pool device] ************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:31 Thursday 21 July 2022 19:05:35 +0000 (0:00:00.057) 0:00:53.319 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "bytes": 10737418240, "changed": false, "lvm": "10g", "parted": "10GiB", "size": "10 GiB" } TASK [debug] ******************************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:36 Thursday 21 July 2022 19:05:36 +0000 (0:00:00.382) 0:00:53.702 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_pool_size": { "bytes": 10737418240, "changed": false, "failed": false, "lvm": "10g", "parted": "10GiB", "size": "10 GiB" } } TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:39 Thursday 21 July 2022 19:05:36 +0000 (0:00:00.049) 0:00:53.751 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_expected_size": "6442450944.0" }, "changed": false } TASK [debug] ******************************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:44 Thursday 21 July 2022 19:05:36 +0000 (0:00:00.085) 0:00:53.836 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_actual_size": { "bytes": 6442450944, "changed": false, "failed": false, "lvm": "6g", "parted": "6GiB", "size": "6 GiB" } } TASK [debug] ******************************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:47 Thursday 21 July 2022 19:05:36 +0000 (0:00:00.090) 0:00:53.927 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_expected_size": "6442450944.0" } TASK [assert] ****************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:50 Thursday 21 July 2022 19:05:36 +0000 (0:00:00.037) 0:00:53.965 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Get information about the LV] ******************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-cache.yml:6 Thursday 21 July 2022 19:05:36 +0000 (0:00:00.097) 0:00:54.062 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "cmd": [ "lvs", "--noheadings", "--nameprefixes", "--units=b", "--nosuffix", "--unquoted", "-o", "name,attr,cache_total_blocks,chunk_size,segtype", "foo/test1" ], "delta": "0:00:00.039117", "end": "2022-07-21 15:05:36.997834", "rc": 0, "start": "2022-07-21 15:05:36.958717" } STDOUT: LVM2_LV_NAME=test1 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-cache.yml:14 Thursday 21 July 2022 19:05:36 +0000 (0:00:00.413) 0:00:54.476 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_lv_segtype": [ "linear" ] }, "changed": false } TASK [check segment type] ****************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-cache.yml:17 Thursday 21 July 2022 19:05:37 +0000 (0:00:00.081) 0:00:54.557 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-cache.yml:22 Thursday 21 July 2022 19:05:37 +0000 (0:00:00.126) 0:00:54.684 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the requested cache size] ****************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-cache.yml:26 Thursday 21 July 2022 19:05:37 +0000 (0:00:00.040) 0:00:54.724 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-cache.yml:32 Thursday 21 July 2022 19:05:37 +0000 (0:00:00.040) 0:00:54.764 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-cache.yml:36 Thursday 21 July 2022 19:05:37 +0000 (0:00:00.039) 0:00:54.804 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume.yml:16 Thursday 21 July 2022 19:05:37 +0000 (0:00:00.036) 0:00:54.841 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume.yml:2 Thursday 21 July 2022 19:05:37 +0000 (0:00:00.033) 0:00:54.874 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_volume_present": true, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume.yml:10 Thursday 21 July 2022 19:05:37 +0000 (0:00:00.049) 0:00:54.924 ********* included: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml for /cache/rhel-8-y.qcow2.snap => (item=mount) included: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-fstab.yml for /cache/rhel-8-y.qcow2.snap => (item=fstab) included: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-fs.yml for /cache/rhel-8-y.qcow2.snap => (item=fs) included: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-device.yml for /cache/rhel-8-y.qcow2.snap => (item=device) included: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml for /cache/rhel-8-y.qcow2.snap => (item=encryption) included: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-md.yml for /cache/rhel-8-y.qcow2.snap => (item=md) included: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml for /cache/rhel-8-y.qcow2.snap => (item=size) included: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-cache.yml for /cache/rhel-8-y.qcow2.snap => (item=cache) TASK [Get expected mount device based on device type] ************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml:6 Thursday 21 July 2022 19:05:37 +0000 (0:00:00.082) 0:00:55.007 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/foo-test2" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml:10 Thursday 21 July 2022 19:05:37 +0000 (0:00:00.042) 0:00:55.049 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_mount_device_matches": [ { "block_available": 954137, "block_size": 4096, "block_total": 1010616, "block_used": 56479, "device": "/dev/mapper/foo-test2", "fstype": "ext4", "inode_available": 262133, "inode_total": 262144, "inode_used": 11, "mount": "/opt/test2", "options": "rw,seclabel,relatime", "size_available": 3908145152, "size_total": 4139483136, "uuid": "64454f89-f2cd-4a8b-958f-4c900f72cf79" } ], "storage_test_mount_expected_match_count": "1", "storage_test_mount_point_matches": [ { "block_available": 954137, "block_size": 4096, "block_total": 1010616, "block_used": 56479, "device": "/dev/mapper/foo-test2", "fstype": "ext4", "inode_available": 262133, "inode_total": 262144, "inode_used": 11, "mount": "/opt/test2", "options": "rw,seclabel,relatime", "size_available": 3908145152, "size_total": 4139483136, "uuid": "64454f89-f2cd-4a8b-958f-4c900f72cf79" } ], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Verify the current mount state by device] ******************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml:20 Thursday 21 July 2022 19:05:37 +0000 (0:00:00.055) 0:00:55.105 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify the current mount state by mount point] *************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml:29 Thursday 21 July 2022 19:05:37 +0000 (0:00:00.047) 0:00:55.153 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify the mount fs type] ************************************************ task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml:37 Thursday 21 July 2022 19:05:37 +0000 (0:00:00.046) 0:00:55.199 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [command] ***************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml:46 Thursday 21 July 2022 19:05:37 +0000 (0:00:00.045) 0:00:55.245 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml:50 Thursday 21 July 2022 19:05:37 +0000 (0:00:00.022) 0:00:55.267 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml:55 Thursday 21 July 2022 19:05:37 +0000 (0:00:00.022) 0:00:55.290 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml:65 Thursday 21 July 2022 19:05:37 +0000 (0:00:00.021) 0:00:55.312 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-fstab.yml:2 Thursday 21 July 2022 19:05:37 +0000 (0:00:00.032) 0:00:55.345 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "1", "storage_test_fstab_expected_mount_options_matches": "1", "storage_test_fstab_expected_mount_point_matches": "1", "storage_test_fstab_id_matches": [ "/dev/mapper/foo-test2 " ], "storage_test_fstab_mount_options_matches": [ " /opt/test2 ext4 defaults " ], "storage_test_fstab_mount_point_matches": [ " /opt/test2 " ] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-fstab.yml:12 Thursday 21 July 2022 19:05:37 +0000 (0:00:00.058) 0:00:55.403 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify the fstab mount point] ******************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-fstab.yml:19 Thursday 21 July 2022 19:05:37 +0000 (0:00:00.049) 0:00:55.452 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-fstab.yml:25 Thursday 21 July 2022 19:05:37 +0000 (0:00:00.046) 0:00:55.499 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-fstab.yml:34 Thursday 21 July 2022 19:05:37 +0000 (0:00:00.034) 0:00:55.533 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-fs.yml:4 Thursday 21 July 2022 19:05:38 +0000 (0:00:00.034) 0:00:55.568 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify fs label] ********************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-fs.yml:10 Thursday 21 July 2022 19:05:38 +0000 (0:00:00.035) 0:00:55.603 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [See whether the device node is present] ********************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-device.yml:4 Thursday 21 July 2022 19:05:38 +0000 (0:00:00.040) 0:00:55.644 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "stat": { "atime": 1658430322.3951166, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1658430322.3951166, "dev": 6, "device_type": 64768, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 59192, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/symlink", "mode": "0660", "mtime": 1658430322.3951166, "nlink": 1, "path": "/dev/mapper/foo-test2", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-device.yml:10 Thursday 21 July 2022 19:05:38 +0000 (0:00:00.373) 0:00:56.017 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about this volume] ********************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-device.yml:15 Thursday 21 July 2022 19:05:38 +0000 (0:00:00.041) 0:00:56.059 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [(1/2) Process volume type (set initial value)] *************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-device.yml:21 Thursday 21 July 2022 19:05:38 +0000 (0:00:00.040) 0:00:56.099 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [(2/2) Process volume type (get RAID value)] ****************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-device.yml:25 Thursday 21 July 2022 19:05:38 +0000 (0:00:00.038) 0:00:56.138 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-device.yml:30 Thursday 21 July 2022 19:05:38 +0000 (0:00:00.024) 0:00:56.162 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Stat the LUKS device, if encrypted] ************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:3 Thursday 21 July 2022 19:05:38 +0000 (0:00:00.044) 0:00:56.206 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:10 Thursday 21 July 2022 19:05:38 +0000 (0:00:00.024) 0:00:56.231 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [Collect LUKS info for this volume] *************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:15 Thursday 21 July 2022 19:05:40 +0000 (0:00:01.421) 0:00:57.652 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:21 Thursday 21 July 2022 19:05:40 +0000 (0:00:00.026) 0:00:57.679 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:27 Thursday 21 July 2022 19:05:40 +0000 (0:00:00.026) 0:00:57.705 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:33 Thursday 21 July 2022 19:05:40 +0000 (0:00:00.055) 0:00:57.761 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:39 Thursday 21 July 2022 19:05:40 +0000 (0:00:00.027) 0:00:57.788 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:44 Thursday 21 July 2022 19:05:40 +0000 (0:00:00.025) 0:00:57.814 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:50 Thursday 21 July 2022 19:05:40 +0000 (0:00:00.025) 0:00:57.840 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:56 Thursday 21 July 2022 19:05:40 +0000 (0:00:00.025) 0:00:57.865 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:62 Thursday 21 July 2022 19:05:40 +0000 (0:00:00.024) 0:00:57.890 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:67 Thursday 21 July 2022 19:05:40 +0000 (0:00:00.049) 0:00:57.940 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:72 Thursday 21 July 2022 19:05:40 +0000 (0:00:00.049) 0:00:57.990 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:78 Thursday 21 July 2022 19:05:40 +0000 (0:00:00.038) 0:00:58.028 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:84 Thursday 21 July 2022 19:05:40 +0000 (0:00:00.038) 0:00:58.067 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:90 Thursday 21 July 2022 19:05:40 +0000 (0:00:00.035) 0:00:58.103 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [get information about RAID] ********************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-md.yml:7 Thursday 21 July 2022 19:05:40 +0000 (0:00:00.034) 0:00:58.137 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-md.yml:13 Thursday 21 July 2022 19:05:40 +0000 (0:00:00.038) 0:00:58.176 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-md.yml:17 Thursday 21 July 2022 19:05:40 +0000 (0:00:00.039) 0:00:58.216 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-md.yml:21 Thursday 21 July 2022 19:05:40 +0000 (0:00:00.036) 0:00:58.252 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-md.yml:25 Thursday 21 July 2022 19:05:40 +0000 (0:00:00.035) 0:00:58.287 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-md.yml:31 Thursday 21 July 2022 19:05:40 +0000 (0:00:00.040) 0:00:58.327 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-md.yml:37 Thursday 21 July 2022 19:05:40 +0000 (0:00:00.036) 0:00:58.364 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the actual size of the volume] ************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:3 Thursday 21 July 2022 19:05:40 +0000 (0:00:00.035) 0:00:58.400 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "bytes": 4294967296, "changed": false, "lvm": "4g", "parted": "4GiB", "size": "4 GiB" } TASK [parse the requested size of the volume] ********************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:9 Thursday 21 July 2022 19:05:41 +0000 (0:00:00.373) 0:00:58.774 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Establish base value for expected size] ********************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:15 Thursday 21 July 2022 19:05:41 +0000 (0:00:00.040) 0:00:58.814 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:20 Thursday 21 July 2022 19:05:41 +0000 (0:00:00.037) 0:00:58.852 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_expected_size": "6442450944.0" } TASK [debug] ******************************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:25 Thursday 21 July 2022 19:05:41 +0000 (0:00:00.072) 0:00:58.924 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_pool": { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "60%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test2", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test2", "_raw_device": "/dev/mapper/foo-test2", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "40%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } } TASK [debug] ******************************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:28 Thursday 21 July 2022 19:05:41 +0000 (0:00:00.118) 0:00:59.043 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_blkinfo": { "changed": false, "failed": false, "info": { "/dev/mapper/foo-test1": { "fstype": "xfs", "label": "", "name": "/dev/mapper/foo-test1", "size": "6G", "type": "lvm", "uuid": "d739e41d-c1a1-4003-b357-25bf62753af5" }, "/dev/mapper/foo-test2": { "fstype": "ext4", "label": "", "name": "/dev/mapper/foo-test2", "size": "4G", "type": "lvm", "uuid": "64454f89-f2cd-4a8b-958f-4c900f72cf79" }, "/dev/nvme0n1": { "fstype": "", "label": "", "name": "/dev/nvme0n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme1n1": { "fstype": "", "label": "", "name": "/dev/nvme1n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme2n1": { "fstype": "", "label": "", "name": "/dev/nvme2n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sda": { "fstype": "LVM2_member", "label": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "HA17fj-P3dB-xyZl-9Zru-ZzXr-ULh6-93Sgui" }, "/dev/sdb": { "fstype": "", "label": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc": { "fstype": "", "label": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sr0": { "fstype": "iso9660", "label": "cidata", "name": "/dev/sr0", "size": "364K", "type": "rom", "uuid": "2022-07-21-19-04-29-00" }, "/dev/vda": { "fstype": "", "label": "", "name": "/dev/vda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vda1": { "fstype": "", "label": "", "name": "/dev/vda1", "size": "1M", "type": "partition", "uuid": "" }, "/dev/vda2": { "fstype": "vfat", "label": "", "name": "/dev/vda2", "size": "100M", "type": "partition", "uuid": "7B77-95E7" }, "/dev/vda3": { "fstype": "xfs", "label": "root", "name": "/dev/vda3", "size": "9.9G", "type": "partition", "uuid": "58dca2db-4ccd-4171-a017-dd53ad8c1c54" }, "/dev/vdb": { "fstype": "", "label": "", "name": "/dev/vdb", "size": "2G", "type": "disk", "uuid": "" }, "/dev/vdb1": { "fstype": "ext4", "label": "yumcache", "name": "/dev/vdb1", "size": "2G", "type": "partition", "uuid": "4d1ed9a1-5608-4ff8-ad89-21d95caa66fa" }, "/dev/vdc": { "fstype": "", "label": "", "name": "/dev/vdc", "size": "2G", "type": "disk", "uuid": "" }, "/dev/vdc1": { "fstype": "ext4", "label": "yumvarlib", "name": "/dev/vdc1", "size": "2G", "type": "partition", "uuid": "a7abcc97-edc5-476c-8ccd-694b4621fd28" }, "/dev/vdd": { "fstype": "", "label": "", "name": "/dev/vdd", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vde": { "fstype": "", "label": "", "name": "/dev/vde", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vdf": { "fstype": "", "label": "", "name": "/dev/vdf", "size": "10G", "type": "disk", "uuid": "" } } } } TASK [Get the size of parent/pool device] ************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:31 Thursday 21 July 2022 19:05:41 +0000 (0:00:00.054) 0:00:59.097 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "bytes": 10737418240, "changed": false, "lvm": "10g", "parted": "10GiB", "size": "10 GiB" } TASK [debug] ******************************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:36 Thursday 21 July 2022 19:05:41 +0000 (0:00:00.397) 0:00:59.495 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_pool_size": { "bytes": 10737418240, "changed": false, "failed": false, "lvm": "10g", "parted": "10GiB", "size": "10 GiB" } } TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:39 Thursday 21 July 2022 19:05:41 +0000 (0:00:00.052) 0:00:59.547 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_expected_size": "4294967296.0" }, "changed": false } TASK [debug] ******************************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:44 Thursday 21 July 2022 19:05:42 +0000 (0:00:00.049) 0:00:59.596 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_actual_size": { "bytes": 4294967296, "changed": false, "failed": false, "lvm": "4g", "parted": "4GiB", "size": "4 GiB" } } TASK [debug] ******************************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:47 Thursday 21 July 2022 19:05:42 +0000 (0:00:00.035) 0:00:59.632 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_expected_size": "4294967296.0" } TASK [assert] ****************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:50 Thursday 21 July 2022 19:05:42 +0000 (0:00:00.035) 0:00:59.667 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Get information about the LV] ******************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-cache.yml:6 Thursday 21 July 2022 19:05:42 +0000 (0:00:00.049) 0:00:59.716 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "cmd": [ "lvs", "--noheadings", "--nameprefixes", "--units=b", "--nosuffix", "--unquoted", "-o", "name,attr,cache_total_blocks,chunk_size,segtype", "foo/test2" ], "delta": "0:00:00.035752", "end": "2022-07-21 15:05:42.653978", "rc": 0, "start": "2022-07-21 15:05:42.618226" } STDOUT: LVM2_LV_NAME=test2 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-cache.yml:14 Thursday 21 July 2022 19:05:42 +0000 (0:00:00.420) 0:01:00.137 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_lv_segtype": [ "linear" ] }, "changed": false } TASK [check segment type] ****************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-cache.yml:17 Thursday 21 July 2022 19:05:42 +0000 (0:00:00.049) 0:01:00.186 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-cache.yml:22 Thursday 21 July 2022 19:05:42 +0000 (0:00:00.056) 0:01:00.243 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the requested cache size] ****************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-cache.yml:26 Thursday 21 July 2022 19:05:42 +0000 (0:00:00.039) 0:01:00.283 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-cache.yml:32 Thursday 21 July 2022 19:05:42 +0000 (0:00:00.041) 0:01:00.324 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-cache.yml:36 Thursday 21 July 2022 19:05:42 +0000 (0:00:00.040) 0:01:00.365 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume.yml:16 Thursday 21 July 2022 19:05:42 +0000 (0:00:00.039) 0:01:00.405 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [Verify the volumes with no pool were correctly managed] ****************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-role-results.yml:43 Thursday 21 July 2022 19:05:42 +0000 (0:00:00.036) 0:01:00.441 ********* TASK [Clean up variable namespace] ********************************************* task path: /tmp/tmpmnqeja9w/tests/storage/verify-role-results.yml:53 Thursday 21 July 2022 19:05:42 +0000 (0:00:00.026) 0:01:00.467 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_blkinfo": null, "storage_test_crypttab": null, "storage_test_fstab": null }, "changed": false } TASK [Repeat the previous invocation to verify idempotence] ******************** task path: /tmp/tmpmnqeja9w/tests/storage/tests_lvm_percent_size.yml:81 Thursday 21 July 2022 19:05:42 +0000 (0:00:00.035) 0:01:00.502 ********* TASK [fedora.linux_system_roles.storage : set platform/version specific variables] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2 Thursday 21 July 2022 19:05:42 +0000 (0:00:00.047) 0:01:00.550 ********* included: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for /cache/rhel-8-y.qcow2.snap TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2 Thursday 21 July 2022 19:05:43 +0000 (0:00:00.049) 0:01:00.600 ********* ok: [/cache/rhel-8-y.qcow2.snap] TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7 Thursday 21 July 2022 19:05:43 +0000 (0:00:00.530) 0:01:01.130 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [/cache/rhel-8-y.qcow2.snap] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [/cache/rhel-8-y.qcow2.snap] => (item=RedHat_8.yml) => { "ansible_facts": { "blivet_package_list": [ "python3-blivet", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap", "vdo", "kmod-kvdo", "xfsprogs" ] }, "ansible_included_var_files": [ "/tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/vars/RedHat_8.yml" ], "ansible_loop_var": "item", "changed": false, "item": "RedHat_8.yml" } skipping: [/cache/rhel-8-y.qcow2.snap] => (item=RedHat_8.7.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat_8.7.yml", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : define an empty list of pools to be used in testing] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5 Thursday 21 July 2022 19:05:43 +0000 (0:00:00.116) 0:01:01.247 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : define an empty list of volumes to be used in testing] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9 Thursday 21 July 2022 19:05:43 +0000 (0:00:00.082) 0:01:01.330 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : include the appropriate provider tasks] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13 Thursday 21 July 2022 19:05:43 +0000 (0:00:00.087) 0:01:01.418 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount included: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for /cache/rhel-8-y.qcow2.snap TASK [fedora.linux_system_roles.storage : get a list of rpm packages installed on host machine] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2 Thursday 21 July 2022 19:05:43 +0000 (0:00:00.101) 0:01:01.519 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : make sure blivet is available] ******* task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:7 Thursday 21 July 2022 19:05:43 +0000 (0:00:00.022) 0:01:01.542 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [fedora.linux_system_roles.storage : show storage_pools] ****************** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:13 Thursday 21 July 2022 19:05:45 +0000 (0:00:01.309) 0:01:02.851 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_pools": [ { "disks": [ "sda" ], "name": "foo", "volumes": [ { "mount_point": "/opt/test1", "name": "test1", "size": "60%" }, { "mount_point": "/opt/test2", "name": "test2", "size": "40%" } ] } ] } TASK [fedora.linux_system_roles.storage : show storage_volumes] **************** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:18 Thursday 21 July 2022 19:05:45 +0000 (0:00:00.044) 0:01:02.895 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [fedora.linux_system_roles.storage : get required packages] *************** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:23 Thursday 21 July 2022 19:05:45 +0000 (0:00:00.041) 0:01:02.936 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [ "lvm2" ], "pools": [], "volumes": [] } TASK [fedora.linux_system_roles.storage : enable copr repositories if needed] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:35 Thursday 21 July 2022 19:05:47 +0000 (0:00:01.628) 0:01:04.564 ********* included: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml for /cache/rhel-8-y.qcow2.snap TASK [fedora.linux_system_roles.storage : check if the COPR support packages should be installed] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:2 Thursday 21 July 2022 19:05:47 +0000 (0:00:00.046) 0:01:04.611 ********* TASK [fedora.linux_system_roles.storage : make sure COPR support packages are present] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:13 Thursday 21 July 2022 19:05:47 +0000 (0:00:00.043) 0:01:04.654 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : enable COPRs] ************************ task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:18 Thursday 21 July 2022 19:05:47 +0000 (0:00:00.053) 0:01:04.707 ********* TASK [fedora.linux_system_roles.storage : make sure required packages are installed] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:41 Thursday 21 July 2022 19:05:47 +0000 (0:00:00.035) 0:01:04.743 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [fedora.linux_system_roles.storage : get service facts] ******************* task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:47 Thursday 21 July 2022 19:05:48 +0000 (0:00:01.290) 0:01:06.033 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "services": { "NetworkManager-dispatcher.service": { "name": "NetworkManager-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "NetworkManager-wait-online.service": { "name": "NetworkManager-wait-online.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "NetworkManager.service": { "name": "NetworkManager.service", "source": "systemd", "state": "running", "status": "enabled" }, "arp-ethers.service": { "name": "arp-ethers.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "auditd.service": { "name": "auditd.service", "source": "systemd", "state": "running", "status": "enabled" }, "auth-rpcgss-module.service": { "name": "auth-rpcgss-module.service", "source": "systemd", "state": "stopped", "status": "static" }, "autovt@.service": { "name": "autovt@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "blivet.service": { "name": "blivet.service", "source": "systemd", "state": "inactive", "status": "static" }, "blk-availability.service": { "name": "blk-availability.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "chrony-dnssrv@.service": { "name": "chrony-dnssrv@.service", "source": "systemd", "state": "unknown", "status": "static" }, "chrony-wait.service": { "name": "chrony-wait.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd.service": { "name": "chronyd.service", "source": "systemd", "state": "running", "status": "enabled" }, "cloud-config.service": { "name": "cloud-config.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-final.service": { "name": "cloud-final.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init-hotplugd.service": { "name": "cloud-init-hotplugd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cloud-init-local.service": { "name": "cloud-init-local.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init.service": { "name": "cloud-init.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cockpit-motd.service": { "name": "cockpit-motd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cockpit-wsinstance-http.service": { "name": "cockpit-wsinstance-http.service", "source": "systemd", "state": "inactive", "status": "static" }, "cockpit-wsinstance-https-factory@.service": { "name": "cockpit-wsinstance-https-factory@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cockpit-wsinstance-https@.service": { "name": "cockpit-wsinstance-https@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cockpit.service": { "name": "cockpit.service", "source": "systemd", "state": "inactive", "status": "static" }, "console-getty.service": { "name": "console-getty.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "container-getty@.service": { "name": "container-getty@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cpupower.service": { "name": "cpupower.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "crond.service": { "name": "crond.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-org.freedesktop.hostname1.service": { "name": "dbus-org.freedesktop.hostname1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.locale1.service": { "name": "dbus-org.freedesktop.locale1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.login1.service": { "name": "dbus-org.freedesktop.login1.service", "source": "systemd", "state": "active", "status": "static" }, "dbus-org.freedesktop.nm-dispatcher.service": { "name": "dbus-org.freedesktop.nm-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "dbus-org.freedesktop.portable1.service": { "name": "dbus-org.freedesktop.portable1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.timedate1.service": { "name": "dbus-org.freedesktop.timedate1.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "dbus.service": { "name": "dbus.service", "source": "systemd", "state": "running", "status": "static" }, "dbxtool.service": { "name": "dbxtool.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "debug-shell.service": { "name": "debug-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "display-manager.service": { "name": "display-manager.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "dm-event.service": { "name": "dm-event.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-makecache.service": { "name": "dnf-makecache.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-cmdline.service": { "name": "dracut-cmdline.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-initqueue.service": { "name": "dracut-initqueue.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-mount.service": { "name": "dracut-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-mount.service": { "name": "dracut-pre-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-pivot.service": { "name": "dracut-pre-pivot.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-trigger.service": { "name": "dracut-pre-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-udev.service": { "name": "dracut-pre-udev.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown.service": { "name": "dracut-shutdown.service", "source": "systemd", "state": "stopped", "status": "static" }, "emergency.service": { "name": "emergency.service", "source": "systemd", "state": "stopped", "status": "static" }, "fcoe.service": { "name": "fcoe.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "fstrim.service": { "name": "fstrim.service", "source": "systemd", "state": "inactive", "status": "static" }, "getty@.service": { "name": "getty@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "getty@tty1.service": { "name": "getty@tty1.service", "source": "systemd", "state": "running", "status": "active" }, "grub-boot-indeterminate.service": { "name": "grub-boot-indeterminate.service", "source": "systemd", "state": "inactive", "status": "static" }, "gssproxy.service": { "name": "gssproxy.service", "source": "systemd", "state": "running", "status": "disabled" }, "halt-local.service": { "name": "halt-local.service", "source": "systemd", "state": "inactive", "status": "static" }, "hv_kvp_daemon.service": { "name": "hv_kvp_daemon.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "import-state.service": { "name": "import-state.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "initrd-cleanup.service": { "name": "initrd-cleanup.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-parse-etc.service": { "name": "initrd-parse-etc.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-switch-root.service": { "name": "initrd-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-udevadm-cleanup-db.service": { "name": "initrd-udevadm-cleanup-db.service", "source": "systemd", "state": "stopped", "status": "static" }, "insights-client-boot.service": { "name": "insights-client-boot.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "insights-client-results.service": { "name": "insights-client-results.service", "source": "systemd", "state": "inactive", "status": "static" }, "insights-client.service": { "name": "insights-client.service", "source": "systemd", "state": "inactive", "status": "static" }, "irqbalance.service": { "name": "irqbalance.service", "source": "systemd", "state": "running", "status": "enabled" }, "iscsi-shutdown.service": { "name": "iscsi-shutdown.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "iscsi.service": { "name": "iscsi.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "iscsid.service": { "name": "iscsid.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "kdump.service": { "name": "kdump.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "kmod-static-nodes.service": { "name": "kmod-static-nodes.service", "source": "systemd", "state": "stopped", "status": "static" }, "kvm_stat.service": { "name": "kvm_stat.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "ldconfig.service": { "name": "ldconfig.service", "source": "systemd", "state": "stopped", "status": "static" }, "loadmodules.service": { "name": "loadmodules.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "lvm2-activation-early.service": { "name": "lvm2-activation-early.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "lvm2-activation.service": { "name": "lvm2-activation.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "lvm2-lvmpolld.service": { "name": "lvm2-lvmpolld.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-monitor.service": { "name": "lvm2-monitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "lvm2-pvscan@.service": { "name": "lvm2-pvscan@.service", "source": "systemd", "state": "unknown", "status": "static" }, "lvm2-pvscan@8:0.service": { "name": "lvm2-pvscan@8:0.service", "source": "systemd", "state": "stopped", "status": "active" }, "man-db-cache-update.service": { "name": "man-db-cache-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "man-db-restart-cache-update.service": { "name": "man-db-restart-cache-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "mdadm-grow-continue@.service": { "name": "mdadm-grow-continue@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdadm-last-resort@.service": { "name": "mdadm-last-resort@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdcheck_continue.service": { "name": "mdcheck_continue.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdcheck_start.service": { "name": "mdcheck_start.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmon@.service": { "name": "mdmon@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdmonitor-oneshot.service": { "name": "mdmonitor-oneshot.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmonitor.service": { "name": "mdmonitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "messagebus.service": { "name": "messagebus.service", "source": "systemd", "state": "active", "status": "static" }, "microcode.service": { "name": "microcode.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "multipathd.service": { "name": "multipathd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "ndctl-monitor.service": { "name": "ndctl-monitor.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "network.service": { "name": "network.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "nfs-blkmap.service": { "name": "nfs-blkmap.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-convert.service": { "name": "nfs-convert.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-idmapd.service": { "name": "nfs-idmapd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-mountd.service": { "name": "nfs-mountd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-server.service": { "name": "nfs-server.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-utils.service": { "name": "nfs-utils.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfsdcld.service": { "name": "nfsdcld.service", "source": "systemd", "state": "stopped", "status": "static" }, "nis-domainname.service": { "name": "nis-domainname.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "ntpd.service": { "name": "ntpd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ntpdate.service": { "name": "ntpdate.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "oddjobd.service": { "name": "oddjobd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "packagekit-offline-update.service": { "name": "packagekit-offline-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "packagekit.service": { "name": "packagekit.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-quit-wait.service": { "name": "plymouth-quit-wait.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "plymouth-start.service": { "name": "plymouth-start.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "polkit.service": { "name": "polkit.service", "source": "systemd", "state": "running", "status": "static" }, "qemu-guest-agent.service": { "name": "qemu-guest-agent.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "quotaon.service": { "name": "quotaon.service", "source": "systemd", "state": "inactive", "status": "static" }, "rbdmap.service": { "name": "rbdmap.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "rc-local.service": { "name": "rc-local.service", "source": "systemd", "state": "stopped", "status": "static" }, "rdisc.service": { "name": "rdisc.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rescue.service": { "name": "rescue.service", "source": "systemd", "state": "stopped", "status": "static" }, "rhcd.service": { "name": "rhcd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhsm-facts.service": { "name": "rhsm-facts.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhsm.service": { "name": "rhsm.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhsmcertd.service": { "name": "rhsmcertd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpc-gssd.service": { "name": "rpc-gssd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd-notify.service": { "name": "rpc-statd-notify.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd.service": { "name": "rpc-statd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpcbind.service": { "name": "rpcbind.service", "source": "systemd", "state": "running", "status": "enabled" }, "rsyslog.service": { "name": "rsyslog.service", "source": "systemd", "state": "running", "status": "enabled" }, "selinux-autorelabel-mark.service": { "name": "selinux-autorelabel-mark.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "selinux-autorelabel.service": { "name": "selinux-autorelabel.service", "source": "systemd", "state": "inactive", "status": "static" }, "serial-getty@.service": { "name": "serial-getty@.service", "source": "systemd", "state": "unknown", "status": "indirect" }, "serial-getty@ttyS0.service": { "name": "serial-getty@ttyS0.service", "source": "systemd", "state": "running", "status": "active" }, "snapd.seeded.service": { "name": "snapd.seeded.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sntp.service": { "name": "sntp.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sshd-keygen.service": { "name": "sshd-keygen.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sshd-keygen@.service": { "name": "sshd-keygen@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@ecdsa.service": { "name": "sshd-keygen@ecdsa.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd-keygen@ed25519.service": { "name": "sshd-keygen@ed25519.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd-keygen@rsa.service": { "name": "sshd-keygen@rsa.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd.service": { "name": "sshd.service", "source": "systemd", "state": "running", "status": "enabled" }, "sshd@.service": { "name": "sshd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "sssd-autofs.service": { "name": "sssd-autofs.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-kcm.service": { "name": "sssd-kcm.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "sssd-nss.service": { "name": "sssd-nss.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pac.service": { "name": "sssd-pac.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pam.service": { "name": "sssd-pam.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-ssh.service": { "name": "sssd-ssh.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-sudo.service": { "name": "sssd-sudo.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd.service": { "name": "sssd.service", "source": "systemd", "state": "running", "status": "enabled" }, "syslog.service": { "name": "syslog.service", "source": "systemd", "state": "active", "status": "enabled" }, "system-update-cleanup.service": { "name": "system-update-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-ask-password-console.service": { "name": "systemd-ask-password-console.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-wall.service": { "name": "systemd-ask-password-wall.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-backlight@.service": { "name": "systemd-backlight@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-binfmt.service": { "name": "systemd-binfmt.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-coredump@.service": { "name": "systemd-coredump@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-exit.service": { "name": "systemd-exit.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-firstboot.service": { "name": "systemd-firstboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck-root.service": { "name": "systemd-fsck-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck@.service": { "name": "systemd-fsck@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-fsck@dev-disk-by\\x2duuid-7B77\\x2d95E7.service": { "name": "systemd-fsck@dev-disk-by\\x2duuid-7B77\\x2d95E7.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-fsck@dev-vdb1.service": { "name": "systemd-fsck@dev-vdb1.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-fsck@dev-vdc1.service": { "name": "systemd-fsck@dev-vdc1.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-halt.service": { "name": "systemd-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hibernate-resume@.service": { "name": "systemd-hibernate-resume@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-hibernate.service": { "name": "systemd-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hostnamed.service": { "name": "systemd-hostnamed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hwdb-update.service": { "name": "systemd-hwdb-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-hybrid-sleep.service": { "name": "systemd-hybrid-sleep.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-initctl.service": { "name": "systemd-initctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-catalog-update.service": { "name": "systemd-journal-catalog-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-flush.service": { "name": "systemd-journal-flush.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journald.service": { "name": "systemd-journald.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-kexec.service": { "name": "systemd-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-localed.service": { "name": "systemd-localed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-logind.service": { "name": "systemd-logind.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-machine-id-commit.service": { "name": "systemd-machine-id-commit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-modules-load.service": { "name": "systemd-modules-load.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-networkd-wait-online.service": { "name": "systemd-networkd-wait-online.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "systemd-portabled.service": { "name": "systemd-portabled.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-poweroff.service": { "name": "systemd-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-quotacheck.service": { "name": "systemd-quotacheck.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-random-seed.service": { "name": "systemd-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-reboot.service": { "name": "systemd-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-remount-fs.service": { "name": "systemd-remount-fs.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-resolved.service": { "name": "systemd-resolved.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-rfkill.service": { "name": "systemd-rfkill.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend-then-hibernate.service": { "name": "systemd-suspend-then-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend.service": { "name": "systemd-suspend.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-sysctl.service": { "name": "systemd-sysctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-sysusers.service": { "name": "systemd-sysusers.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-timedated.service": { "name": "systemd-timedated.service", "source": "systemd", "state": "inactive", "status": "masked" }, "systemd-timesyncd.service": { "name": "systemd-timesyncd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "systemd-tmpfiles-clean.service": { "name": "systemd-tmpfiles-clean.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev.service": { "name": "systemd-tmpfiles-setup-dev.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup.service": { "name": "systemd-tmpfiles-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-settle.service": { "name": "systemd-udev-settle.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-trigger.service": { "name": "systemd-udev-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udevd.service": { "name": "systemd-udevd.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-update-done.service": { "name": "systemd-update-done.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp-runlevel.service": { "name": "systemd-update-utmp-runlevel.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp.service": { "name": "systemd-update-utmp.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-user-sessions.service": { "name": "systemd-user-sessions.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-vconsole-setup.service": { "name": "systemd-vconsole-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-volatile-root.service": { "name": "systemd-volatile-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "tcsd.service": { "name": "tcsd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "teamd@.service": { "name": "teamd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "timedatex.service": { "name": "timedatex.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "tuned.service": { "name": "tuned.service", "source": "systemd", "state": "running", "status": "enabled" }, "unbound-anchor.service": { "name": "unbound-anchor.service", "source": "systemd", "state": "stopped", "status": "static" }, "user-runtime-dir@.service": { "name": "user-runtime-dir@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user-runtime-dir@0.service": { "name": "user-runtime-dir@0.service", "source": "systemd", "state": "stopped", "status": "active" }, "user@.service": { "name": "user@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user@0.service": { "name": "user@0.service", "source": "systemd", "state": "running", "status": "active" }, "vdo-start-by-dev@.service": { "name": "vdo-start-by-dev@.service", "source": "systemd", "state": "unknown", "status": "static" }, "vdo.service": { "name": "vdo.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "ypbind.service": { "name": "ypbind.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "yppasswdd.service": { "name": "yppasswdd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ypserv.service": { "name": "ypserv.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ypxfrd.service": { "name": "ypxfrd.service", "source": "systemd", "state": "stopped", "status": "not-found" } } }, "changed": false } TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] ***** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:53 Thursday 21 July 2022 19:05:51 +0000 (0:00:02.758) 0:01:08.792 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:58 Thursday 21 July 2022 19:05:51 +0000 (0:00:00.101) 0:01:08.893 ********* TASK [fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64 Thursday 21 July 2022 19:05:51 +0000 (0:00:00.025) 0:01:08.919 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "actions": [], "changed": false, "crypts": [], "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/vda2", "/dev/vda3", "/dev/mapper/foo-test1", "/dev/mapper/foo-test2", "/dev/sdb", "/dev/sdc", "/dev/nvme0n1", "/dev/nvme1n1", "/dev/nvme2n1", "/dev/vdb1", "/dev/vdc1", "/dev/vdd", "/dev/vde", "/dev/vdf" ], "mounts": [ { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" }, { "dump": 0, "fstype": "ext4", "opts": "defaults", "passno": 0, "path": "/opt/test2", "src": "/dev/mapper/foo-test2", "state": "mounted" } ], "packages": [ "e2fsprogs", "xfsprogs", "lvm2", "dosfstools" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "60%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test2", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test2", "_raw_device": "/dev/mapper/foo-test2", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "40%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:78 Thursday 21 July 2022 19:05:53 +0000 (0:00:01.807) 0:01:10.727 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90 Thursday 21 July 2022 19:05:53 +0000 (0:00:00.038) 0:01:10.765 ********* TASK [fedora.linux_system_roles.storage : show blivet_output] ****************** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:96 Thursday 21 July 2022 19:05:53 +0000 (0:00:00.023) 0:01:10.788 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "blivet_output": { "actions": [], "changed": false, "crypts": [], "failed": false, "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/vda2", "/dev/vda3", "/dev/mapper/foo-test1", "/dev/mapper/foo-test2", "/dev/sdb", "/dev/sdc", "/dev/nvme0n1", "/dev/nvme1n1", "/dev/nvme2n1", "/dev/vdb1", "/dev/vdc1", "/dev/vdd", "/dev/vde", "/dev/vdf" ], "mounts": [ { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" }, { "dump": 0, "fstype": "ext4", "opts": "defaults", "passno": 0, "path": "/opt/test2", "src": "/dev/mapper/foo-test2", "state": "mounted" } ], "packages": [ "e2fsprogs", "xfsprogs", "lvm2", "dosfstools" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "60%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test2", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test2", "_raw_device": "/dev/mapper/foo-test2", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "40%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } } TASK [fedora.linux_system_roles.storage : set the list of pools for test verification] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:101 Thursday 21 July 2022 19:05:53 +0000 (0:00:00.110) 0:01:10.899 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "60%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test2", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test2", "_raw_device": "/dev/mapper/foo-test2", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "40%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] }, "changed": false } TASK [fedora.linux_system_roles.storage : set the list of volumes for test verification] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:105 Thursday 21 July 2022 19:05:53 +0000 (0:00:00.041) 0:01:10.941 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : remove obsolete mounts] ************** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:121 Thursday 21 July 2022 19:05:53 +0000 (0:00:00.037) 0:01:10.978 ********* TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132 Thursday 21 July 2022 19:05:53 +0000 (0:00:00.038) 0:01:11.016 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.storage : set up new/current mounts] *********** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:137 Thursday 21 July 2022 19:05:54 +0000 (0:00:00.631) 0:01:11.648 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount ok: [/cache/rhel-8-y.qcow2.snap] => (item={'src': '/dev/mapper/foo-test1', 'path': '/opt/test1', 'fstype': 'xfs', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted'}) => { "ansible_loop_var": "mount_info", "backup_file": "", "boot": "yes", "changed": false, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" }, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test1" } redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount ok: [/cache/rhel-8-y.qcow2.snap] => (item={'src': '/dev/mapper/foo-test2', 'path': '/opt/test2', 'fstype': 'ext4', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted'}) => { "ansible_loop_var": "mount_info", "backup_file": "", "boot": "yes", "changed": false, "dump": "0", "fstab": "/etc/fstab", "fstype": "ext4", "mount_info": { "dump": 0, "fstype": "ext4", "opts": "defaults", "passno": 0, "path": "/opt/test2", "src": "/dev/mapper/foo-test2", "state": "mounted" }, "name": "/opt/test2", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test2" } TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148 Thursday 21 July 2022 19:05:54 +0000 (0:00:00.801) 0:01:12.449 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.storage : retrieve facts for the /etc/crypttab file] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:156 Thursday 21 July 2022 19:05:55 +0000 (0:00:00.638) 0:01:13.088 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "stat": { "exists": false } } TASK [fedora.linux_system_roles.storage : manage /etc/crypttab to account for changes we just made] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:161 Thursday 21 July 2022 19:05:55 +0000 (0:00:00.376) 0:01:13.465 ********* TASK [fedora.linux_system_roles.storage : Update facts] ************************ task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:183 Thursday 21 July 2022 19:05:55 +0000 (0:00:00.025) 0:01:13.490 ********* ok: [/cache/rhel-8-y.qcow2.snap] META: role_complete for /cache/rhel-8-y.qcow2.snap TASK [include_tasks] *********************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/tests_lvm_percent_size.yml:95 Thursday 21 July 2022 19:05:56 +0000 (0:00:00.991) 0:01:14.482 ********* included: /tmp/tmpmnqeja9w/tests/storage/verify-role-results.yml for /cache/rhel-8-y.qcow2.snap TASK [Print out pool information] ********************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-role-results.yml:1 Thursday 21 July 2022 19:05:56 +0000 (0:00:00.043) 0:01:14.526 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "60%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test2", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test2", "_raw_device": "/dev/mapper/foo-test2", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "40%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] } TASK [Print out volume information] ******************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-role-results.yml:6 Thursday 21 July 2022 19:05:57 +0000 (0:00:00.051) 0:01:14.577 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => {} TASK [Collect info about the volumes.] ***************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-role-results.yml:14 Thursday 21 July 2022 19:05:57 +0000 (0:00:00.035) 0:01:14.613 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "info": { "/dev/mapper/foo-test1": { "fstype": "xfs", "label": "", "name": "/dev/mapper/foo-test1", "size": "6G", "type": "lvm", "uuid": "d739e41d-c1a1-4003-b357-25bf62753af5" }, "/dev/mapper/foo-test2": { "fstype": "ext4", "label": "", "name": "/dev/mapper/foo-test2", "size": "4G", "type": "lvm", "uuid": "64454f89-f2cd-4a8b-958f-4c900f72cf79" }, "/dev/nvme0n1": { "fstype": "", "label": "", "name": "/dev/nvme0n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme1n1": { "fstype": "", "label": "", "name": "/dev/nvme1n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme2n1": { "fstype": "", "label": "", "name": "/dev/nvme2n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sda": { "fstype": "LVM2_member", "label": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "HA17fj-P3dB-xyZl-9Zru-ZzXr-ULh6-93Sgui" }, "/dev/sdb": { "fstype": "", "label": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc": { "fstype": "", "label": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sr0": { "fstype": "iso9660", "label": "cidata", "name": "/dev/sr0", "size": "364K", "type": "rom", "uuid": "2022-07-21-19-04-29-00" }, "/dev/vda": { "fstype": "", "label": "", "name": "/dev/vda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vda1": { "fstype": "", "label": "", "name": "/dev/vda1", "size": "1M", "type": "partition", "uuid": "" }, "/dev/vda2": { "fstype": "vfat", "label": "", "name": "/dev/vda2", "size": "100M", "type": "partition", "uuid": "7B77-95E7" }, "/dev/vda3": { "fstype": "xfs", "label": "root", "name": "/dev/vda3", "size": "9.9G", "type": "partition", "uuid": "58dca2db-4ccd-4171-a017-dd53ad8c1c54" }, "/dev/vdb": { "fstype": "", "label": "", "name": "/dev/vdb", "size": "2G", "type": "disk", "uuid": "" }, "/dev/vdb1": { "fstype": "ext4", "label": "yumcache", "name": "/dev/vdb1", "size": "2G", "type": "partition", "uuid": "4d1ed9a1-5608-4ff8-ad89-21d95caa66fa" }, "/dev/vdc": { "fstype": "", "label": "", "name": "/dev/vdc", "size": "2G", "type": "disk", "uuid": "" }, "/dev/vdc1": { "fstype": "ext4", "label": "yumvarlib", "name": "/dev/vdc1", "size": "2G", "type": "partition", "uuid": "a7abcc97-edc5-476c-8ccd-694b4621fd28" }, "/dev/vdd": { "fstype": "", "label": "", "name": "/dev/vdd", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vde": { "fstype": "", "label": "", "name": "/dev/vde", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vdf": { "fstype": "", "label": "", "name": "/dev/vdf", "size": "10G", "type": "disk", "uuid": "" } } } TASK [Read the /etc/fstab file for volume existence] *************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-role-results.yml:19 Thursday 21 July 2022 19:05:57 +0000 (0:00:00.368) 0:01:14.981 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "cmd": [ "cat", "/etc/fstab" ], "delta": "0:00:00.003428", "end": "2022-07-21 15:05:57.868959", "rc": 0, "start": "2022-07-21 15:05:57.865531" } STDOUT: UUID=58dca2db-4ccd-4171-a017-dd53ad8c1c54 / xfs defaults 0 0 UUID=7B77-95E7 /boot/efi vfat defaults,uid=0,gid=0,umask=077,shortname=winnt 0 2 /dev/vdb1 /var/cache/dnf auto defaults,nofail,x-systemd.requires=cloud-init.service,_netdev,comment=cloudconfig 0 2 /dev/vdc1 /var/lib/dnf auto defaults,nofail,x-systemd.requires=cloud-init.service,_netdev,comment=cloudconfig 0 2 /dev/mapper/foo-test1 /opt/test1 xfs defaults 0 0 /dev/mapper/foo-test2 /opt/test2 ext4 defaults 0 0 TASK [Read the /etc/crypttab file] ********************************************* task path: /tmp/tmpmnqeja9w/tests/storage/verify-role-results.yml:24 Thursday 21 July 2022 19:05:57 +0000 (0:00:00.370) 0:01:15.352 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "cmd": [ "cat", "/etc/crypttab" ], "delta": "0:00:00.003050", "end": "2022-07-21 15:05:58.295528", "failed_when_result": false, "rc": 1, "start": "2022-07-21 15:05:58.292478" } STDERR: cat: /etc/crypttab: No such file or directory MSG: non-zero return code TASK [Verify the volumes listed in storage_pools were correctly managed] ******* task path: /tmp/tmpmnqeja9w/tests/storage/verify-role-results.yml:33 Thursday 21 July 2022 19:05:58 +0000 (0:00:00.423) 0:01:15.776 ********* included: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool.yml for /cache/rhel-8-y.qcow2.snap => (item={'disks': ['sda'], 'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'name': 'foo', 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_metadata_version': None, 'state': 'present', 'type': 'lvm', 'volumes': [{'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'name': 'test1', 'raid_level': None, 'size': '60%', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test1', '_raw_device': '/dev/mapper/foo-test1', '_mount_id': '/dev/mapper/foo-test1', '_kernel_device': '/dev/dm-1', '_raw_kernel_device': '/dev/dm-1'}, {'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'ext4', 'mount_options': 'defaults', 'mount_point': '/opt/test2', 'name': 'test2', 'raid_level': None, 'size': '40%', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test2', '_raw_device': '/dev/mapper/foo-test2', '_mount_id': '/dev/mapper/foo-test2', '_kernel_device': '/dev/dm-0', '_raw_kernel_device': '/dev/dm-0'}], 'raid_chunk_size': None}) TASK [Set _storage_pool_tests] ************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool.yml:5 Thursday 21 July 2022 19:05:58 +0000 (0:00:00.110) 0:01:15.886 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_pool_tests": [ "members", "volumes" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool.yml:18 Thursday 21 July 2022 19:05:58 +0000 (0:00:00.037) 0:01:15.923 ********* included: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml for /cache/rhel-8-y.qcow2.snap => (item=members) included: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-volumes.yml for /cache/rhel-8-y.qcow2.snap => (item=volumes) TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:1 Thursday 21 July 2022 19:05:58 +0000 (0:00:00.051) 0:01:15.974 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_pv_count": "1", "_storage_test_pool_pvs_lvm": [ "/dev/sda" ] }, "changed": false } TASK [Get the canonical device path for each member device] ******************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:6 Thursday 21 July 2022 19:05:58 +0000 (0:00:00.057) 0:01:16.031 ********* ok: [/cache/rhel-8-y.qcow2.snap] => (item=/dev/sda) => { "ansible_loop_var": "pv", "changed": false, "device": "/dev/sda", "pv": "/dev/sda" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:15 Thursday 21 July 2022 19:05:58 +0000 (0:00:00.380) 0:01:16.412 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "__pvs_lvm_len": "1" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:19 Thursday 21 July 2022 19:05:58 +0000 (0:00:00.050) 0:01:16.462 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_pool_pvs": [ "/dev/sda" ] }, "changed": false } TASK [Verify PV count] ********************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:23 Thursday 21 July 2022 19:05:58 +0000 (0:00:00.050) 0:01:16.513 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:29 Thursday 21 July 2022 19:05:59 +0000 (0:00:00.050) 0:01:16.563 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:33 Thursday 21 July 2022 19:05:59 +0000 (0:00:00.037) 0:01:16.600 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:37 Thursday 21 July 2022 19:05:59 +0000 (0:00:00.049) 0:01:16.650 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check the type of each PV] *********************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:41 Thursday 21 July 2022 19:05:59 +0000 (0:00:00.023) 0:01:16.673 ********* ok: [/cache/rhel-8-y.qcow2.snap] => (item=/dev/sda) => { "ansible_loop_var": "pv", "changed": false, "pv": "/dev/sda" } MSG: All assertions passed TASK [Check MD RAID] *********************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:50 Thursday 21 July 2022 19:05:59 +0000 (0:00:00.044) 0:01:16.717 ********* included: /tmp/tmpmnqeja9w/tests/storage/verify-pool-md.yml for /cache/rhel-8-y.qcow2.snap TASK [get information about RAID] ********************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-md.yml:6 Thursday 21 July 2022 19:05:59 +0000 (0:00:00.043) 0:01:16.761 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-md.yml:12 Thursday 21 July 2022 19:05:59 +0000 (0:00:00.025) 0:01:16.787 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-md.yml:16 Thursday 21 July 2022 19:05:59 +0000 (0:00:00.028) 0:01:16.815 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-md.yml:20 Thursday 21 July 2022 19:05:59 +0000 (0:00:00.024) 0:01:16.839 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-md.yml:24 Thursday 21 July 2022 19:05:59 +0000 (0:00:00.025) 0:01:16.865 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-md.yml:30 Thursday 21 July 2022 19:05:59 +0000 (0:00:00.025) 0:01:16.890 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-md.yml:36 Thursday 21 July 2022 19:05:59 +0000 (0:00:00.028) 0:01:16.918 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-md.yml:44 Thursday 21 July 2022 19:05:59 +0000 (0:00:00.024) 0:01:16.942 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_md_active_devices_re": null, "storage_test_md_metadata_version_re": null, "storage_test_md_spare_devices_re": null }, "changed": false } TASK [Check LVM RAID] ********************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:53 Thursday 21 July 2022 19:05:59 +0000 (0:00:00.037) 0:01:16.980 ********* included: /tmp/tmpmnqeja9w/tests/storage/verify-pool-members-lvmraid.yml for /cache/rhel-8-y.qcow2.snap TASK [Validate pool member LVM RAID settings] ********************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-members-lvmraid.yml:1 Thursday 21 July 2022 19:05:59 +0000 (0:00:00.047) 0:01:17.027 ********* included: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-lvmraid.yml for /cache/rhel-8-y.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'name': 'test1', 'raid_level': None, 'size': '60%', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test1', '_raw_device': '/dev/mapper/foo-test1', '_mount_id': '/dev/mapper/foo-test1', '_kernel_device': '/dev/dm-1', '_raw_kernel_device': '/dev/dm-1'}) included: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-lvmraid.yml for /cache/rhel-8-y.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'ext4', 'mount_options': 'defaults', 'mount_point': '/opt/test2', 'name': 'test2', 'raid_level': None, 'size': '40%', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test2', '_raw_device': '/dev/mapper/foo-test2', '_mount_id': '/dev/mapper/foo-test2', '_kernel_device': '/dev/dm-0', '_raw_kernel_device': '/dev/dm-0'}) TASK [Get information about LVM RAID] ****************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-lvmraid.yml:3 Thursday 21 July 2022 19:05:59 +0000 (0:00:00.052) 0:01:17.080 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is LVM RAID] ******************************************* task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-lvmraid.yml:8 Thursday 21 July 2022 19:05:59 +0000 (0:00:00.031) 0:01:17.112 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-lvmraid.yml:12 Thursday 21 July 2022 19:05:59 +0000 (0:00:00.029) 0:01:17.141 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about LVM RAID] ****************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-lvmraid.yml:3 Thursday 21 July 2022 19:05:59 +0000 (0:00:00.029) 0:01:17.171 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is LVM RAID] ******************************************* task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-lvmraid.yml:8 Thursday 21 July 2022 19:05:59 +0000 (0:00:00.029) 0:01:17.201 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-lvmraid.yml:12 Thursday 21 July 2022 19:05:59 +0000 (0:00:00.033) 0:01:17.235 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check Thin Pools] ******************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:56 Thursday 21 July 2022 19:05:59 +0000 (0:00:00.039) 0:01:17.275 ********* included: /tmp/tmpmnqeja9w/tests/storage/verify-pool-members-thin.yml for /cache/rhel-8-y.qcow2.snap TASK [Validate pool member thinpool settings] ********************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-members-thin.yml:1 Thursday 21 July 2022 19:05:59 +0000 (0:00:00.047) 0:01:17.322 ********* included: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-thin.yml for /cache/rhel-8-y.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'name': 'test1', 'raid_level': None, 'size': '60%', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test1', '_raw_device': '/dev/mapper/foo-test1', '_mount_id': '/dev/mapper/foo-test1', '_kernel_device': '/dev/dm-1', '_raw_kernel_device': '/dev/dm-1'}) included: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-thin.yml for /cache/rhel-8-y.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'ext4', 'mount_options': 'defaults', 'mount_point': '/opt/test2', 'name': 'test2', 'raid_level': None, 'size': '40%', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test2', '_raw_device': '/dev/mapper/foo-test2', '_mount_id': '/dev/mapper/foo-test2', '_kernel_device': '/dev/dm-0', '_raw_kernel_device': '/dev/dm-0'}) TASK [Get information about thinpool] ****************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-thin.yml:3 Thursday 21 July 2022 19:05:59 +0000 (0:00:00.093) 0:01:17.416 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-thin.yml:8 Thursday 21 July 2022 19:05:59 +0000 (0:00:00.025) 0:01:17.441 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-thin.yml:13 Thursday 21 July 2022 19:05:59 +0000 (0:00:00.023) 0:01:17.465 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-thin.yml:17 Thursday 21 July 2022 19:05:59 +0000 (0:00:00.026) 0:01:17.491 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about thinpool] ****************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-thin.yml:3 Thursday 21 July 2022 19:05:59 +0000 (0:00:00.025) 0:01:17.516 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-thin.yml:8 Thursday 21 July 2022 19:05:59 +0000 (0:00:00.024) 0:01:17.541 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-thin.yml:13 Thursday 21 July 2022 19:06:00 +0000 (0:00:00.024) 0:01:17.566 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-thin.yml:17 Thursday 21 July 2022 19:06:00 +0000 (0:00:00.026) 0:01:17.592 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check member encryption] ************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:59 Thursday 21 July 2022 19:06:00 +0000 (0:00:00.023) 0:01:17.616 ********* included: /tmp/tmpmnqeja9w/tests/storage/verify-pool-members-encryption.yml for /cache/rhel-8-y.qcow2.snap TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-members-encryption.yml:4 Thursday 21 July 2022 19:06:00 +0000 (0:00:00.046) 0:01:17.663 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Validate pool member LUKS settings] ************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-members-encryption.yml:8 Thursday 21 July 2022 19:06:00 +0000 (0:00:00.051) 0:01:17.714 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => (item=/dev/sda) => { "_storage_test_pool_member_path": "/dev/sda", "ansible_loop_var": "_storage_test_pool_member_path", "changed": false, "skip_reason": "Conditional result was False" } TASK [Validate pool member crypttab entries] *********************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-members-encryption.yml:15 Thursday 21 July 2022 19:06:00 +0000 (0:00:00.030) 0:01:17.744 ********* included: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-crypttab.yml for /cache/rhel-8-y.qcow2.snap => (item=/dev/sda) TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-crypttab.yml:1 Thursday 21 July 2022 19:06:00 +0000 (0:00:00.045) 0:01:17.790 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": [] }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-crypttab.yml:4 Thursday 21 July 2022 19:06:00 +0000 (0:00:00.051) 0:01:17.842 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-crypttab.yml:9 Thursday 21 July 2022 19:06:00 +0000 (0:00:00.049) 0:01:17.892 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-crypttab.yml:15 Thursday 21 July 2022 19:06:00 +0000 (0:00:00.036) 0:01:17.928 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-crypttab.yml:21 Thursday 21 July 2022 19:06:00 +0000 (0:00:00.034) 0:01:17.963 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-crypttab.yml:27 Thursday 21 July 2022 19:06:00 +0000 (0:00:00.034) 0:01:17.998 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-members-encryption.yml:22 Thursday 21 July 2022 19:06:00 +0000 (0:00:00.035) 0:01:18.034 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_crypttab_key_file": null }, "changed": false } TASK [Check VDO] *************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:62 Thursday 21 July 2022 19:06:00 +0000 (0:00:00.040) 0:01:18.074 ********* included: /tmp/tmpmnqeja9w/tests/storage/verify-pool-members-vdo.yml for /cache/rhel-8-y.qcow2.snap TASK [Validate pool member VDO settings] *************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-members-vdo.yml:1 Thursday 21 July 2022 19:06:00 +0000 (0:00:00.049) 0:01:18.123 ********* included: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-vdo.yml for /cache/rhel-8-y.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'name': 'test1', 'raid_level': None, 'size': '60%', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test1', '_raw_device': '/dev/mapper/foo-test1', '_mount_id': '/dev/mapper/foo-test1', '_kernel_device': '/dev/dm-1', '_raw_kernel_device': '/dev/dm-1'}) included: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-vdo.yml for /cache/rhel-8-y.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'ext4', 'mount_options': 'defaults', 'mount_point': '/opt/test2', 'name': 'test2', 'raid_level': None, 'size': '40%', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test2', '_raw_device': '/dev/mapper/foo-test2', '_mount_id': '/dev/mapper/foo-test2', '_kernel_device': '/dev/dm-0', '_raw_kernel_device': '/dev/dm-0'}) TASK [get information about VDO deduplication] ********************************* task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-vdo.yml:3 Thursday 21 July 2022 19:06:00 +0000 (0:00:00.057) 0:01:18.181 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-vdo.yml:8 Thursday 21 July 2022 19:06:00 +0000 (0:00:00.023) 0:01:18.204 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-vdo.yml:11 Thursday 21 July 2022 19:06:00 +0000 (0:00:00.023) 0:01:18.228 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-vdo.yml:16 Thursday 21 July 2022 19:06:00 +0000 (0:00:00.024) 0:01:18.252 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-vdo.yml:21 Thursday 21 July 2022 19:06:00 +0000 (0:00:00.022) 0:01:18.275 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-vdo.yml:24 Thursday 21 July 2022 19:06:00 +0000 (0:00:00.024) 0:01:18.299 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-vdo.yml:29 Thursday 21 July 2022 19:06:00 +0000 (0:00:00.024) 0:01:18.324 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-vdo.yml:39 Thursday 21 July 2022 19:06:00 +0000 (0:00:00.026) 0:01:18.350 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [get information about VDO deduplication] ********************************* task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-vdo.yml:3 Thursday 21 July 2022 19:06:00 +0000 (0:00:00.036) 0:01:18.387 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-vdo.yml:8 Thursday 21 July 2022 19:06:00 +0000 (0:00:00.024) 0:01:18.411 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-vdo.yml:11 Thursday 21 July 2022 19:06:00 +0000 (0:00:00.025) 0:01:18.437 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-vdo.yml:16 Thursday 21 July 2022 19:06:00 +0000 (0:00:00.026) 0:01:18.464 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-vdo.yml:21 Thursday 21 July 2022 19:06:00 +0000 (0:00:00.025) 0:01:18.489 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-vdo.yml:24 Thursday 21 July 2022 19:06:00 +0000 (0:00:00.024) 0:01:18.514 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-vdo.yml:29 Thursday 21 July 2022 19:06:00 +0000 (0:00:00.025) 0:01:18.539 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-vdo.yml:39 Thursday 21 July 2022 19:06:01 +0000 (0:00:00.026) 0:01:18.566 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:65 Thursday 21 July 2022 19:06:01 +0000 (0:00:00.072) 0:01:18.638 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "__pvs_lvm_len": null, "_storage_test_expected_pv_count": null, "_storage_test_expected_pv_type": null, "_storage_test_pool_pvs": [], "_storage_test_pool_pvs_lvm": [] }, "changed": false } TASK [verify the volumes] ****************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-volumes.yml:3 Thursday 21 July 2022 19:06:01 +0000 (0:00:00.034) 0:01:18.673 ********* included: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume.yml for /cache/rhel-8-y.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'name': 'test1', 'raid_level': None, 'size': '60%', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test1', '_raw_device': '/dev/mapper/foo-test1', '_mount_id': '/dev/mapper/foo-test1', '_kernel_device': '/dev/dm-1', '_raw_kernel_device': '/dev/dm-1'}) included: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume.yml for /cache/rhel-8-y.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'ext4', 'mount_options': 'defaults', 'mount_point': '/opt/test2', 'name': 'test2', 'raid_level': None, 'size': '40%', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test2', '_raw_device': '/dev/mapper/foo-test2', '_mount_id': '/dev/mapper/foo-test2', '_kernel_device': '/dev/dm-0', '_raw_kernel_device': '/dev/dm-0'}) TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume.yml:2 Thursday 21 July 2022 19:06:01 +0000 (0:00:00.049) 0:01:18.722 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_volume_present": true, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume.yml:10 Thursday 21 July 2022 19:06:01 +0000 (0:00:00.046) 0:01:18.769 ********* included: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml for /cache/rhel-8-y.qcow2.snap => (item=mount) included: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-fstab.yml for /cache/rhel-8-y.qcow2.snap => (item=fstab) included: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-fs.yml for /cache/rhel-8-y.qcow2.snap => (item=fs) included: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-device.yml for /cache/rhel-8-y.qcow2.snap => (item=device) included: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml for /cache/rhel-8-y.qcow2.snap => (item=encryption) included: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-md.yml for /cache/rhel-8-y.qcow2.snap => (item=md) included: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml for /cache/rhel-8-y.qcow2.snap => (item=size) included: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-cache.yml for /cache/rhel-8-y.qcow2.snap => (item=cache) TASK [Get expected mount device based on device type] ************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml:6 Thursday 21 July 2022 19:06:01 +0000 (0:00:00.093) 0:01:18.863 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/foo-test1" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml:10 Thursday 21 July 2022 19:06:01 +0000 (0:00:00.048) 0:01:18.912 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_mount_device_matches": [ { "block_available": 1551090, "block_size": 4096, "block_total": 1570304, "block_used": 19214, "device": "/dev/mapper/foo-test1", "fstype": "xfs", "inode_available": 3145725, "inode_total": 3145728, "inode_used": 3, "mount": "/opt/test1", "options": "rw,seclabel,relatime,attr2,inode64,logbufs=8,logbsize=32k,noquota", "size_available": 6353264640, "size_total": 6431965184, "uuid": "d739e41d-c1a1-4003-b357-25bf62753af5" } ], "storage_test_mount_expected_match_count": "1", "storage_test_mount_point_matches": [ { "block_available": 1551090, "block_size": 4096, "block_total": 1570304, "block_used": 19214, "device": "/dev/mapper/foo-test1", "fstype": "xfs", "inode_available": 3145725, "inode_total": 3145728, "inode_used": 3, "mount": "/opt/test1", "options": "rw,seclabel,relatime,attr2,inode64,logbufs=8,logbsize=32k,noquota", "size_available": 6353264640, "size_total": 6431965184, "uuid": "d739e41d-c1a1-4003-b357-25bf62753af5" } ], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Verify the current mount state by device] ******************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml:20 Thursday 21 July 2022 19:06:01 +0000 (0:00:00.056) 0:01:18.969 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify the current mount state by mount point] *************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml:29 Thursday 21 July 2022 19:06:01 +0000 (0:00:00.049) 0:01:19.018 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify the mount fs type] ************************************************ task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml:37 Thursday 21 July 2022 19:06:01 +0000 (0:00:00.052) 0:01:19.071 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [command] ***************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml:46 Thursday 21 July 2022 19:06:01 +0000 (0:00:00.050) 0:01:19.122 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml:50 Thursday 21 July 2022 19:06:01 +0000 (0:00:00.023) 0:01:19.146 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml:55 Thursday 21 July 2022 19:06:01 +0000 (0:00:00.023) 0:01:19.169 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml:65 Thursday 21 July 2022 19:06:01 +0000 (0:00:00.024) 0:01:19.194 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-fstab.yml:2 Thursday 21 July 2022 19:06:01 +0000 (0:00:00.032) 0:01:19.227 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "1", "storage_test_fstab_expected_mount_options_matches": "1", "storage_test_fstab_expected_mount_point_matches": "1", "storage_test_fstab_id_matches": [ "/dev/mapper/foo-test1 " ], "storage_test_fstab_mount_options_matches": [ " /opt/test1 xfs defaults " ], "storage_test_fstab_mount_point_matches": [ " /opt/test1 " ] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-fstab.yml:12 Thursday 21 July 2022 19:06:01 +0000 (0:00:00.061) 0:01:19.288 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify the fstab mount point] ******************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-fstab.yml:19 Thursday 21 July 2022 19:06:01 +0000 (0:00:00.054) 0:01:19.342 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-fstab.yml:25 Thursday 21 July 2022 19:06:01 +0000 (0:00:00.049) 0:01:19.392 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-fstab.yml:34 Thursday 21 July 2022 19:06:01 +0000 (0:00:00.034) 0:01:19.427 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-fs.yml:4 Thursday 21 July 2022 19:06:01 +0000 (0:00:00.035) 0:01:19.462 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify fs label] ********************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-fs.yml:10 Thursday 21 July 2022 19:06:01 +0000 (0:00:00.039) 0:01:19.502 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [See whether the device node is present] ********************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-device.yml:4 Thursday 21 July 2022 19:06:01 +0000 (0:00:00.039) 0:01:19.541 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "stat": { "atime": 1658430322.6091166, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1658430322.6091166, "dev": 6, "device_type": 64769, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 60295, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/symlink", "mode": "0660", "mtime": 1658430322.6091166, "nlink": 1, "path": "/dev/mapper/foo-test1", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-device.yml:10 Thursday 21 July 2022 19:06:02 +0000 (0:00:00.377) 0:01:19.918 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about this volume] ********************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-device.yml:15 Thursday 21 July 2022 19:06:02 +0000 (0:00:00.039) 0:01:19.958 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [(1/2) Process volume type (set initial value)] *************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-device.yml:21 Thursday 21 July 2022 19:06:02 +0000 (0:00:00.040) 0:01:19.998 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [(2/2) Process volume type (get RAID value)] ****************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-device.yml:25 Thursday 21 July 2022 19:06:02 +0000 (0:00:00.036) 0:01:20.035 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-device.yml:30 Thursday 21 July 2022 19:06:02 +0000 (0:00:00.023) 0:01:20.059 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Stat the LUKS device, if encrypted] ************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:3 Thursday 21 July 2022 19:06:02 +0000 (0:00:00.039) 0:01:20.098 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:10 Thursday 21 July 2022 19:06:02 +0000 (0:00:00.026) 0:01:20.124 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [Collect LUKS info for this volume] *************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:15 Thursday 21 July 2022 19:06:03 +0000 (0:00:01.389) 0:01:21.514 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:21 Thursday 21 July 2022 19:06:03 +0000 (0:00:00.026) 0:01:21.540 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:27 Thursday 21 July 2022 19:06:04 +0000 (0:00:00.027) 0:01:21.567 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:33 Thursday 21 July 2022 19:06:04 +0000 (0:00:00.056) 0:01:21.624 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:39 Thursday 21 July 2022 19:06:04 +0000 (0:00:00.025) 0:01:21.650 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:44 Thursday 21 July 2022 19:06:04 +0000 (0:00:00.026) 0:01:21.676 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:50 Thursday 21 July 2022 19:06:04 +0000 (0:00:00.025) 0:01:21.702 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:56 Thursday 21 July 2022 19:06:04 +0000 (0:00:00.024) 0:01:21.726 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:62 Thursday 21 July 2022 19:06:04 +0000 (0:00:00.024) 0:01:21.750 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:67 Thursday 21 July 2022 19:06:04 +0000 (0:00:00.052) 0:01:21.803 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:72 Thursday 21 July 2022 19:06:04 +0000 (0:00:00.048) 0:01:21.851 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:78 Thursday 21 July 2022 19:06:04 +0000 (0:00:00.038) 0:01:21.890 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:84 Thursday 21 July 2022 19:06:04 +0000 (0:00:00.038) 0:01:21.929 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:90 Thursday 21 July 2022 19:06:04 +0000 (0:00:00.042) 0:01:21.971 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [get information about RAID] ********************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-md.yml:7 Thursday 21 July 2022 19:06:04 +0000 (0:00:00.037) 0:01:22.008 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-md.yml:13 Thursday 21 July 2022 19:06:04 +0000 (0:00:00.038) 0:01:22.047 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-md.yml:17 Thursday 21 July 2022 19:06:04 +0000 (0:00:00.037) 0:01:22.084 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-md.yml:21 Thursday 21 July 2022 19:06:04 +0000 (0:00:00.040) 0:01:22.125 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-md.yml:25 Thursday 21 July 2022 19:06:04 +0000 (0:00:00.039) 0:01:22.164 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-md.yml:31 Thursday 21 July 2022 19:06:04 +0000 (0:00:00.037) 0:01:22.202 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-md.yml:37 Thursday 21 July 2022 19:06:04 +0000 (0:00:00.036) 0:01:22.239 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the actual size of the volume] ************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:3 Thursday 21 July 2022 19:06:04 +0000 (0:00:00.040) 0:01:22.280 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "bytes": 6442450944, "changed": false, "lvm": "6g", "parted": "6GiB", "size": "6 GiB" } TASK [parse the requested size of the volume] ********************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:9 Thursday 21 July 2022 19:06:05 +0000 (0:00:00.393) 0:01:22.674 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Establish base value for expected size] ********************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:15 Thursday 21 July 2022 19:06:05 +0000 (0:00:00.038) 0:01:22.713 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:20 Thursday 21 July 2022 19:06:05 +0000 (0:00:00.041) 0:01:22.754 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_expected_size": "4294967296.0" } TASK [debug] ******************************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:25 Thursday 21 July 2022 19:06:05 +0000 (0:00:00.037) 0:01:22.792 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_pool": { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "60%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test2", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test2", "_raw_device": "/dev/mapper/foo-test2", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "40%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } } TASK [debug] ******************************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:28 Thursday 21 July 2022 19:06:05 +0000 (0:00:00.095) 0:01:22.888 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_blkinfo": { "changed": false, "failed": false, "info": { "/dev/mapper/foo-test1": { "fstype": "xfs", "label": "", "name": "/dev/mapper/foo-test1", "size": "6G", "type": "lvm", "uuid": "d739e41d-c1a1-4003-b357-25bf62753af5" }, "/dev/mapper/foo-test2": { "fstype": "ext4", "label": "", "name": "/dev/mapper/foo-test2", "size": "4G", "type": "lvm", "uuid": "64454f89-f2cd-4a8b-958f-4c900f72cf79" }, "/dev/nvme0n1": { "fstype": "", "label": "", "name": "/dev/nvme0n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme1n1": { "fstype": "", "label": "", "name": "/dev/nvme1n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme2n1": { "fstype": "", "label": "", "name": "/dev/nvme2n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sda": { "fstype": "LVM2_member", "label": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "HA17fj-P3dB-xyZl-9Zru-ZzXr-ULh6-93Sgui" }, "/dev/sdb": { "fstype": "", "label": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc": { "fstype": "", "label": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sr0": { "fstype": "iso9660", "label": "cidata", "name": "/dev/sr0", "size": "364K", "type": "rom", "uuid": "2022-07-21-19-04-29-00" }, "/dev/vda": { "fstype": "", "label": "", "name": "/dev/vda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vda1": { "fstype": "", "label": "", "name": "/dev/vda1", "size": "1M", "type": "partition", "uuid": "" }, "/dev/vda2": { "fstype": "vfat", "label": "", "name": "/dev/vda2", "size": "100M", "type": "partition", "uuid": "7B77-95E7" }, "/dev/vda3": { "fstype": "xfs", "label": "root", "name": "/dev/vda3", "size": "9.9G", "type": "partition", "uuid": "58dca2db-4ccd-4171-a017-dd53ad8c1c54" }, "/dev/vdb": { "fstype": "", "label": "", "name": "/dev/vdb", "size": "2G", "type": "disk", "uuid": "" }, "/dev/vdb1": { "fstype": "ext4", "label": "yumcache", "name": "/dev/vdb1", "size": "2G", "type": "partition", "uuid": "4d1ed9a1-5608-4ff8-ad89-21d95caa66fa" }, "/dev/vdc": { "fstype": "", "label": "", "name": "/dev/vdc", "size": "2G", "type": "disk", "uuid": "" }, "/dev/vdc1": { "fstype": "ext4", "label": "yumvarlib", "name": "/dev/vdc1", "size": "2G", "type": "partition", "uuid": "a7abcc97-edc5-476c-8ccd-694b4621fd28" }, "/dev/vdd": { "fstype": "", "label": "", "name": "/dev/vdd", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vde": { "fstype": "", "label": "", "name": "/dev/vde", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vdf": { "fstype": "", "label": "", "name": "/dev/vdf", "size": "10G", "type": "disk", "uuid": "" } } } } TASK [Get the size of parent/pool device] ************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:31 Thursday 21 July 2022 19:06:05 +0000 (0:00:00.054) 0:01:22.942 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "bytes": 10737418240, "changed": false, "lvm": "10g", "parted": "10GiB", "size": "10 GiB" } TASK [debug] ******************************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:36 Thursday 21 July 2022 19:06:05 +0000 (0:00:00.433) 0:01:23.376 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_pool_size": { "bytes": 10737418240, "changed": false, "failed": false, "lvm": "10g", "parted": "10GiB", "size": "10 GiB" } } TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:39 Thursday 21 July 2022 19:06:05 +0000 (0:00:00.092) 0:01:23.468 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_expected_size": "6442450944.0" }, "changed": false } TASK [debug] ******************************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:44 Thursday 21 July 2022 19:06:06 +0000 (0:00:00.090) 0:01:23.559 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_actual_size": { "bytes": 6442450944, "changed": false, "failed": false, "lvm": "6g", "parted": "6GiB", "size": "6 GiB" } } TASK [debug] ******************************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:47 Thursday 21 July 2022 19:06:06 +0000 (0:00:00.078) 0:01:23.637 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_expected_size": "6442450944.0" } TASK [assert] ****************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:50 Thursday 21 July 2022 19:06:06 +0000 (0:00:00.110) 0:01:23.747 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Get information about the LV] ******************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-cache.yml:6 Thursday 21 July 2022 19:06:06 +0000 (0:00:00.055) 0:01:23.803 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "cmd": [ "lvs", "--noheadings", "--nameprefixes", "--units=b", "--nosuffix", "--unquoted", "-o", "name,attr,cache_total_blocks,chunk_size,segtype", "foo/test1" ], "delta": "0:00:00.036126", "end": "2022-07-21 15:06:06.740813", "rc": 0, "start": "2022-07-21 15:06:06.704687" } STDOUT: LVM2_LV_NAME=test1 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-cache.yml:14 Thursday 21 July 2022 19:06:06 +0000 (0:00:00.420) 0:01:24.224 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_lv_segtype": [ "linear" ] }, "changed": false } TASK [check segment type] ****************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-cache.yml:17 Thursday 21 July 2022 19:06:06 +0000 (0:00:00.051) 0:01:24.276 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-cache.yml:22 Thursday 21 July 2022 19:06:06 +0000 (0:00:00.055) 0:01:24.331 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the requested cache size] ****************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-cache.yml:26 Thursday 21 July 2022 19:06:06 +0000 (0:00:00.042) 0:01:24.374 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-cache.yml:32 Thursday 21 July 2022 19:06:06 +0000 (0:00:00.038) 0:01:24.413 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-cache.yml:36 Thursday 21 July 2022 19:06:06 +0000 (0:00:00.039) 0:01:24.452 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume.yml:16 Thursday 21 July 2022 19:06:06 +0000 (0:00:00.037) 0:01:24.489 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume.yml:2 Thursday 21 July 2022 19:06:06 +0000 (0:00:00.041) 0:01:24.531 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_volume_present": true, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume.yml:10 Thursday 21 July 2022 19:06:07 +0000 (0:00:00.051) 0:01:24.582 ********* included: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml for /cache/rhel-8-y.qcow2.snap => (item=mount) included: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-fstab.yml for /cache/rhel-8-y.qcow2.snap => (item=fstab) included: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-fs.yml for /cache/rhel-8-y.qcow2.snap => (item=fs) included: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-device.yml for /cache/rhel-8-y.qcow2.snap => (item=device) included: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml for /cache/rhel-8-y.qcow2.snap => (item=encryption) included: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-md.yml for /cache/rhel-8-y.qcow2.snap => (item=md) included: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml for /cache/rhel-8-y.qcow2.snap => (item=size) included: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-cache.yml for /cache/rhel-8-y.qcow2.snap => (item=cache) TASK [Get expected mount device based on device type] ************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml:6 Thursday 21 July 2022 19:06:07 +0000 (0:00:00.081) 0:01:24.663 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/foo-test2" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml:10 Thursday 21 July 2022 19:06:07 +0000 (0:00:00.041) 0:01:24.705 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_mount_device_matches": [ { "block_available": 954137, "block_size": 4096, "block_total": 1010616, "block_used": 56479, "device": "/dev/mapper/foo-test2", "fstype": "ext4", "inode_available": 262133, "inode_total": 262144, "inode_used": 11, "mount": "/opt/test2", "options": "rw,seclabel,relatime", "size_available": 3908145152, "size_total": 4139483136, "uuid": "64454f89-f2cd-4a8b-958f-4c900f72cf79" } ], "storage_test_mount_expected_match_count": "1", "storage_test_mount_point_matches": [ { "block_available": 954137, "block_size": 4096, "block_total": 1010616, "block_used": 56479, "device": "/dev/mapper/foo-test2", "fstype": "ext4", "inode_available": 262133, "inode_total": 262144, "inode_used": 11, "mount": "/opt/test2", "options": "rw,seclabel,relatime", "size_available": 3908145152, "size_total": 4139483136, "uuid": "64454f89-f2cd-4a8b-958f-4c900f72cf79" } ], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Verify the current mount state by device] ******************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml:20 Thursday 21 July 2022 19:06:07 +0000 (0:00:00.057) 0:01:24.762 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify the current mount state by mount point] *************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml:29 Thursday 21 July 2022 19:06:07 +0000 (0:00:00.051) 0:01:24.814 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify the mount fs type] ************************************************ task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml:37 Thursday 21 July 2022 19:06:07 +0000 (0:00:00.052) 0:01:24.867 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [command] ***************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml:46 Thursday 21 July 2022 19:06:07 +0000 (0:00:00.058) 0:01:24.925 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml:50 Thursday 21 July 2022 19:06:07 +0000 (0:00:00.027) 0:01:24.952 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml:55 Thursday 21 July 2022 19:06:07 +0000 (0:00:00.026) 0:01:24.979 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml:65 Thursday 21 July 2022 19:06:07 +0000 (0:00:00.026) 0:01:25.005 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-fstab.yml:2 Thursday 21 July 2022 19:06:07 +0000 (0:00:00.040) 0:01:25.046 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "1", "storage_test_fstab_expected_mount_options_matches": "1", "storage_test_fstab_expected_mount_point_matches": "1", "storage_test_fstab_id_matches": [ "/dev/mapper/foo-test2 " ], "storage_test_fstab_mount_options_matches": [ " /opt/test2 ext4 defaults " ], "storage_test_fstab_mount_point_matches": [ " /opt/test2 " ] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-fstab.yml:12 Thursday 21 July 2022 19:06:07 +0000 (0:00:00.063) 0:01:25.110 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify the fstab mount point] ******************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-fstab.yml:19 Thursday 21 July 2022 19:06:07 +0000 (0:00:00.053) 0:01:25.163 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-fstab.yml:25 Thursday 21 July 2022 19:06:07 +0000 (0:00:00.052) 0:01:25.216 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-fstab.yml:34 Thursday 21 July 2022 19:06:07 +0000 (0:00:00.035) 0:01:25.252 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-fs.yml:4 Thursday 21 July 2022 19:06:07 +0000 (0:00:00.033) 0:01:25.285 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify fs label] ********************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-fs.yml:10 Thursday 21 July 2022 19:06:07 +0000 (0:00:00.037) 0:01:25.323 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [See whether the device node is present] ********************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-device.yml:4 Thursday 21 July 2022 19:06:07 +0000 (0:00:00.044) 0:01:25.367 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "stat": { "atime": 1658430353.2021167, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1658430353.2021167, "dev": 6, "device_type": 64768, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 59192, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/symlink", "mode": "0660", "mtime": 1658430353.2021167, "nlink": 1, "path": "/dev/mapper/foo-test2", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-device.yml:10 Thursday 21 July 2022 19:06:08 +0000 (0:00:00.385) 0:01:25.753 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about this volume] ********************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-device.yml:15 Thursday 21 July 2022 19:06:08 +0000 (0:00:00.038) 0:01:25.792 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [(1/2) Process volume type (set initial value)] *************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-device.yml:21 Thursday 21 July 2022 19:06:08 +0000 (0:00:00.039) 0:01:25.831 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [(2/2) Process volume type (get RAID value)] ****************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-device.yml:25 Thursday 21 July 2022 19:06:08 +0000 (0:00:00.038) 0:01:25.869 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-device.yml:30 Thursday 21 July 2022 19:06:08 +0000 (0:00:00.028) 0:01:25.898 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Stat the LUKS device, if encrypted] ************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:3 Thursday 21 July 2022 19:06:08 +0000 (0:00:00.040) 0:01:25.938 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:10 Thursday 21 July 2022 19:06:08 +0000 (0:00:00.024) 0:01:25.963 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [Collect LUKS info for this volume] *************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:15 Thursday 21 July 2022 19:06:09 +0000 (0:00:01.352) 0:01:27.315 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:21 Thursday 21 July 2022 19:06:09 +0000 (0:00:00.024) 0:01:27.340 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:27 Thursday 21 July 2022 19:06:09 +0000 (0:00:00.022) 0:01:27.363 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:33 Thursday 21 July 2022 19:06:09 +0000 (0:00:00.051) 0:01:27.415 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:39 Thursday 21 July 2022 19:06:09 +0000 (0:00:00.023) 0:01:27.438 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:44 Thursday 21 July 2022 19:06:09 +0000 (0:00:00.028) 0:01:27.467 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:50 Thursday 21 July 2022 19:06:09 +0000 (0:00:00.028) 0:01:27.496 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:56 Thursday 21 July 2022 19:06:09 +0000 (0:00:00.029) 0:01:27.525 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:62 Thursday 21 July 2022 19:06:10 +0000 (0:00:00.026) 0:01:27.552 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:67 Thursday 21 July 2022 19:06:10 +0000 (0:00:00.055) 0:01:27.608 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:72 Thursday 21 July 2022 19:06:10 +0000 (0:00:00.068) 0:01:27.677 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:78 Thursday 21 July 2022 19:06:10 +0000 (0:00:00.040) 0:01:27.718 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:84 Thursday 21 July 2022 19:06:10 +0000 (0:00:00.040) 0:01:27.758 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:90 Thursday 21 July 2022 19:06:10 +0000 (0:00:00.037) 0:01:27.795 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [get information about RAID] ********************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-md.yml:7 Thursday 21 July 2022 19:06:10 +0000 (0:00:00.037) 0:01:27.833 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-md.yml:13 Thursday 21 July 2022 19:06:10 +0000 (0:00:00.036) 0:01:27.869 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-md.yml:17 Thursday 21 July 2022 19:06:10 +0000 (0:00:00.036) 0:01:27.906 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-md.yml:21 Thursday 21 July 2022 19:06:10 +0000 (0:00:00.036) 0:01:27.942 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-md.yml:25 Thursday 21 July 2022 19:06:10 +0000 (0:00:00.040) 0:01:27.983 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-md.yml:31 Thursday 21 July 2022 19:06:10 +0000 (0:00:00.039) 0:01:28.023 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-md.yml:37 Thursday 21 July 2022 19:06:10 +0000 (0:00:00.043) 0:01:28.066 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the actual size of the volume] ************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:3 Thursday 21 July 2022 19:06:10 +0000 (0:00:00.037) 0:01:28.104 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "bytes": 4294967296, "changed": false, "lvm": "4g", "parted": "4GiB", "size": "4 GiB" } TASK [parse the requested size of the volume] ********************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:9 Thursday 21 July 2022 19:06:10 +0000 (0:00:00.380) 0:01:28.484 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Establish base value for expected size] ********************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:15 Thursday 21 July 2022 19:06:10 +0000 (0:00:00.040) 0:01:28.525 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:20 Thursday 21 July 2022 19:06:11 +0000 (0:00:00.039) 0:01:28.564 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_expected_size": "6442450944.0" } TASK [debug] ******************************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:25 Thursday 21 July 2022 19:06:11 +0000 (0:00:00.040) 0:01:28.605 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_pool": { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "60%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test2", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test2", "_raw_device": "/dev/mapper/foo-test2", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "40%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } } TASK [debug] ******************************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:28 Thursday 21 July 2022 19:06:11 +0000 (0:00:00.053) 0:01:28.658 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_blkinfo": { "changed": false, "failed": false, "info": { "/dev/mapper/foo-test1": { "fstype": "xfs", "label": "", "name": "/dev/mapper/foo-test1", "size": "6G", "type": "lvm", "uuid": "d739e41d-c1a1-4003-b357-25bf62753af5" }, "/dev/mapper/foo-test2": { "fstype": "ext4", "label": "", "name": "/dev/mapper/foo-test2", "size": "4G", "type": "lvm", "uuid": "64454f89-f2cd-4a8b-958f-4c900f72cf79" }, "/dev/nvme0n1": { "fstype": "", "label": "", "name": "/dev/nvme0n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme1n1": { "fstype": "", "label": "", "name": "/dev/nvme1n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme2n1": { "fstype": "", "label": "", "name": "/dev/nvme2n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sda": { "fstype": "LVM2_member", "label": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "HA17fj-P3dB-xyZl-9Zru-ZzXr-ULh6-93Sgui" }, "/dev/sdb": { "fstype": "", "label": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc": { "fstype": "", "label": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sr0": { "fstype": "iso9660", "label": "cidata", "name": "/dev/sr0", "size": "364K", "type": "rom", "uuid": "2022-07-21-19-04-29-00" }, "/dev/vda": { "fstype": "", "label": "", "name": "/dev/vda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vda1": { "fstype": "", "label": "", "name": "/dev/vda1", "size": "1M", "type": "partition", "uuid": "" }, "/dev/vda2": { "fstype": "vfat", "label": "", "name": "/dev/vda2", "size": "100M", "type": "partition", "uuid": "7B77-95E7" }, "/dev/vda3": { "fstype": "xfs", "label": "root", "name": "/dev/vda3", "size": "9.9G", "type": "partition", "uuid": "58dca2db-4ccd-4171-a017-dd53ad8c1c54" }, "/dev/vdb": { "fstype": "", "label": "", "name": "/dev/vdb", "size": "2G", "type": "disk", "uuid": "" }, "/dev/vdb1": { "fstype": "ext4", "label": "yumcache", "name": "/dev/vdb1", "size": "2G", "type": "partition", "uuid": "4d1ed9a1-5608-4ff8-ad89-21d95caa66fa" }, "/dev/vdc": { "fstype": "", "label": "", "name": "/dev/vdc", "size": "2G", "type": "disk", "uuid": "" }, "/dev/vdc1": { "fstype": "ext4", "label": "yumvarlib", "name": "/dev/vdc1", "size": "2G", "type": "partition", "uuid": "a7abcc97-edc5-476c-8ccd-694b4621fd28" }, "/dev/vdd": { "fstype": "", "label": "", "name": "/dev/vdd", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vde": { "fstype": "", "label": "", "name": "/dev/vde", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vdf": { "fstype": "", "label": "", "name": "/dev/vdf", "size": "10G", "type": "disk", "uuid": "" } } } } TASK [Get the size of parent/pool device] ************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:31 Thursday 21 July 2022 19:06:11 +0000 (0:00:00.096) 0:01:28.755 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "bytes": 10737418240, "changed": false, "lvm": "10g", "parted": "10GiB", "size": "10 GiB" } TASK [debug] ******************************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:36 Thursday 21 July 2022 19:06:11 +0000 (0:00:00.440) 0:01:29.195 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_pool_size": { "bytes": 10737418240, "changed": false, "failed": false, "lvm": "10g", "parted": "10GiB", "size": "10 GiB" } } TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:39 Thursday 21 July 2022 19:06:11 +0000 (0:00:00.053) 0:01:29.249 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_expected_size": "4294967296.0" }, "changed": false } TASK [debug] ******************************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:44 Thursday 21 July 2022 19:06:11 +0000 (0:00:00.056) 0:01:29.305 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_actual_size": { "bytes": 4294967296, "changed": false, "failed": false, "lvm": "4g", "parted": "4GiB", "size": "4 GiB" } } TASK [debug] ******************************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:47 Thursday 21 July 2022 19:06:11 +0000 (0:00:00.042) 0:01:29.347 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_expected_size": "4294967296.0" } TASK [assert] ****************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:50 Thursday 21 July 2022 19:06:11 +0000 (0:00:00.038) 0:01:29.386 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Get information about the LV] ******************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-cache.yml:6 Thursday 21 July 2022 19:06:11 +0000 (0:00:00.055) 0:01:29.441 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "cmd": [ "lvs", "--noheadings", "--nameprefixes", "--units=b", "--nosuffix", "--unquoted", "-o", "name,attr,cache_total_blocks,chunk_size,segtype", "foo/test2" ], "delta": "0:00:00.035281", "end": "2022-07-21 15:06:12.386789", "rc": 0, "start": "2022-07-21 15:06:12.351508" } STDOUT: LVM2_LV_NAME=test2 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-cache.yml:14 Thursday 21 July 2022 19:06:12 +0000 (0:00:00.428) 0:01:29.870 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_lv_segtype": [ "linear" ] }, "changed": false } TASK [check segment type] ****************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-cache.yml:17 Thursday 21 July 2022 19:06:12 +0000 (0:00:00.053) 0:01:29.923 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-cache.yml:22 Thursday 21 July 2022 19:06:12 +0000 (0:00:00.053) 0:01:29.977 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the requested cache size] ****************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-cache.yml:26 Thursday 21 July 2022 19:06:12 +0000 (0:00:00.040) 0:01:30.018 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-cache.yml:32 Thursday 21 July 2022 19:06:12 +0000 (0:00:00.040) 0:01:30.058 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-cache.yml:36 Thursday 21 July 2022 19:06:12 +0000 (0:00:00.040) 0:01:30.098 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume.yml:16 Thursday 21 July 2022 19:06:12 +0000 (0:00:00.039) 0:01:30.138 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [Verify the volumes with no pool were correctly managed] ****************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-role-results.yml:43 Thursday 21 July 2022 19:06:12 +0000 (0:00:00.039) 0:01:30.178 ********* TASK [Clean up variable namespace] ********************************************* task path: /tmp/tmpmnqeja9w/tests/storage/verify-role-results.yml:53 Thursday 21 July 2022 19:06:12 +0000 (0:00:00.021) 0:01:30.199 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_blkinfo": null, "storage_test_crypttab": null, "storage_test_fstab": null }, "changed": false } TASK [Shrink test2 volume via percentage-based size spec] ********************** task path: /tmp/tmpmnqeja9w/tests/storage/tests_lvm_percent_size.yml:97 Thursday 21 July 2022 19:06:12 +0000 (0:00:00.035) 0:01:30.235 ********* TASK [fedora.linux_system_roles.storage : set platform/version specific variables] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2 Thursday 21 July 2022 19:06:12 +0000 (0:00:00.054) 0:01:30.290 ********* included: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for /cache/rhel-8-y.qcow2.snap TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2 Thursday 21 July 2022 19:06:12 +0000 (0:00:00.038) 0:01:30.328 ********* ok: [/cache/rhel-8-y.qcow2.snap] TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7 Thursday 21 July 2022 19:06:13 +0000 (0:00:00.550) 0:01:30.879 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [/cache/rhel-8-y.qcow2.snap] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [/cache/rhel-8-y.qcow2.snap] => (item=RedHat_8.yml) => { "ansible_facts": { "blivet_package_list": [ "python3-blivet", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap", "vdo", "kmod-kvdo", "xfsprogs" ] }, "ansible_included_var_files": [ "/tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/vars/RedHat_8.yml" ], "ansible_loop_var": "item", "changed": false, "item": "RedHat_8.yml" } skipping: [/cache/rhel-8-y.qcow2.snap] => (item=RedHat_8.7.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat_8.7.yml", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : define an empty list of pools to be used in testing] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5 Thursday 21 July 2022 19:06:13 +0000 (0:00:00.063) 0:01:30.942 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : define an empty list of volumes to be used in testing] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9 Thursday 21 July 2022 19:06:13 +0000 (0:00:00.038) 0:01:30.981 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : include the appropriate provider tasks] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13 Thursday 21 July 2022 19:06:13 +0000 (0:00:00.085) 0:01:31.066 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount included: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for /cache/rhel-8-y.qcow2.snap TASK [fedora.linux_system_roles.storage : get a list of rpm packages installed on host machine] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2 Thursday 21 July 2022 19:06:13 +0000 (0:00:00.065) 0:01:31.132 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : make sure blivet is available] ******* task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:7 Thursday 21 July 2022 19:06:13 +0000 (0:00:00.022) 0:01:31.154 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [fedora.linux_system_roles.storage : show storage_pools] ****************** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:13 Thursday 21 July 2022 19:06:14 +0000 (0:00:01.374) 0:01:32.529 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_pools": [ { "disks": [ "sda" ], "name": "foo", "volumes": [ { "mount_point": "/opt/test1", "name": "test1", "size": "60%" }, { "mount_point": "/opt/test2", "name": "test2", "size": "25%" } ] } ] } TASK [fedora.linux_system_roles.storage : show storage_volumes] **************** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:18 Thursday 21 July 2022 19:06:15 +0000 (0:00:00.043) 0:01:32.572 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [fedora.linux_system_roles.storage : get required packages] *************** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:23 Thursday 21 July 2022 19:06:15 +0000 (0:00:00.040) 0:01:32.613 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [ "lvm2" ], "pools": [], "volumes": [] } TASK [fedora.linux_system_roles.storage : enable copr repositories if needed] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:35 Thursday 21 July 2022 19:06:16 +0000 (0:00:01.686) 0:01:34.300 ********* included: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml for /cache/rhel-8-y.qcow2.snap TASK [fedora.linux_system_roles.storage : check if the COPR support packages should be installed] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:2 Thursday 21 July 2022 19:06:16 +0000 (0:00:00.050) 0:01:34.350 ********* TASK [fedora.linux_system_roles.storage : make sure COPR support packages are present] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:13 Thursday 21 July 2022 19:06:16 +0000 (0:00:00.042) 0:01:34.393 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : enable COPRs] ************************ task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:18 Thursday 21 July 2022 19:06:16 +0000 (0:00:00.040) 0:01:34.434 ********* TASK [fedora.linux_system_roles.storage : make sure required packages are installed] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:41 Thursday 21 July 2022 19:06:16 +0000 (0:00:00.038) 0:01:34.472 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [fedora.linux_system_roles.storage : get service facts] ******************* task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:47 Thursday 21 July 2022 19:06:18 +0000 (0:00:01.394) 0:01:35.867 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "services": { "NetworkManager-dispatcher.service": { "name": "NetworkManager-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "NetworkManager-wait-online.service": { "name": "NetworkManager-wait-online.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "NetworkManager.service": { "name": "NetworkManager.service", "source": "systemd", "state": "running", "status": "enabled" }, "arp-ethers.service": { "name": "arp-ethers.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "auditd.service": { "name": "auditd.service", "source": "systemd", "state": "running", "status": "enabled" }, "auth-rpcgss-module.service": { "name": "auth-rpcgss-module.service", "source": "systemd", "state": "stopped", "status": "static" }, "autovt@.service": { "name": "autovt@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "blivet.service": { "name": "blivet.service", "source": "systemd", "state": "inactive", "status": "static" }, "blk-availability.service": { "name": "blk-availability.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "chrony-dnssrv@.service": { "name": "chrony-dnssrv@.service", "source": "systemd", "state": "unknown", "status": "static" }, "chrony-wait.service": { "name": "chrony-wait.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd.service": { "name": "chronyd.service", "source": "systemd", "state": "running", "status": "enabled" }, "cloud-config.service": { "name": "cloud-config.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-final.service": { "name": "cloud-final.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init-hotplugd.service": { "name": "cloud-init-hotplugd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cloud-init-local.service": { "name": "cloud-init-local.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init.service": { "name": "cloud-init.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cockpit-motd.service": { "name": "cockpit-motd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cockpit-wsinstance-http.service": { "name": "cockpit-wsinstance-http.service", "source": "systemd", "state": "inactive", "status": "static" }, "cockpit-wsinstance-https-factory@.service": { "name": "cockpit-wsinstance-https-factory@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cockpit-wsinstance-https@.service": { "name": "cockpit-wsinstance-https@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cockpit.service": { "name": "cockpit.service", "source": "systemd", "state": "inactive", "status": "static" }, "console-getty.service": { "name": "console-getty.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "container-getty@.service": { "name": "container-getty@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cpupower.service": { "name": "cpupower.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "crond.service": { "name": "crond.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-org.freedesktop.hostname1.service": { "name": "dbus-org.freedesktop.hostname1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.locale1.service": { "name": "dbus-org.freedesktop.locale1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.login1.service": { "name": "dbus-org.freedesktop.login1.service", "source": "systemd", "state": "active", "status": "static" }, "dbus-org.freedesktop.nm-dispatcher.service": { "name": "dbus-org.freedesktop.nm-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "dbus-org.freedesktop.portable1.service": { "name": "dbus-org.freedesktop.portable1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.timedate1.service": { "name": "dbus-org.freedesktop.timedate1.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "dbus.service": { "name": "dbus.service", "source": "systemd", "state": "running", "status": "static" }, "dbxtool.service": { "name": "dbxtool.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "debug-shell.service": { "name": "debug-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "display-manager.service": { "name": "display-manager.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "dm-event.service": { "name": "dm-event.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-makecache.service": { "name": "dnf-makecache.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-cmdline.service": { "name": "dracut-cmdline.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-initqueue.service": { "name": "dracut-initqueue.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-mount.service": { "name": "dracut-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-mount.service": { "name": "dracut-pre-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-pivot.service": { "name": "dracut-pre-pivot.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-trigger.service": { "name": "dracut-pre-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-udev.service": { "name": "dracut-pre-udev.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown.service": { "name": "dracut-shutdown.service", "source": "systemd", "state": "stopped", "status": "static" }, "emergency.service": { "name": "emergency.service", "source": "systemd", "state": "stopped", "status": "static" }, "fcoe.service": { "name": "fcoe.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "fstrim.service": { "name": "fstrim.service", "source": "systemd", "state": "inactive", "status": "static" }, "getty@.service": { "name": "getty@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "getty@tty1.service": { "name": "getty@tty1.service", "source": "systemd", "state": "running", "status": "active" }, "grub-boot-indeterminate.service": { "name": "grub-boot-indeterminate.service", "source": "systemd", "state": "inactive", "status": "static" }, "gssproxy.service": { "name": "gssproxy.service", "source": "systemd", "state": "running", "status": "disabled" }, "halt-local.service": { "name": "halt-local.service", "source": "systemd", "state": "inactive", "status": "static" }, "hv_kvp_daemon.service": { "name": "hv_kvp_daemon.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "import-state.service": { "name": "import-state.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "initrd-cleanup.service": { "name": "initrd-cleanup.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-parse-etc.service": { "name": "initrd-parse-etc.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-switch-root.service": { "name": "initrd-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-udevadm-cleanup-db.service": { "name": "initrd-udevadm-cleanup-db.service", "source": "systemd", "state": "stopped", "status": "static" }, "insights-client-boot.service": { "name": "insights-client-boot.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "insights-client-results.service": { "name": "insights-client-results.service", "source": "systemd", "state": "inactive", "status": "static" }, "insights-client.service": { "name": "insights-client.service", "source": "systemd", "state": "inactive", "status": "static" }, "irqbalance.service": { "name": "irqbalance.service", "source": "systemd", "state": "running", "status": "enabled" }, "iscsi-shutdown.service": { "name": "iscsi-shutdown.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "iscsi.service": { "name": "iscsi.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "iscsid.service": { "name": "iscsid.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "kdump.service": { "name": "kdump.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "kmod-static-nodes.service": { "name": "kmod-static-nodes.service", "source": "systemd", "state": "stopped", "status": "static" }, "kvm_stat.service": { "name": "kvm_stat.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "ldconfig.service": { "name": "ldconfig.service", "source": "systemd", "state": "stopped", "status": "static" }, "loadmodules.service": { "name": "loadmodules.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "lvm2-activation-early.service": { "name": "lvm2-activation-early.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "lvm2-activation.service": { "name": "lvm2-activation.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "lvm2-lvmpolld.service": { "name": "lvm2-lvmpolld.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-monitor.service": { "name": "lvm2-monitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "lvm2-pvscan@.service": { "name": "lvm2-pvscan@.service", "source": "systemd", "state": "unknown", "status": "static" }, "lvm2-pvscan@8:0.service": { "name": "lvm2-pvscan@8:0.service", "source": "systemd", "state": "stopped", "status": "active" }, "man-db-cache-update.service": { "name": "man-db-cache-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "man-db-restart-cache-update.service": { "name": "man-db-restart-cache-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "mdadm-grow-continue@.service": { "name": "mdadm-grow-continue@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdadm-last-resort@.service": { "name": "mdadm-last-resort@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdcheck_continue.service": { "name": "mdcheck_continue.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdcheck_start.service": { "name": "mdcheck_start.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmon@.service": { "name": "mdmon@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdmonitor-oneshot.service": { "name": "mdmonitor-oneshot.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmonitor.service": { "name": "mdmonitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "messagebus.service": { "name": "messagebus.service", "source": "systemd", "state": "active", "status": "static" }, "microcode.service": { "name": "microcode.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "multipathd.service": { "name": "multipathd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "ndctl-monitor.service": { "name": "ndctl-monitor.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "network.service": { "name": "network.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "nfs-blkmap.service": { "name": "nfs-blkmap.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-convert.service": { "name": "nfs-convert.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-idmapd.service": { "name": "nfs-idmapd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-mountd.service": { "name": "nfs-mountd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-server.service": { "name": "nfs-server.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-utils.service": { "name": "nfs-utils.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfsdcld.service": { "name": "nfsdcld.service", "source": "systemd", "state": "stopped", "status": "static" }, "nis-domainname.service": { "name": "nis-domainname.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "ntpd.service": { "name": "ntpd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ntpdate.service": { "name": "ntpdate.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "oddjobd.service": { "name": "oddjobd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "packagekit-offline-update.service": { "name": "packagekit-offline-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "packagekit.service": { "name": "packagekit.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-quit-wait.service": { "name": "plymouth-quit-wait.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "plymouth-start.service": { "name": "plymouth-start.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "polkit.service": { "name": "polkit.service", "source": "systemd", "state": "running", "status": "static" }, "qemu-guest-agent.service": { "name": "qemu-guest-agent.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "quotaon.service": { "name": "quotaon.service", "source": "systemd", "state": "inactive", "status": "static" }, "rbdmap.service": { "name": "rbdmap.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "rc-local.service": { "name": "rc-local.service", "source": "systemd", "state": "stopped", "status": "static" }, "rdisc.service": { "name": "rdisc.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rescue.service": { "name": "rescue.service", "source": "systemd", "state": "stopped", "status": "static" }, "rhcd.service": { "name": "rhcd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhsm-facts.service": { "name": "rhsm-facts.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhsm.service": { "name": "rhsm.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhsmcertd.service": { "name": "rhsmcertd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpc-gssd.service": { "name": "rpc-gssd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd-notify.service": { "name": "rpc-statd-notify.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd.service": { "name": "rpc-statd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpcbind.service": { "name": "rpcbind.service", "source": "systemd", "state": "running", "status": "enabled" }, "rsyslog.service": { "name": "rsyslog.service", "source": "systemd", "state": "running", "status": "enabled" }, "selinux-autorelabel-mark.service": { "name": "selinux-autorelabel-mark.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "selinux-autorelabel.service": { "name": "selinux-autorelabel.service", "source": "systemd", "state": "inactive", "status": "static" }, "serial-getty@.service": { "name": "serial-getty@.service", "source": "systemd", "state": "unknown", "status": "indirect" }, "serial-getty@ttyS0.service": { "name": "serial-getty@ttyS0.service", "source": "systemd", "state": "running", "status": "active" }, "snapd.seeded.service": { "name": "snapd.seeded.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sntp.service": { "name": "sntp.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sshd-keygen.service": { "name": "sshd-keygen.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sshd-keygen@.service": { "name": "sshd-keygen@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@ecdsa.service": { "name": "sshd-keygen@ecdsa.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd-keygen@ed25519.service": { "name": "sshd-keygen@ed25519.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd-keygen@rsa.service": { "name": "sshd-keygen@rsa.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd.service": { "name": "sshd.service", "source": "systemd", "state": "running", "status": "enabled" }, "sshd@.service": { "name": "sshd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "sssd-autofs.service": { "name": "sssd-autofs.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-kcm.service": { "name": "sssd-kcm.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "sssd-nss.service": { "name": "sssd-nss.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pac.service": { "name": "sssd-pac.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pam.service": { "name": "sssd-pam.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-ssh.service": { "name": "sssd-ssh.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-sudo.service": { "name": "sssd-sudo.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd.service": { "name": "sssd.service", "source": "systemd", "state": "running", "status": "enabled" }, "syslog.service": { "name": "syslog.service", "source": "systemd", "state": "active", "status": "enabled" }, "system-update-cleanup.service": { "name": "system-update-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-ask-password-console.service": { "name": "systemd-ask-password-console.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-wall.service": { "name": "systemd-ask-password-wall.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-backlight@.service": { "name": "systemd-backlight@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-binfmt.service": { "name": "systemd-binfmt.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-coredump@.service": { "name": "systemd-coredump@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-exit.service": { "name": "systemd-exit.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-firstboot.service": { "name": "systemd-firstboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck-root.service": { "name": "systemd-fsck-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck@.service": { "name": "systemd-fsck@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-fsck@dev-disk-by\\x2duuid-7B77\\x2d95E7.service": { "name": "systemd-fsck@dev-disk-by\\x2duuid-7B77\\x2d95E7.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-fsck@dev-vdb1.service": { "name": "systemd-fsck@dev-vdb1.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-fsck@dev-vdc1.service": { "name": "systemd-fsck@dev-vdc1.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-halt.service": { "name": "systemd-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hibernate-resume@.service": { "name": "systemd-hibernate-resume@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-hibernate.service": { "name": "systemd-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hostnamed.service": { "name": "systemd-hostnamed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hwdb-update.service": { "name": "systemd-hwdb-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-hybrid-sleep.service": { "name": "systemd-hybrid-sleep.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-initctl.service": { "name": "systemd-initctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-catalog-update.service": { "name": "systemd-journal-catalog-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-flush.service": { "name": "systemd-journal-flush.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journald.service": { "name": "systemd-journald.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-kexec.service": { "name": "systemd-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-localed.service": { "name": "systemd-localed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-logind.service": { "name": "systemd-logind.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-machine-id-commit.service": { "name": "systemd-machine-id-commit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-modules-load.service": { "name": "systemd-modules-load.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-networkd-wait-online.service": { "name": "systemd-networkd-wait-online.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "systemd-portabled.service": { "name": "systemd-portabled.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-poweroff.service": { "name": "systemd-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-quotacheck.service": { "name": "systemd-quotacheck.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-random-seed.service": { "name": "systemd-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-reboot.service": { "name": "systemd-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-remount-fs.service": { "name": "systemd-remount-fs.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-resolved.service": { "name": "systemd-resolved.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-rfkill.service": { "name": "systemd-rfkill.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend-then-hibernate.service": { "name": "systemd-suspend-then-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend.service": { "name": "systemd-suspend.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-sysctl.service": { "name": "systemd-sysctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-sysusers.service": { "name": "systemd-sysusers.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-timedated.service": { "name": "systemd-timedated.service", "source": "systemd", "state": "inactive", "status": "masked" }, "systemd-timesyncd.service": { "name": "systemd-timesyncd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "systemd-tmpfiles-clean.service": { "name": "systemd-tmpfiles-clean.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev.service": { "name": "systemd-tmpfiles-setup-dev.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup.service": { "name": "systemd-tmpfiles-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-settle.service": { "name": "systemd-udev-settle.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-trigger.service": { "name": "systemd-udev-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udevd.service": { "name": "systemd-udevd.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-update-done.service": { "name": "systemd-update-done.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp-runlevel.service": { "name": "systemd-update-utmp-runlevel.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp.service": { "name": "systemd-update-utmp.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-user-sessions.service": { "name": "systemd-user-sessions.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-vconsole-setup.service": { "name": "systemd-vconsole-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-volatile-root.service": { "name": "systemd-volatile-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "tcsd.service": { "name": "tcsd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "teamd@.service": { "name": "teamd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "timedatex.service": { "name": "timedatex.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "tuned.service": { "name": "tuned.service", "source": "systemd", "state": "running", "status": "enabled" }, "unbound-anchor.service": { "name": "unbound-anchor.service", "source": "systemd", "state": "stopped", "status": "static" }, "user-runtime-dir@.service": { "name": "user-runtime-dir@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user-runtime-dir@0.service": { "name": "user-runtime-dir@0.service", "source": "systemd", "state": "stopped", "status": "active" }, "user@.service": { "name": "user@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user@0.service": { "name": "user@0.service", "source": "systemd", "state": "running", "status": "active" }, "vdo-start-by-dev@.service": { "name": "vdo-start-by-dev@.service", "source": "systemd", "state": "unknown", "status": "static" }, "vdo.service": { "name": "vdo.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "ypbind.service": { "name": "ypbind.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "yppasswdd.service": { "name": "yppasswdd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ypserv.service": { "name": "ypserv.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ypxfrd.service": { "name": "ypxfrd.service", "source": "systemd", "state": "stopped", "status": "not-found" } } }, "changed": false } TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] ***** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:53 Thursday 21 July 2022 19:06:21 +0000 (0:00:02.771) 0:01:38.638 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:58 Thursday 21 July 2022 19:06:21 +0000 (0:00:00.061) 0:01:38.700 ********* TASK [fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64 Thursday 21 July 2022 19:06:21 +0000 (0:00:00.023) 0:01:38.724 ********* changed: [/cache/rhel-8-y.qcow2.snap] => { "actions": [ { "action": "resize format", "device": "/dev/mapper/foo-test2", "fs_type": "ext4" }, { "action": "resize device", "device": "/dev/mapper/foo-test2", "fs_type": null } ], "changed": true, "crypts": [], "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/vda2", "/dev/vda3", "/dev/mapper/foo-test1", "/dev/mapper/foo-test2", "/dev/sdb", "/dev/sdc", "/dev/nvme0n1", "/dev/nvme1n1", "/dev/nvme2n1", "/dev/vdb1", "/dev/vdc1", "/dev/vdd", "/dev/vde", "/dev/vdf" ], "mounts": [ { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" }, { "dump": 0, "fstype": "ext4", "opts": "defaults", "passno": 0, "path": "/opt/test2", "src": "/dev/mapper/foo-test2", "state": "mounted" } ], "packages": [ "lvm2", "dosfstools", "e2fsprogs", "xfsprogs" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "60%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test2", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test2", "_raw_device": "/dev/mapper/foo-test2", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "25%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:78 Thursday 21 July 2022 19:06:23 +0000 (0:00:02.431) 0:01:41.156 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90 Thursday 21 July 2022 19:06:23 +0000 (0:00:00.039) 0:01:41.195 ********* TASK [fedora.linux_system_roles.storage : show blivet_output] ****************** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:96 Thursday 21 July 2022 19:06:23 +0000 (0:00:00.025) 0:01:41.221 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "blivet_output": { "actions": [ { "action": "resize format", "device": "/dev/mapper/foo-test2", "fs_type": "ext4" }, { "action": "resize device", "device": "/dev/mapper/foo-test2", "fs_type": null } ], "changed": true, "crypts": [], "failed": false, "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/vda2", "/dev/vda3", "/dev/mapper/foo-test1", "/dev/mapper/foo-test2", "/dev/sdb", "/dev/sdc", "/dev/nvme0n1", "/dev/nvme1n1", "/dev/nvme2n1", "/dev/vdb1", "/dev/vdc1", "/dev/vdd", "/dev/vde", "/dev/vdf" ], "mounts": [ { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" }, { "dump": 0, "fstype": "ext4", "opts": "defaults", "passno": 0, "path": "/opt/test2", "src": "/dev/mapper/foo-test2", "state": "mounted" } ], "packages": [ "lvm2", "dosfstools", "e2fsprogs", "xfsprogs" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "60%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test2", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test2", "_raw_device": "/dev/mapper/foo-test2", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "25%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } } TASK [fedora.linux_system_roles.storage : set the list of pools for test verification] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:101 Thursday 21 July 2022 19:06:23 +0000 (0:00:00.084) 0:01:41.305 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "60%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test2", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test2", "_raw_device": "/dev/mapper/foo-test2", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "25%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] }, "changed": false } TASK [fedora.linux_system_roles.storage : set the list of volumes for test verification] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:105 Thursday 21 July 2022 19:06:23 +0000 (0:00:00.044) 0:01:41.349 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : remove obsolete mounts] ************** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:121 Thursday 21 July 2022 19:06:23 +0000 (0:00:00.036) 0:01:41.386 ********* TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132 Thursday 21 July 2022 19:06:23 +0000 (0:00:00.037) 0:01:41.424 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.storage : set up new/current mounts] *********** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:137 Thursday 21 July 2022 19:06:24 +0000 (0:00:00.666) 0:01:42.090 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount ok: [/cache/rhel-8-y.qcow2.snap] => (item={'src': '/dev/mapper/foo-test1', 'path': '/opt/test1', 'fstype': 'xfs', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted'}) => { "ansible_loop_var": "mount_info", "backup_file": "", "boot": "yes", "changed": false, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" }, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test1" } redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount changed: [/cache/rhel-8-y.qcow2.snap] => (item={'src': '/dev/mapper/foo-test2', 'path': '/opt/test2', 'fstype': 'ext4', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted'}) => { "ansible_loop_var": "mount_info", "backup_file": "", "boot": "yes", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "ext4", "mount_info": { "dump": 0, "fstype": "ext4", "opts": "defaults", "passno": 0, "path": "/opt/test2", "src": "/dev/mapper/foo-test2", "state": "mounted" }, "name": "/opt/test2", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test2" } TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148 Thursday 21 July 2022 19:06:25 +0000 (0:00:00.777) 0:01:42.868 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.storage : retrieve facts for the /etc/crypttab file] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:156 Thursday 21 July 2022 19:06:25 +0000 (0:00:00.674) 0:01:43.542 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "stat": { "exists": false } } TASK [fedora.linux_system_roles.storage : manage /etc/crypttab to account for changes we just made] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:161 Thursday 21 July 2022 19:06:26 +0000 (0:00:00.378) 0:01:43.921 ********* TASK [fedora.linux_system_roles.storage : Update facts] ************************ task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:183 Thursday 21 July 2022 19:06:26 +0000 (0:00:00.025) 0:01:43.946 ********* ok: [/cache/rhel-8-y.qcow2.snap] META: role_complete for /cache/rhel-8-y.qcow2.snap TASK [include_tasks] *********************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/tests_lvm_percent_size.yml:112 Thursday 21 July 2022 19:06:27 +0000 (0:00:01.002) 0:01:44.948 ********* included: /tmp/tmpmnqeja9w/tests/storage/verify-role-results.yml for /cache/rhel-8-y.qcow2.snap TASK [Print out pool information] ********************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-role-results.yml:1 Thursday 21 July 2022 19:06:27 +0000 (0:00:00.045) 0:01:44.994 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "60%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test2", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test2", "_raw_device": "/dev/mapper/foo-test2", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "25%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] } TASK [Print out volume information] ******************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-role-results.yml:6 Thursday 21 July 2022 19:06:27 +0000 (0:00:00.054) 0:01:45.049 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => {} TASK [Collect info about the volumes.] ***************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-role-results.yml:14 Thursday 21 July 2022 19:06:27 +0000 (0:00:00.038) 0:01:45.087 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "info": { "/dev/mapper/foo-test1": { "fstype": "xfs", "label": "", "name": "/dev/mapper/foo-test1", "size": "6G", "type": "lvm", "uuid": "d739e41d-c1a1-4003-b357-25bf62753af5" }, "/dev/mapper/foo-test2": { "fstype": "ext4", "label": "", "name": "/dev/mapper/foo-test2", "size": "2.5G", "type": "lvm", "uuid": "64454f89-f2cd-4a8b-958f-4c900f72cf79" }, "/dev/nvme0n1": { "fstype": "", "label": "", "name": "/dev/nvme0n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme1n1": { "fstype": "", "label": "", "name": "/dev/nvme1n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme2n1": { "fstype": "", "label": "", "name": "/dev/nvme2n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sda": { "fstype": "LVM2_member", "label": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "HA17fj-P3dB-xyZl-9Zru-ZzXr-ULh6-93Sgui" }, "/dev/sdb": { "fstype": "", "label": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc": { "fstype": "", "label": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sr0": { "fstype": "iso9660", "label": "cidata", "name": "/dev/sr0", "size": "364K", "type": "rom", "uuid": "2022-07-21-19-04-29-00" }, "/dev/vda": { "fstype": "", "label": "", "name": "/dev/vda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vda1": { "fstype": "", "label": "", "name": "/dev/vda1", "size": "1M", "type": "partition", "uuid": "" }, "/dev/vda2": { "fstype": "vfat", "label": "", "name": "/dev/vda2", "size": "100M", "type": "partition", "uuid": "7B77-95E7" }, "/dev/vda3": { "fstype": "xfs", "label": "root", "name": "/dev/vda3", "size": "9.9G", "type": "partition", "uuid": "58dca2db-4ccd-4171-a017-dd53ad8c1c54" }, "/dev/vdb": { "fstype": "", "label": "", "name": "/dev/vdb", "size": "2G", "type": "disk", "uuid": "" }, "/dev/vdb1": { "fstype": "ext4", "label": "yumcache", "name": "/dev/vdb1", "size": "2G", "type": "partition", "uuid": "4d1ed9a1-5608-4ff8-ad89-21d95caa66fa" }, "/dev/vdc": { "fstype": "", "label": "", "name": "/dev/vdc", "size": "2G", "type": "disk", "uuid": "" }, "/dev/vdc1": { "fstype": "ext4", "label": "yumvarlib", "name": "/dev/vdc1", "size": "2G", "type": "partition", "uuid": "a7abcc97-edc5-476c-8ccd-694b4621fd28" }, "/dev/vdd": { "fstype": "", "label": "", "name": "/dev/vdd", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vde": { "fstype": "", "label": "", "name": "/dev/vde", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vdf": { "fstype": "", "label": "", "name": "/dev/vdf", "size": "10G", "type": "disk", "uuid": "" } } } TASK [Read the /etc/fstab file for volume existence] *************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-role-results.yml:19 Thursday 21 July 2022 19:06:27 +0000 (0:00:00.385) 0:01:45.472 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "cmd": [ "cat", "/etc/fstab" ], "delta": "0:00:00.002884", "end": "2022-07-21 15:06:28.377872", "rc": 0, "start": "2022-07-21 15:06:28.374988" } STDOUT: UUID=58dca2db-4ccd-4171-a017-dd53ad8c1c54 / xfs defaults 0 0 UUID=7B77-95E7 /boot/efi vfat defaults,uid=0,gid=0,umask=077,shortname=winnt 0 2 /dev/vdb1 /var/cache/dnf auto defaults,nofail,x-systemd.requires=cloud-init.service,_netdev,comment=cloudconfig 0 2 /dev/vdc1 /var/lib/dnf auto defaults,nofail,x-systemd.requires=cloud-init.service,_netdev,comment=cloudconfig 0 2 /dev/mapper/foo-test1 /opt/test1 xfs defaults 0 0 /dev/mapper/foo-test2 /opt/test2 ext4 defaults 0 0 TASK [Read the /etc/crypttab file] ********************************************* task path: /tmp/tmpmnqeja9w/tests/storage/verify-role-results.yml:24 Thursday 21 July 2022 19:06:28 +0000 (0:00:00.386) 0:01:45.858 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "cmd": [ "cat", "/etc/crypttab" ], "delta": "0:00:01.005563", "end": "2022-07-21 15:06:29.763005", "failed_when_result": false, "rc": 1, "start": "2022-07-21 15:06:28.757442" } STDERR: cat: /etc/crypttab: No such file or directory MSG: non-zero return code TASK [Verify the volumes listed in storage_pools were correctly managed] ******* task path: /tmp/tmpmnqeja9w/tests/storage/verify-role-results.yml:33 Thursday 21 July 2022 19:06:29 +0000 (0:00:01.390) 0:01:47.249 ********* included: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool.yml for /cache/rhel-8-y.qcow2.snap => (item={'disks': ['sda'], 'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'name': 'foo', 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_metadata_version': None, 'state': 'present', 'type': 'lvm', 'volumes': [{'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'name': 'test1', 'raid_level': None, 'size': '60%', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test1', '_raw_device': '/dev/mapper/foo-test1', '_mount_id': '/dev/mapper/foo-test1', '_kernel_device': '/dev/dm-1', '_raw_kernel_device': '/dev/dm-1'}, {'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'ext4', 'mount_options': 'defaults', 'mount_point': '/opt/test2', 'name': 'test2', 'raid_level': None, 'size': '25%', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test2', '_raw_device': '/dev/mapper/foo-test2', '_mount_id': '/dev/mapper/foo-test2', '_kernel_device': '/dev/dm-0', '_raw_kernel_device': '/dev/dm-0'}], 'raid_chunk_size': None}) TASK [Set _storage_pool_tests] ************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool.yml:5 Thursday 21 July 2022 19:06:29 +0000 (0:00:00.061) 0:01:47.311 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_pool_tests": [ "members", "volumes" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool.yml:18 Thursday 21 July 2022 19:06:29 +0000 (0:00:00.067) 0:01:47.378 ********* included: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml for /cache/rhel-8-y.qcow2.snap => (item=members) included: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-volumes.yml for /cache/rhel-8-y.qcow2.snap => (item=volumes) TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:1 Thursday 21 July 2022 19:06:29 +0000 (0:00:00.047) 0:01:47.426 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_pv_count": "1", "_storage_test_pool_pvs_lvm": [ "/dev/sda" ] }, "changed": false } TASK [Get the canonical device path for each member device] ******************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:6 Thursday 21 July 2022 19:06:29 +0000 (0:00:00.094) 0:01:47.521 ********* ok: [/cache/rhel-8-y.qcow2.snap] => (item=/dev/sda) => { "ansible_loop_var": "pv", "changed": false, "device": "/dev/sda", "pv": "/dev/sda" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:15 Thursday 21 July 2022 19:06:30 +0000 (0:00:00.388) 0:01:47.909 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "__pvs_lvm_len": "1" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:19 Thursday 21 July 2022 19:06:30 +0000 (0:00:00.050) 0:01:47.959 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_pool_pvs": [ "/dev/sda" ] }, "changed": false } TASK [Verify PV count] ********************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:23 Thursday 21 July 2022 19:06:30 +0000 (0:00:00.054) 0:01:48.014 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:29 Thursday 21 July 2022 19:06:30 +0000 (0:00:00.052) 0:01:48.067 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:33 Thursday 21 July 2022 19:06:30 +0000 (0:00:00.037) 0:01:48.104 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:37 Thursday 21 July 2022 19:06:30 +0000 (0:00:00.050) 0:01:48.154 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check the type of each PV] *********************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:41 Thursday 21 July 2022 19:06:30 +0000 (0:00:00.028) 0:01:48.183 ********* ok: [/cache/rhel-8-y.qcow2.snap] => (item=/dev/sda) => { "ansible_loop_var": "pv", "changed": false, "pv": "/dev/sda" } MSG: All assertions passed TASK [Check MD RAID] *********************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:50 Thursday 21 July 2022 19:06:30 +0000 (0:00:00.043) 0:01:48.227 ********* included: /tmp/tmpmnqeja9w/tests/storage/verify-pool-md.yml for /cache/rhel-8-y.qcow2.snap TASK [get information about RAID] ********************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-md.yml:6 Thursday 21 July 2022 19:06:30 +0000 (0:00:00.047) 0:01:48.274 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-md.yml:12 Thursday 21 July 2022 19:06:30 +0000 (0:00:00.031) 0:01:48.305 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-md.yml:16 Thursday 21 July 2022 19:06:30 +0000 (0:00:00.025) 0:01:48.331 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-md.yml:20 Thursday 21 July 2022 19:06:30 +0000 (0:00:00.025) 0:01:48.356 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-md.yml:24 Thursday 21 July 2022 19:06:30 +0000 (0:00:00.024) 0:01:48.380 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-md.yml:30 Thursday 21 July 2022 19:06:30 +0000 (0:00:00.025) 0:01:48.405 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-md.yml:36 Thursday 21 July 2022 19:06:30 +0000 (0:00:00.023) 0:01:48.429 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-md.yml:44 Thursday 21 July 2022 19:06:30 +0000 (0:00:00.025) 0:01:48.455 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_md_active_devices_re": null, "storage_test_md_metadata_version_re": null, "storage_test_md_spare_devices_re": null }, "changed": false } TASK [Check LVM RAID] ********************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:53 Thursday 21 July 2022 19:06:30 +0000 (0:00:00.040) 0:01:48.495 ********* included: /tmp/tmpmnqeja9w/tests/storage/verify-pool-members-lvmraid.yml for /cache/rhel-8-y.qcow2.snap TASK [Validate pool member LVM RAID settings] ********************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-members-lvmraid.yml:1 Thursday 21 July 2022 19:06:30 +0000 (0:00:00.048) 0:01:48.544 ********* included: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-lvmraid.yml for /cache/rhel-8-y.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'name': 'test1', 'raid_level': None, 'size': '60%', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test1', '_raw_device': '/dev/mapper/foo-test1', '_mount_id': '/dev/mapper/foo-test1', '_kernel_device': '/dev/dm-1', '_raw_kernel_device': '/dev/dm-1'}) included: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-lvmraid.yml for /cache/rhel-8-y.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'ext4', 'mount_options': 'defaults', 'mount_point': '/opt/test2', 'name': 'test2', 'raid_level': None, 'size': '25%', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test2', '_raw_device': '/dev/mapper/foo-test2', '_mount_id': '/dev/mapper/foo-test2', '_kernel_device': '/dev/dm-0', '_raw_kernel_device': '/dev/dm-0'}) TASK [Get information about LVM RAID] ****************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-lvmraid.yml:3 Thursday 21 July 2022 19:06:31 +0000 (0:00:00.053) 0:01:48.598 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is LVM RAID] ******************************************* task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-lvmraid.yml:8 Thursday 21 July 2022 19:06:31 +0000 (0:00:00.029) 0:01:48.627 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-lvmraid.yml:12 Thursday 21 July 2022 19:06:31 +0000 (0:00:00.027) 0:01:48.655 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about LVM RAID] ****************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-lvmraid.yml:3 Thursday 21 July 2022 19:06:31 +0000 (0:00:00.027) 0:01:48.682 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is LVM RAID] ******************************************* task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-lvmraid.yml:8 Thursday 21 July 2022 19:06:31 +0000 (0:00:00.028) 0:01:48.710 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-lvmraid.yml:12 Thursday 21 July 2022 19:06:31 +0000 (0:00:00.026) 0:01:48.737 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check Thin Pools] ******************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:56 Thursday 21 July 2022 19:06:31 +0000 (0:00:00.027) 0:01:48.764 ********* included: /tmp/tmpmnqeja9w/tests/storage/verify-pool-members-thin.yml for /cache/rhel-8-y.qcow2.snap TASK [Validate pool member thinpool settings] ********************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-members-thin.yml:1 Thursday 21 July 2022 19:06:31 +0000 (0:00:00.046) 0:01:48.810 ********* included: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-thin.yml for /cache/rhel-8-y.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'name': 'test1', 'raid_level': None, 'size': '60%', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test1', '_raw_device': '/dev/mapper/foo-test1', '_mount_id': '/dev/mapper/foo-test1', '_kernel_device': '/dev/dm-1', '_raw_kernel_device': '/dev/dm-1'}) included: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-thin.yml for /cache/rhel-8-y.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'ext4', 'mount_options': 'defaults', 'mount_point': '/opt/test2', 'name': 'test2', 'raid_level': None, 'size': '25%', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test2', '_raw_device': '/dev/mapper/foo-test2', '_mount_id': '/dev/mapper/foo-test2', '_kernel_device': '/dev/dm-0', '_raw_kernel_device': '/dev/dm-0'}) TASK [Get information about thinpool] ****************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-thin.yml:3 Thursday 21 July 2022 19:06:31 +0000 (0:00:00.049) 0:01:48.860 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-thin.yml:8 Thursday 21 July 2022 19:06:31 +0000 (0:00:00.024) 0:01:48.885 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-thin.yml:13 Thursday 21 July 2022 19:06:31 +0000 (0:00:00.023) 0:01:48.909 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-thin.yml:17 Thursday 21 July 2022 19:06:31 +0000 (0:00:00.023) 0:01:48.933 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about thinpool] ****************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-thin.yml:3 Thursday 21 July 2022 19:06:31 +0000 (0:00:00.023) 0:01:48.957 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-thin.yml:8 Thursday 21 July 2022 19:06:31 +0000 (0:00:00.024) 0:01:48.982 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-thin.yml:13 Thursday 21 July 2022 19:06:31 +0000 (0:00:00.023) 0:01:49.005 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-thin.yml:17 Thursday 21 July 2022 19:06:31 +0000 (0:00:00.023) 0:01:49.028 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check member encryption] ************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:59 Thursday 21 July 2022 19:06:31 +0000 (0:00:00.025) 0:01:49.054 ********* included: /tmp/tmpmnqeja9w/tests/storage/verify-pool-members-encryption.yml for /cache/rhel-8-y.qcow2.snap TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-members-encryption.yml:4 Thursday 21 July 2022 19:06:31 +0000 (0:00:00.083) 0:01:49.137 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Validate pool member LUKS settings] ************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-members-encryption.yml:8 Thursday 21 July 2022 19:06:31 +0000 (0:00:00.053) 0:01:49.190 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => (item=/dev/sda) => { "_storage_test_pool_member_path": "/dev/sda", "ansible_loop_var": "_storage_test_pool_member_path", "changed": false, "skip_reason": "Conditional result was False" } TASK [Validate pool member crypttab entries] *********************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-members-encryption.yml:15 Thursday 21 July 2022 19:06:31 +0000 (0:00:00.029) 0:01:49.220 ********* included: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-crypttab.yml for /cache/rhel-8-y.qcow2.snap => (item=/dev/sda) TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-crypttab.yml:1 Thursday 21 July 2022 19:06:31 +0000 (0:00:00.049) 0:01:49.270 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": [] }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-crypttab.yml:4 Thursday 21 July 2022 19:06:31 +0000 (0:00:00.051) 0:01:49.321 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-crypttab.yml:9 Thursday 21 July 2022 19:06:31 +0000 (0:00:00.052) 0:01:49.374 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-crypttab.yml:15 Thursday 21 July 2022 19:06:31 +0000 (0:00:00.039) 0:01:49.413 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-crypttab.yml:21 Thursday 21 July 2022 19:06:31 +0000 (0:00:00.036) 0:01:49.450 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-crypttab.yml:27 Thursday 21 July 2022 19:06:31 +0000 (0:00:00.035) 0:01:49.485 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-members-encryption.yml:22 Thursday 21 July 2022 19:06:31 +0000 (0:00:00.040) 0:01:49.526 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_crypttab_key_file": null }, "changed": false } TASK [Check VDO] *************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:62 Thursday 21 July 2022 19:06:32 +0000 (0:00:00.033) 0:01:49.560 ********* included: /tmp/tmpmnqeja9w/tests/storage/verify-pool-members-vdo.yml for /cache/rhel-8-y.qcow2.snap TASK [Validate pool member VDO settings] *************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-members-vdo.yml:1 Thursday 21 July 2022 19:06:32 +0000 (0:00:00.048) 0:01:49.608 ********* included: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-vdo.yml for /cache/rhel-8-y.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'name': 'test1', 'raid_level': None, 'size': '60%', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test1', '_raw_device': '/dev/mapper/foo-test1', '_mount_id': '/dev/mapper/foo-test1', '_kernel_device': '/dev/dm-1', '_raw_kernel_device': '/dev/dm-1'}) included: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-vdo.yml for /cache/rhel-8-y.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'ext4', 'mount_options': 'defaults', 'mount_point': '/opt/test2', 'name': 'test2', 'raid_level': None, 'size': '25%', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test2', '_raw_device': '/dev/mapper/foo-test2', '_mount_id': '/dev/mapper/foo-test2', '_kernel_device': '/dev/dm-0', '_raw_kernel_device': '/dev/dm-0'}) TASK [get information about VDO deduplication] ********************************* task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-vdo.yml:3 Thursday 21 July 2022 19:06:32 +0000 (0:00:00.054) 0:01:49.663 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-vdo.yml:8 Thursday 21 July 2022 19:06:32 +0000 (0:00:00.024) 0:01:49.687 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-vdo.yml:11 Thursday 21 July 2022 19:06:32 +0000 (0:00:00.024) 0:01:49.712 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-vdo.yml:16 Thursday 21 July 2022 19:06:32 +0000 (0:00:00.024) 0:01:49.736 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-vdo.yml:21 Thursday 21 July 2022 19:06:32 +0000 (0:00:00.024) 0:01:49.760 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-vdo.yml:24 Thursday 21 July 2022 19:06:32 +0000 (0:00:00.023) 0:01:49.784 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-vdo.yml:29 Thursday 21 July 2022 19:06:32 +0000 (0:00:00.026) 0:01:49.810 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-vdo.yml:39 Thursday 21 July 2022 19:06:32 +0000 (0:00:00.024) 0:01:49.835 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [get information about VDO deduplication] ********************************* task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-vdo.yml:3 Thursday 21 July 2022 19:06:32 +0000 (0:00:00.036) 0:01:49.871 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-vdo.yml:8 Thursday 21 July 2022 19:06:32 +0000 (0:00:00.026) 0:01:49.897 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-vdo.yml:11 Thursday 21 July 2022 19:06:32 +0000 (0:00:00.027) 0:01:49.925 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-vdo.yml:16 Thursday 21 July 2022 19:06:32 +0000 (0:00:00.027) 0:01:49.952 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-vdo.yml:21 Thursday 21 July 2022 19:06:32 +0000 (0:00:00.025) 0:01:49.978 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-vdo.yml:24 Thursday 21 July 2022 19:06:32 +0000 (0:00:00.023) 0:01:50.002 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-vdo.yml:29 Thursday 21 July 2022 19:06:32 +0000 (0:00:00.026) 0:01:50.028 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-vdo.yml:39 Thursday 21 July 2022 19:06:32 +0000 (0:00:00.024) 0:01:50.053 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:65 Thursday 21 July 2022 19:06:32 +0000 (0:00:00.036) 0:01:50.089 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "__pvs_lvm_len": null, "_storage_test_expected_pv_count": null, "_storage_test_expected_pv_type": null, "_storage_test_pool_pvs": [], "_storage_test_pool_pvs_lvm": [] }, "changed": false } TASK [verify the volumes] ****************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-volumes.yml:3 Thursday 21 July 2022 19:06:32 +0000 (0:00:00.036) 0:01:50.126 ********* included: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume.yml for /cache/rhel-8-y.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'name': 'test1', 'raid_level': None, 'size': '60%', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test1', '_raw_device': '/dev/mapper/foo-test1', '_mount_id': '/dev/mapper/foo-test1', '_kernel_device': '/dev/dm-1', '_raw_kernel_device': '/dev/dm-1'}) included: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume.yml for /cache/rhel-8-y.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'ext4', 'mount_options': 'defaults', 'mount_point': '/opt/test2', 'name': 'test2', 'raid_level': None, 'size': '25%', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test2', '_raw_device': '/dev/mapper/foo-test2', '_mount_id': '/dev/mapper/foo-test2', '_kernel_device': '/dev/dm-0', '_raw_kernel_device': '/dev/dm-0'}) TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume.yml:2 Thursday 21 July 2022 19:06:32 +0000 (0:00:00.049) 0:01:50.175 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_volume_present": true, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume.yml:10 Thursday 21 July 2022 19:06:32 +0000 (0:00:00.049) 0:01:50.225 ********* included: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml for /cache/rhel-8-y.qcow2.snap => (item=mount) included: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-fstab.yml for /cache/rhel-8-y.qcow2.snap => (item=fstab) included: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-fs.yml for /cache/rhel-8-y.qcow2.snap => (item=fs) included: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-device.yml for /cache/rhel-8-y.qcow2.snap => (item=device) included: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml for /cache/rhel-8-y.qcow2.snap => (item=encryption) included: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-md.yml for /cache/rhel-8-y.qcow2.snap => (item=md) included: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml for /cache/rhel-8-y.qcow2.snap => (item=size) included: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-cache.yml for /cache/rhel-8-y.qcow2.snap => (item=cache) TASK [Get expected mount device based on device type] ************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml:6 Thursday 21 July 2022 19:06:32 +0000 (0:00:00.076) 0:01:50.301 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/foo-test1" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml:10 Thursday 21 July 2022 19:06:32 +0000 (0:00:00.103) 0:01:50.405 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_mount_device_matches": [ { "block_available": 1551090, "block_size": 4096, "block_total": 1570304, "block_used": 19214, "device": "/dev/mapper/foo-test1", "fstype": "xfs", "inode_available": 3145725, "inode_total": 3145728, "inode_used": 3, "mount": "/opt/test1", "options": "rw,seclabel,relatime,attr2,inode64,logbufs=8,logbsize=32k,noquota", "size_available": 6353264640, "size_total": 6431965184, "uuid": "d739e41d-c1a1-4003-b357-25bf62753af5" } ], "storage_test_mount_expected_match_count": "1", "storage_test_mount_point_matches": [ { "block_available": 1551090, "block_size": 4096, "block_total": 1570304, "block_used": 19214, "device": "/dev/mapper/foo-test1", "fstype": "xfs", "inode_available": 3145725, "inode_total": 3145728, "inode_used": 3, "mount": "/opt/test1", "options": "rw,seclabel,relatime,attr2,inode64,logbufs=8,logbsize=32k,noquota", "size_available": 6353264640, "size_total": 6431965184, "uuid": "d739e41d-c1a1-4003-b357-25bf62753af5" } ], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Verify the current mount state by device] ******************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml:20 Thursday 21 July 2022 19:06:32 +0000 (0:00:00.053) 0:01:50.459 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify the current mount state by mount point] *************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml:29 Thursday 21 July 2022 19:06:32 +0000 (0:00:00.063) 0:01:50.523 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify the mount fs type] ************************************************ task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml:37 Thursday 21 July 2022 19:06:33 +0000 (0:00:00.051) 0:01:50.574 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [command] ***************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml:46 Thursday 21 July 2022 19:06:33 +0000 (0:00:00.049) 0:01:50.623 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml:50 Thursday 21 July 2022 19:06:33 +0000 (0:00:00.023) 0:01:50.647 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml:55 Thursday 21 July 2022 19:06:33 +0000 (0:00:00.023) 0:01:50.671 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml:65 Thursday 21 July 2022 19:06:33 +0000 (0:00:00.023) 0:01:50.695 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-fstab.yml:2 Thursday 21 July 2022 19:06:33 +0000 (0:00:00.032) 0:01:50.727 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "1", "storage_test_fstab_expected_mount_options_matches": "1", "storage_test_fstab_expected_mount_point_matches": "1", "storage_test_fstab_id_matches": [ "/dev/mapper/foo-test1 " ], "storage_test_fstab_mount_options_matches": [ " /opt/test1 xfs defaults " ], "storage_test_fstab_mount_point_matches": [ " /opt/test1 " ] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-fstab.yml:12 Thursday 21 July 2022 19:06:33 +0000 (0:00:00.058) 0:01:50.786 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify the fstab mount point] ******************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-fstab.yml:19 Thursday 21 July 2022 19:06:33 +0000 (0:00:00.051) 0:01:50.837 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-fstab.yml:25 Thursday 21 July 2022 19:06:33 +0000 (0:00:00.053) 0:01:50.891 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-fstab.yml:34 Thursday 21 July 2022 19:06:33 +0000 (0:00:00.036) 0:01:50.928 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-fs.yml:4 Thursday 21 July 2022 19:06:33 +0000 (0:00:00.037) 0:01:50.966 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify fs label] ********************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-fs.yml:10 Thursday 21 July 2022 19:06:33 +0000 (0:00:00.044) 0:01:51.010 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [See whether the device node is present] ********************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-device.yml:4 Thursday 21 July 2022 19:06:33 +0000 (0:00:00.042) 0:01:51.053 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "stat": { "atime": 1658430322.6091166, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1658430322.6091166, "dev": 6, "device_type": 64769, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 60295, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/symlink", "mode": "0660", "mtime": 1658430322.6091166, "nlink": 1, "path": "/dev/mapper/foo-test1", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-device.yml:10 Thursday 21 July 2022 19:06:33 +0000 (0:00:00.377) 0:01:51.431 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about this volume] ********************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-device.yml:15 Thursday 21 July 2022 19:06:33 +0000 (0:00:00.042) 0:01:51.473 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [(1/2) Process volume type (set initial value)] *************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-device.yml:21 Thursday 21 July 2022 19:06:33 +0000 (0:00:00.041) 0:01:51.515 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [(2/2) Process volume type (get RAID value)] ****************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-device.yml:25 Thursday 21 July 2022 19:06:34 +0000 (0:00:00.036) 0:01:51.551 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-device.yml:30 Thursday 21 July 2022 19:06:34 +0000 (0:00:00.024) 0:01:51.576 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Stat the LUKS device, if encrypted] ************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:3 Thursday 21 July 2022 19:06:34 +0000 (0:00:00.039) 0:01:51.616 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:10 Thursday 21 July 2022 19:06:34 +0000 (0:00:00.027) 0:01:51.643 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [Collect LUKS info for this volume] *************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:15 Thursday 21 July 2022 19:06:35 +0000 (0:00:01.365) 0:01:53.009 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:21 Thursday 21 July 2022 19:06:35 +0000 (0:00:00.025) 0:01:53.034 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:27 Thursday 21 July 2022 19:06:35 +0000 (0:00:00.025) 0:01:53.060 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:33 Thursday 21 July 2022 19:06:35 +0000 (0:00:00.084) 0:01:53.144 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:39 Thursday 21 July 2022 19:06:35 +0000 (0:00:00.026) 0:01:53.171 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:44 Thursday 21 July 2022 19:06:35 +0000 (0:00:00.025) 0:01:53.197 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:50 Thursday 21 July 2022 19:06:35 +0000 (0:00:00.025) 0:01:53.222 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:56 Thursday 21 July 2022 19:06:35 +0000 (0:00:00.026) 0:01:53.248 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:62 Thursday 21 July 2022 19:06:35 +0000 (0:00:00.024) 0:01:53.273 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:67 Thursday 21 July 2022 19:06:35 +0000 (0:00:00.094) 0:01:53.368 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:72 Thursday 21 July 2022 19:06:35 +0000 (0:00:00.053) 0:01:53.421 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:78 Thursday 21 July 2022 19:06:35 +0000 (0:00:00.039) 0:01:53.460 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:84 Thursday 21 July 2022 19:06:35 +0000 (0:00:00.043) 0:01:53.503 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:90 Thursday 21 July 2022 19:06:35 +0000 (0:00:00.036) 0:01:53.540 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [get information about RAID] ********************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-md.yml:7 Thursday 21 July 2022 19:06:36 +0000 (0:00:00.037) 0:01:53.578 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-md.yml:13 Thursday 21 July 2022 19:06:36 +0000 (0:00:00.036) 0:01:53.614 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-md.yml:17 Thursday 21 July 2022 19:06:36 +0000 (0:00:00.040) 0:01:53.655 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-md.yml:21 Thursday 21 July 2022 19:06:36 +0000 (0:00:00.039) 0:01:53.694 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-md.yml:25 Thursday 21 July 2022 19:06:36 +0000 (0:00:00.037) 0:01:53.732 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-md.yml:31 Thursday 21 July 2022 19:06:36 +0000 (0:00:00.038) 0:01:53.771 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-md.yml:37 Thursday 21 July 2022 19:06:36 +0000 (0:00:00.042) 0:01:53.813 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the actual size of the volume] ************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:3 Thursday 21 July 2022 19:06:36 +0000 (0:00:00.036) 0:01:53.849 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "bytes": 6442450944, "changed": false, "lvm": "6g", "parted": "6GiB", "size": "6 GiB" } TASK [parse the requested size of the volume] ********************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:9 Thursday 21 July 2022 19:06:36 +0000 (0:00:00.370) 0:01:54.220 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Establish base value for expected size] ********************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:15 Thursday 21 July 2022 19:06:36 +0000 (0:00:00.037) 0:01:54.258 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:20 Thursday 21 July 2022 19:06:36 +0000 (0:00:00.039) 0:01:54.297 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_expected_size": "4294967296.0" } TASK [debug] ******************************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:25 Thursday 21 July 2022 19:06:36 +0000 (0:00:00.036) 0:01:54.334 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_pool": { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "60%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test2", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test2", "_raw_device": "/dev/mapper/foo-test2", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "25%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } } TASK [debug] ******************************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:28 Thursday 21 July 2022 19:06:36 +0000 (0:00:00.056) 0:01:54.390 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_blkinfo": { "changed": false, "failed": false, "info": { "/dev/mapper/foo-test1": { "fstype": "xfs", "label": "", "name": "/dev/mapper/foo-test1", "size": "6G", "type": "lvm", "uuid": "d739e41d-c1a1-4003-b357-25bf62753af5" }, "/dev/mapper/foo-test2": { "fstype": "ext4", "label": "", "name": "/dev/mapper/foo-test2", "size": "2.5G", "type": "lvm", "uuid": "64454f89-f2cd-4a8b-958f-4c900f72cf79" }, "/dev/nvme0n1": { "fstype": "", "label": "", "name": "/dev/nvme0n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme1n1": { "fstype": "", "label": "", "name": "/dev/nvme1n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme2n1": { "fstype": "", "label": "", "name": "/dev/nvme2n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sda": { "fstype": "LVM2_member", "label": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "HA17fj-P3dB-xyZl-9Zru-ZzXr-ULh6-93Sgui" }, "/dev/sdb": { "fstype": "", "label": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc": { "fstype": "", "label": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sr0": { "fstype": "iso9660", "label": "cidata", "name": "/dev/sr0", "size": "364K", "type": "rom", "uuid": "2022-07-21-19-04-29-00" }, "/dev/vda": { "fstype": "", "label": "", "name": "/dev/vda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vda1": { "fstype": "", "label": "", "name": "/dev/vda1", "size": "1M", "type": "partition", "uuid": "" }, "/dev/vda2": { "fstype": "vfat", "label": "", "name": "/dev/vda2", "size": "100M", "type": "partition", "uuid": "7B77-95E7" }, "/dev/vda3": { "fstype": "xfs", "label": "root", "name": "/dev/vda3", "size": "9.9G", "type": "partition", "uuid": "58dca2db-4ccd-4171-a017-dd53ad8c1c54" }, "/dev/vdb": { "fstype": "", "label": "", "name": "/dev/vdb", "size": "2G", "type": "disk", "uuid": "" }, "/dev/vdb1": { "fstype": "ext4", "label": "yumcache", "name": "/dev/vdb1", "size": "2G", "type": "partition", "uuid": "4d1ed9a1-5608-4ff8-ad89-21d95caa66fa" }, "/dev/vdc": { "fstype": "", "label": "", "name": "/dev/vdc", "size": "2G", "type": "disk", "uuid": "" }, "/dev/vdc1": { "fstype": "ext4", "label": "yumvarlib", "name": "/dev/vdc1", "size": "2G", "type": "partition", "uuid": "a7abcc97-edc5-476c-8ccd-694b4621fd28" }, "/dev/vdd": { "fstype": "", "label": "", "name": "/dev/vdd", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vde": { "fstype": "", "label": "", "name": "/dev/vde", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vdf": { "fstype": "", "label": "", "name": "/dev/vdf", "size": "10G", "type": "disk", "uuid": "" } } } } TASK [Get the size of parent/pool device] ************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:31 Thursday 21 July 2022 19:06:36 +0000 (0:00:00.060) 0:01:54.450 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "bytes": 10737418240, "changed": false, "lvm": "10g", "parted": "10GiB", "size": "10 GiB" } TASK [debug] ******************************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:36 Thursday 21 July 2022 19:06:37 +0000 (0:00:00.395) 0:01:54.846 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_pool_size": { "bytes": 10737418240, "changed": false, "failed": false, "lvm": "10g", "parted": "10GiB", "size": "10 GiB" } } TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:39 Thursday 21 July 2022 19:06:37 +0000 (0:00:00.058) 0:01:54.904 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_expected_size": "6442450944.0" }, "changed": false } TASK [debug] ******************************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:44 Thursday 21 July 2022 19:06:37 +0000 (0:00:00.058) 0:01:54.963 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_actual_size": { "bytes": 6442450944, "changed": false, "failed": false, "lvm": "6g", "parted": "6GiB", "size": "6 GiB" } } TASK [debug] ******************************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:47 Thursday 21 July 2022 19:06:37 +0000 (0:00:00.039) 0:01:55.003 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_expected_size": "6442450944.0" } TASK [assert] ****************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:50 Thursday 21 July 2022 19:06:37 +0000 (0:00:00.041) 0:01:55.044 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Get information about the LV] ******************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-cache.yml:6 Thursday 21 July 2022 19:06:37 +0000 (0:00:00.058) 0:01:55.103 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "cmd": [ "lvs", "--noheadings", "--nameprefixes", "--units=b", "--nosuffix", "--unquoted", "-o", "name,attr,cache_total_blocks,chunk_size,segtype", "foo/test1" ], "delta": "0:00:00.034984", "end": "2022-07-21 15:06:38.042890", "rc": 0, "start": "2022-07-21 15:06:38.007906" } STDOUT: LVM2_LV_NAME=test1 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-cache.yml:14 Thursday 21 July 2022 19:06:37 +0000 (0:00:00.425) 0:01:55.528 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_lv_segtype": [ "linear" ] }, "changed": false } TASK [check segment type] ****************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-cache.yml:17 Thursday 21 July 2022 19:06:38 +0000 (0:00:00.105) 0:01:55.634 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-cache.yml:22 Thursday 21 July 2022 19:06:38 +0000 (0:00:00.053) 0:01:55.688 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the requested cache size] ****************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-cache.yml:26 Thursday 21 July 2022 19:06:38 +0000 (0:00:00.039) 0:01:55.727 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-cache.yml:32 Thursday 21 July 2022 19:06:38 +0000 (0:00:00.041) 0:01:55.769 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-cache.yml:36 Thursday 21 July 2022 19:06:38 +0000 (0:00:00.082) 0:01:55.852 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume.yml:16 Thursday 21 July 2022 19:06:38 +0000 (0:00:00.040) 0:01:55.892 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume.yml:2 Thursday 21 July 2022 19:06:38 +0000 (0:00:00.038) 0:01:55.930 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_volume_present": true, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume.yml:10 Thursday 21 July 2022 19:06:38 +0000 (0:00:00.052) 0:01:55.983 ********* included: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml for /cache/rhel-8-y.qcow2.snap => (item=mount) included: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-fstab.yml for /cache/rhel-8-y.qcow2.snap => (item=fstab) included: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-fs.yml for /cache/rhel-8-y.qcow2.snap => (item=fs) included: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-device.yml for /cache/rhel-8-y.qcow2.snap => (item=device) included: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml for /cache/rhel-8-y.qcow2.snap => (item=encryption) included: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-md.yml for /cache/rhel-8-y.qcow2.snap => (item=md) included: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml for /cache/rhel-8-y.qcow2.snap => (item=size) included: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-cache.yml for /cache/rhel-8-y.qcow2.snap => (item=cache) TASK [Get expected mount device based on device type] ************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml:6 Thursday 21 July 2022 19:06:38 +0000 (0:00:00.080) 0:01:56.063 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/foo-test2" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml:10 Thursday 21 July 2022 19:06:38 +0000 (0:00:00.043) 0:01:56.107 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_mount_device_matches": [ { "block_available": 588749, "block_size": 4096, "block_total": 625618, "block_used": 36869, "device": "/dev/mapper/foo-test2", "fstype": "ext4", "inode_available": 163829, "inode_total": 163840, "inode_used": 11, "mount": "/opt/test2", "options": "rw,seclabel,relatime", "size_available": 2411515904, "size_total": 2562531328, "uuid": "64454f89-f2cd-4a8b-958f-4c900f72cf79" } ], "storage_test_mount_expected_match_count": "1", "storage_test_mount_point_matches": [ { "block_available": 588749, "block_size": 4096, "block_total": 625618, "block_used": 36869, "device": "/dev/mapper/foo-test2", "fstype": "ext4", "inode_available": 163829, "inode_total": 163840, "inode_used": 11, "mount": "/opt/test2", "options": "rw,seclabel,relatime", "size_available": 2411515904, "size_total": 2562531328, "uuid": "64454f89-f2cd-4a8b-958f-4c900f72cf79" } ], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Verify the current mount state by device] ******************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml:20 Thursday 21 July 2022 19:06:38 +0000 (0:00:00.058) 0:01:56.166 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify the current mount state by mount point] *************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml:29 Thursday 21 July 2022 19:06:38 +0000 (0:00:00.054) 0:01:56.220 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify the mount fs type] ************************************************ task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml:37 Thursday 21 July 2022 19:06:38 +0000 (0:00:00.053) 0:01:56.273 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [command] ***************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml:46 Thursday 21 July 2022 19:06:38 +0000 (0:00:00.053) 0:01:56.327 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml:50 Thursday 21 July 2022 19:06:38 +0000 (0:00:00.027) 0:01:56.354 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml:55 Thursday 21 July 2022 19:06:38 +0000 (0:00:00.026) 0:01:56.380 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml:65 Thursday 21 July 2022 19:06:38 +0000 (0:00:00.025) 0:01:56.406 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-fstab.yml:2 Thursday 21 July 2022 19:06:38 +0000 (0:00:00.037) 0:01:56.444 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "1", "storage_test_fstab_expected_mount_options_matches": "1", "storage_test_fstab_expected_mount_point_matches": "1", "storage_test_fstab_id_matches": [ "/dev/mapper/foo-test2 " ], "storage_test_fstab_mount_options_matches": [ " /opt/test2 ext4 defaults " ], "storage_test_fstab_mount_point_matches": [ " /opt/test2 " ] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-fstab.yml:12 Thursday 21 July 2022 19:06:38 +0000 (0:00:00.064) 0:01:56.508 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify the fstab mount point] ******************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-fstab.yml:19 Thursday 21 July 2022 19:06:39 +0000 (0:00:00.053) 0:01:56.562 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-fstab.yml:25 Thursday 21 July 2022 19:06:39 +0000 (0:00:00.052) 0:01:56.615 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-fstab.yml:34 Thursday 21 July 2022 19:06:39 +0000 (0:00:00.037) 0:01:56.652 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-fs.yml:4 Thursday 21 July 2022 19:06:39 +0000 (0:00:00.038) 0:01:56.691 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify fs label] ********************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-fs.yml:10 Thursday 21 July 2022 19:06:39 +0000 (0:00:00.040) 0:01:56.731 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [See whether the device node is present] ********************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-device.yml:4 Thursday 21 July 2022 19:06:39 +0000 (0:00:00.043) 0:01:56.775 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "stat": { "atime": 1658430383.5851166, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1658430383.5851166, "dev": 6, "device_type": 64768, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 59192, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/symlink", "mode": "0660", "mtime": 1658430383.5851166, "nlink": 1, "path": "/dev/mapper/foo-test2", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-device.yml:10 Thursday 21 July 2022 19:06:39 +0000 (0:00:00.378) 0:01:57.154 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about this volume] ********************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-device.yml:15 Thursday 21 July 2022 19:06:39 +0000 (0:00:00.038) 0:01:57.192 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [(1/2) Process volume type (set initial value)] *************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-device.yml:21 Thursday 21 July 2022 19:06:39 +0000 (0:00:00.036) 0:01:57.229 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [(2/2) Process volume type (get RAID value)] ****************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-device.yml:25 Thursday 21 July 2022 19:06:39 +0000 (0:00:00.035) 0:01:57.265 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-device.yml:30 Thursday 21 July 2022 19:06:39 +0000 (0:00:00.023) 0:01:57.288 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Stat the LUKS device, if encrypted] ************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:3 Thursday 21 July 2022 19:06:39 +0000 (0:00:00.040) 0:01:57.329 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:10 Thursday 21 July 2022 19:06:39 +0000 (0:00:00.025) 0:01:57.354 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [Collect LUKS info for this volume] *************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:15 Thursday 21 July 2022 19:06:41 +0000 (0:00:01.295) 0:01:58.649 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:21 Thursday 21 July 2022 19:06:41 +0000 (0:00:00.028) 0:01:58.678 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:27 Thursday 21 July 2022 19:06:41 +0000 (0:00:00.060) 0:01:58.738 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:33 Thursday 21 July 2022 19:06:41 +0000 (0:00:00.055) 0:01:58.793 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:39 Thursday 21 July 2022 19:06:41 +0000 (0:00:00.030) 0:01:58.824 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:44 Thursday 21 July 2022 19:06:41 +0000 (0:00:00.030) 0:01:58.854 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:50 Thursday 21 July 2022 19:06:41 +0000 (0:00:00.026) 0:01:58.881 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:56 Thursday 21 July 2022 19:06:41 +0000 (0:00:00.027) 0:01:58.909 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:62 Thursday 21 July 2022 19:06:41 +0000 (0:00:00.026) 0:01:58.935 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:67 Thursday 21 July 2022 19:06:41 +0000 (0:00:00.068) 0:01:59.003 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:72 Thursday 21 July 2022 19:06:41 +0000 (0:00:00.052) 0:01:59.056 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:78 Thursday 21 July 2022 19:06:41 +0000 (0:00:00.039) 0:01:59.095 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:84 Thursday 21 July 2022 19:06:41 +0000 (0:00:00.039) 0:01:59.134 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:90 Thursday 21 July 2022 19:06:41 +0000 (0:00:00.040) 0:01:59.175 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [get information about RAID] ********************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-md.yml:7 Thursday 21 July 2022 19:06:41 +0000 (0:00:00.037) 0:01:59.212 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-md.yml:13 Thursday 21 July 2022 19:06:41 +0000 (0:00:00.041) 0:01:59.254 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-md.yml:17 Thursday 21 July 2022 19:06:41 +0000 (0:00:00.041) 0:01:59.295 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-md.yml:21 Thursday 21 July 2022 19:06:41 +0000 (0:00:00.043) 0:01:59.338 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-md.yml:25 Thursday 21 July 2022 19:06:41 +0000 (0:00:00.039) 0:01:59.378 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-md.yml:31 Thursday 21 July 2022 19:06:41 +0000 (0:00:00.040) 0:01:59.418 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-md.yml:37 Thursday 21 July 2022 19:06:41 +0000 (0:00:00.039) 0:01:59.458 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the actual size of the volume] ************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:3 Thursday 21 July 2022 19:06:41 +0000 (0:00:00.043) 0:01:59.502 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "bytes": 2684354560, "changed": false, "lvm": "2g", "parted": "2GiB", "size": "2 GiB" } TASK [parse the requested size of the volume] ********************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:9 Thursday 21 July 2022 19:06:42 +0000 (0:00:00.390) 0:01:59.892 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Establish base value for expected size] ********************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:15 Thursday 21 July 2022 19:06:42 +0000 (0:00:00.037) 0:01:59.930 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:20 Thursday 21 July 2022 19:06:42 +0000 (0:00:00.040) 0:01:59.971 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_expected_size": "6442450944.0" } TASK [debug] ******************************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:25 Thursday 21 July 2022 19:06:42 +0000 (0:00:00.043) 0:02:00.015 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_pool": { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "60%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test2", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test2", "_raw_device": "/dev/mapper/foo-test2", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "25%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } } TASK [debug] ******************************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:28 Thursday 21 July 2022 19:06:42 +0000 (0:00:00.055) 0:02:00.071 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_blkinfo": { "changed": false, "failed": false, "info": { "/dev/mapper/foo-test1": { "fstype": "xfs", "label": "", "name": "/dev/mapper/foo-test1", "size": "6G", "type": "lvm", "uuid": "d739e41d-c1a1-4003-b357-25bf62753af5" }, "/dev/mapper/foo-test2": { "fstype": "ext4", "label": "", "name": "/dev/mapper/foo-test2", "size": "2.5G", "type": "lvm", "uuid": "64454f89-f2cd-4a8b-958f-4c900f72cf79" }, "/dev/nvme0n1": { "fstype": "", "label": "", "name": "/dev/nvme0n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme1n1": { "fstype": "", "label": "", "name": "/dev/nvme1n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme2n1": { "fstype": "", "label": "", "name": "/dev/nvme2n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sda": { "fstype": "LVM2_member", "label": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "HA17fj-P3dB-xyZl-9Zru-ZzXr-ULh6-93Sgui" }, "/dev/sdb": { "fstype": "", "label": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc": { "fstype": "", "label": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sr0": { "fstype": "iso9660", "label": "cidata", "name": "/dev/sr0", "size": "364K", "type": "rom", "uuid": "2022-07-21-19-04-29-00" }, "/dev/vda": { "fstype": "", "label": "", "name": "/dev/vda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vda1": { "fstype": "", "label": "", "name": "/dev/vda1", "size": "1M", "type": "partition", "uuid": "" }, "/dev/vda2": { "fstype": "vfat", "label": "", "name": "/dev/vda2", "size": "100M", "type": "partition", "uuid": "7B77-95E7" }, "/dev/vda3": { "fstype": "xfs", "label": "root", "name": "/dev/vda3", "size": "9.9G", "type": "partition", "uuid": "58dca2db-4ccd-4171-a017-dd53ad8c1c54" }, "/dev/vdb": { "fstype": "", "label": "", "name": "/dev/vdb", "size": "2G", "type": "disk", "uuid": "" }, "/dev/vdb1": { "fstype": "ext4", "label": "yumcache", "name": "/dev/vdb1", "size": "2G", "type": "partition", "uuid": "4d1ed9a1-5608-4ff8-ad89-21d95caa66fa" }, "/dev/vdc": { "fstype": "", "label": "", "name": "/dev/vdc", "size": "2G", "type": "disk", "uuid": "" }, "/dev/vdc1": { "fstype": "ext4", "label": "yumvarlib", "name": "/dev/vdc1", "size": "2G", "type": "partition", "uuid": "a7abcc97-edc5-476c-8ccd-694b4621fd28" }, "/dev/vdd": { "fstype": "", "label": "", "name": "/dev/vdd", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vde": { "fstype": "", "label": "", "name": "/dev/vde", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vdf": { "fstype": "", "label": "", "name": "/dev/vdf", "size": "10G", "type": "disk", "uuid": "" } } } } TASK [Get the size of parent/pool device] ************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:31 Thursday 21 July 2022 19:06:42 +0000 (0:00:00.058) 0:02:00.129 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "bytes": 10737418240, "changed": false, "lvm": "10g", "parted": "10GiB", "size": "10 GiB" } TASK [debug] ******************************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:36 Thursday 21 July 2022 19:06:42 +0000 (0:00:00.391) 0:02:00.520 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_pool_size": { "bytes": 10737418240, "changed": false, "failed": false, "lvm": "10g", "parted": "10GiB", "size": "10 GiB" } } TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:39 Thursday 21 July 2022 19:06:43 +0000 (0:00:00.060) 0:02:00.581 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_expected_size": "2684354560.0" }, "changed": false } TASK [debug] ******************************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:44 Thursday 21 July 2022 19:06:43 +0000 (0:00:00.052) 0:02:00.634 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_actual_size": { "bytes": 2684354560, "changed": false, "failed": false, "lvm": "2g", "parted": "2GiB", "size": "2 GiB" } } TASK [debug] ******************************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:47 Thursday 21 July 2022 19:06:43 +0000 (0:00:00.038) 0:02:00.672 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_expected_size": "2684354560.0" } TASK [assert] ****************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:50 Thursday 21 July 2022 19:06:43 +0000 (0:00:00.088) 0:02:00.760 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Get information about the LV] ******************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-cache.yml:6 Thursday 21 July 2022 19:06:43 +0000 (0:00:00.094) 0:02:00.855 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "cmd": [ "lvs", "--noheadings", "--nameprefixes", "--units=b", "--nosuffix", "--unquoted", "-o", "name,attr,cache_total_blocks,chunk_size,segtype", "foo/test2" ], "delta": "0:00:00.032325", "end": "2022-07-21 15:06:43.792523", "rc": 0, "start": "2022-07-21 15:06:43.760198" } STDOUT: LVM2_LV_NAME=test2 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-cache.yml:14 Thursday 21 July 2022 19:06:43 +0000 (0:00:00.416) 0:02:01.272 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_lv_segtype": [ "linear" ] }, "changed": false } TASK [check segment type] ****************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-cache.yml:17 Thursday 21 July 2022 19:06:43 +0000 (0:00:00.051) 0:02:01.323 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-cache.yml:22 Thursday 21 July 2022 19:06:43 +0000 (0:00:00.054) 0:02:01.378 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the requested cache size] ****************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-cache.yml:26 Thursday 21 July 2022 19:06:43 +0000 (0:00:00.040) 0:02:01.418 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-cache.yml:32 Thursday 21 July 2022 19:06:43 +0000 (0:00:00.037) 0:02:01.455 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-cache.yml:36 Thursday 21 July 2022 19:06:43 +0000 (0:00:00.037) 0:02:01.493 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume.yml:16 Thursday 21 July 2022 19:06:43 +0000 (0:00:00.040) 0:02:01.534 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [Verify the volumes with no pool were correctly managed] ****************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-role-results.yml:43 Thursday 21 July 2022 19:06:44 +0000 (0:00:00.034) 0:02:01.568 ********* TASK [Clean up variable namespace] ********************************************* task path: /tmp/tmpmnqeja9w/tests/storage/verify-role-results.yml:53 Thursday 21 July 2022 19:06:44 +0000 (0:00:00.021) 0:02:01.590 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_blkinfo": null, "storage_test_crypttab": null, "storage_test_fstab": null }, "changed": false } TASK [Get the size of test2 volume] ******************************************** task path: /tmp/tmpmnqeja9w/tests/storage/tests_lvm_percent_size.yml:114 Thursday 21 July 2022 19:06:44 +0000 (0:00:00.034) 0:02:01.624 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "cmd": [ "lsblk", "--noheadings", "-o", "SIZE", "/dev/mapper/foo-test2" ], "delta": "0:00:00.004915", "end": "2022-07-21 15:06:44.524052", "rc": 0, "start": "2022-07-21 15:06:44.519137" } STDOUT: 2.5G TASK [Remove the test1 volume without changing its size] *********************** task path: /tmp/tmpmnqeja9w/tests/storage/tests_lvm_percent_size.yml:119 Thursday 21 July 2022 19:06:44 +0000 (0:00:00.383) 0:02:02.008 ********* TASK [fedora.linux_system_roles.storage : set platform/version specific variables] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2 Thursday 21 July 2022 19:06:44 +0000 (0:00:00.058) 0:02:02.066 ********* included: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for /cache/rhel-8-y.qcow2.snap TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2 Thursday 21 July 2022 19:06:44 +0000 (0:00:00.037) 0:02:02.103 ********* ok: [/cache/rhel-8-y.qcow2.snap] TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7 Thursday 21 July 2022 19:06:45 +0000 (0:00:00.533) 0:02:02.637 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [/cache/rhel-8-y.qcow2.snap] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [/cache/rhel-8-y.qcow2.snap] => (item=RedHat_8.yml) => { "ansible_facts": { "blivet_package_list": [ "python3-blivet", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap", "vdo", "kmod-kvdo", "xfsprogs" ] }, "ansible_included_var_files": [ "/tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/vars/RedHat_8.yml" ], "ansible_loop_var": "item", "changed": false, "item": "RedHat_8.yml" } skipping: [/cache/rhel-8-y.qcow2.snap] => (item=RedHat_8.7.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat_8.7.yml", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : define an empty list of pools to be used in testing] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5 Thursday 21 July 2022 19:06:45 +0000 (0:00:00.063) 0:02:02.701 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : define an empty list of volumes to be used in testing] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9 Thursday 21 July 2022 19:06:45 +0000 (0:00:00.033) 0:02:02.735 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : include the appropriate provider tasks] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13 Thursday 21 July 2022 19:06:45 +0000 (0:00:00.043) 0:02:02.778 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount included: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for /cache/rhel-8-y.qcow2.snap TASK [fedora.linux_system_roles.storage : get a list of rpm packages installed on host machine] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2 Thursday 21 July 2022 19:06:45 +0000 (0:00:00.063) 0:02:02.841 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : make sure blivet is available] ******* task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:7 Thursday 21 July 2022 19:06:45 +0000 (0:00:00.023) 0:02:02.865 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [fedora.linux_system_roles.storage : show storage_pools] ****************** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:13 Thursday 21 July 2022 19:06:46 +0000 (0:00:01.375) 0:02:04.241 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_pools": [ { "disks": [ "sda" ], "name": "foo", "state": "present", "volumes": [ { "mount_point": "/opt/test1", "name": "test1", "size": "60%", "state": "absent" }, { "mount_point": "/opt/test2", "name": "test2", "size": "25%" } ] } ] } TASK [fedora.linux_system_roles.storage : show storage_volumes] **************** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:18 Thursday 21 July 2022 19:06:46 +0000 (0:00:00.044) 0:02:04.285 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [fedora.linux_system_roles.storage : get required packages] *************** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:23 Thursday 21 July 2022 19:06:46 +0000 (0:00:00.083) 0:02:04.368 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [ "lvm2" ], "pools": [], "volumes": [] } TASK [fedora.linux_system_roles.storage : enable copr repositories if needed] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:35 Thursday 21 July 2022 19:06:48 +0000 (0:00:01.705) 0:02:06.073 ********* included: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml for /cache/rhel-8-y.qcow2.snap TASK [fedora.linux_system_roles.storage : check if the COPR support packages should be installed] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:2 Thursday 21 July 2022 19:06:48 +0000 (0:00:00.050) 0:02:06.124 ********* TASK [fedora.linux_system_roles.storage : make sure COPR support packages are present] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:13 Thursday 21 July 2022 19:06:48 +0000 (0:00:00.083) 0:02:06.207 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : enable COPRs] ************************ task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:18 Thursday 21 July 2022 19:06:48 +0000 (0:00:00.042) 0:02:06.250 ********* TASK [fedora.linux_system_roles.storage : make sure required packages are installed] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:41 Thursday 21 July 2022 19:06:48 +0000 (0:00:00.037) 0:02:06.287 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [fedora.linux_system_roles.storage : get service facts] ******************* task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:47 Thursday 21 July 2022 19:06:50 +0000 (0:00:01.339) 0:02:07.627 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "services": { "NetworkManager-dispatcher.service": { "name": "NetworkManager-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "NetworkManager-wait-online.service": { "name": "NetworkManager-wait-online.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "NetworkManager.service": { "name": "NetworkManager.service", "source": "systemd", "state": "running", "status": "enabled" }, "arp-ethers.service": { "name": "arp-ethers.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "auditd.service": { "name": "auditd.service", "source": "systemd", "state": "running", "status": "enabled" }, "auth-rpcgss-module.service": { "name": "auth-rpcgss-module.service", "source": "systemd", "state": "stopped", "status": "static" }, "autovt@.service": { "name": "autovt@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "blivet.service": { "name": "blivet.service", "source": "systemd", "state": "inactive", "status": "static" }, "blk-availability.service": { "name": "blk-availability.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "chrony-dnssrv@.service": { "name": "chrony-dnssrv@.service", "source": "systemd", "state": "unknown", "status": "static" }, "chrony-wait.service": { "name": "chrony-wait.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd.service": { "name": "chronyd.service", "source": "systemd", "state": "running", "status": "enabled" }, "cloud-config.service": { "name": "cloud-config.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-final.service": { "name": "cloud-final.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init-hotplugd.service": { "name": "cloud-init-hotplugd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cloud-init-local.service": { "name": "cloud-init-local.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init.service": { "name": "cloud-init.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cockpit-motd.service": { "name": "cockpit-motd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cockpit-wsinstance-http.service": { "name": "cockpit-wsinstance-http.service", "source": "systemd", "state": "inactive", "status": "static" }, "cockpit-wsinstance-https-factory@.service": { "name": "cockpit-wsinstance-https-factory@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cockpit-wsinstance-https@.service": { "name": "cockpit-wsinstance-https@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cockpit.service": { "name": "cockpit.service", "source": "systemd", "state": "inactive", "status": "static" }, "console-getty.service": { "name": "console-getty.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "container-getty@.service": { "name": "container-getty@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cpupower.service": { "name": "cpupower.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "crond.service": { "name": "crond.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-org.freedesktop.hostname1.service": { "name": "dbus-org.freedesktop.hostname1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.locale1.service": { "name": "dbus-org.freedesktop.locale1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.login1.service": { "name": "dbus-org.freedesktop.login1.service", "source": "systemd", "state": "active", "status": "static" }, "dbus-org.freedesktop.nm-dispatcher.service": { "name": "dbus-org.freedesktop.nm-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "dbus-org.freedesktop.portable1.service": { "name": "dbus-org.freedesktop.portable1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.timedate1.service": { "name": "dbus-org.freedesktop.timedate1.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "dbus.service": { "name": "dbus.service", "source": "systemd", "state": "running", "status": "static" }, "dbxtool.service": { "name": "dbxtool.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "debug-shell.service": { "name": "debug-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "display-manager.service": { "name": "display-manager.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "dm-event.service": { "name": "dm-event.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-makecache.service": { "name": "dnf-makecache.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-cmdline.service": { "name": "dracut-cmdline.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-initqueue.service": { "name": "dracut-initqueue.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-mount.service": { "name": "dracut-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-mount.service": { "name": "dracut-pre-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-pivot.service": { "name": "dracut-pre-pivot.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-trigger.service": { "name": "dracut-pre-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-udev.service": { "name": "dracut-pre-udev.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown.service": { "name": "dracut-shutdown.service", "source": "systemd", "state": "stopped", "status": "static" }, "emergency.service": { "name": "emergency.service", "source": "systemd", "state": "stopped", "status": "static" }, "fcoe.service": { "name": "fcoe.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "fstrim.service": { "name": "fstrim.service", "source": "systemd", "state": "inactive", "status": "static" }, "getty@.service": { "name": "getty@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "getty@tty1.service": { "name": "getty@tty1.service", "source": "systemd", "state": "running", "status": "active" }, "grub-boot-indeterminate.service": { "name": "grub-boot-indeterminate.service", "source": "systemd", "state": "inactive", "status": "static" }, "gssproxy.service": { "name": "gssproxy.service", "source": "systemd", "state": "running", "status": "disabled" }, "halt-local.service": { "name": "halt-local.service", "source": "systemd", "state": "inactive", "status": "static" }, "hv_kvp_daemon.service": { "name": "hv_kvp_daemon.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "import-state.service": { "name": "import-state.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "initrd-cleanup.service": { "name": "initrd-cleanup.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-parse-etc.service": { "name": "initrd-parse-etc.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-switch-root.service": { "name": "initrd-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-udevadm-cleanup-db.service": { "name": "initrd-udevadm-cleanup-db.service", "source": "systemd", "state": "stopped", "status": "static" }, "insights-client-boot.service": { "name": "insights-client-boot.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "insights-client-results.service": { "name": "insights-client-results.service", "source": "systemd", "state": "inactive", "status": "static" }, "insights-client.service": { "name": "insights-client.service", "source": "systemd", "state": "inactive", "status": "static" }, "irqbalance.service": { "name": "irqbalance.service", "source": "systemd", "state": "running", "status": "enabled" }, "iscsi-shutdown.service": { "name": "iscsi-shutdown.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "iscsi.service": { "name": "iscsi.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "iscsid.service": { "name": "iscsid.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "kdump.service": { "name": "kdump.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "kmod-static-nodes.service": { "name": "kmod-static-nodes.service", "source": "systemd", "state": "stopped", "status": "static" }, "kvm_stat.service": { "name": "kvm_stat.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "ldconfig.service": { "name": "ldconfig.service", "source": "systemd", "state": "stopped", "status": "static" }, "loadmodules.service": { "name": "loadmodules.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "lvm2-activation-early.service": { "name": "lvm2-activation-early.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "lvm2-activation.service": { "name": "lvm2-activation.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "lvm2-lvmpolld.service": { "name": "lvm2-lvmpolld.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-monitor.service": { "name": "lvm2-monitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "lvm2-pvscan@.service": { "name": "lvm2-pvscan@.service", "source": "systemd", "state": "unknown", "status": "static" }, "lvm2-pvscan@8:0.service": { "name": "lvm2-pvscan@8:0.service", "source": "systemd", "state": "stopped", "status": "active" }, "man-db-cache-update.service": { "name": "man-db-cache-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "man-db-restart-cache-update.service": { "name": "man-db-restart-cache-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "mdadm-grow-continue@.service": { "name": "mdadm-grow-continue@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdadm-last-resort@.service": { "name": "mdadm-last-resort@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdcheck_continue.service": { "name": "mdcheck_continue.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdcheck_start.service": { "name": "mdcheck_start.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmon@.service": { "name": "mdmon@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdmonitor-oneshot.service": { "name": "mdmonitor-oneshot.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmonitor.service": { "name": "mdmonitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "messagebus.service": { "name": "messagebus.service", "source": "systemd", "state": "active", "status": "static" }, "microcode.service": { "name": "microcode.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "multipathd.service": { "name": "multipathd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "ndctl-monitor.service": { "name": "ndctl-monitor.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "network.service": { "name": "network.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "nfs-blkmap.service": { "name": "nfs-blkmap.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-convert.service": { "name": "nfs-convert.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-idmapd.service": { "name": "nfs-idmapd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-mountd.service": { "name": "nfs-mountd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-server.service": { "name": "nfs-server.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-utils.service": { "name": "nfs-utils.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfsdcld.service": { "name": "nfsdcld.service", "source": "systemd", "state": "stopped", "status": "static" }, "nis-domainname.service": { "name": "nis-domainname.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "ntpd.service": { "name": "ntpd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ntpdate.service": { "name": "ntpdate.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "oddjobd.service": { "name": "oddjobd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "packagekit-offline-update.service": { "name": "packagekit-offline-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "packagekit.service": { "name": "packagekit.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-quit-wait.service": { "name": "plymouth-quit-wait.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "plymouth-start.service": { "name": "plymouth-start.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "polkit.service": { "name": "polkit.service", "source": "systemd", "state": "running", "status": "static" }, "qemu-guest-agent.service": { "name": "qemu-guest-agent.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "quotaon.service": { "name": "quotaon.service", "source": "systemd", "state": "inactive", "status": "static" }, "rbdmap.service": { "name": "rbdmap.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "rc-local.service": { "name": "rc-local.service", "source": "systemd", "state": "stopped", "status": "static" }, "rdisc.service": { "name": "rdisc.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rescue.service": { "name": "rescue.service", "source": "systemd", "state": "stopped", "status": "static" }, "rhcd.service": { "name": "rhcd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhsm-facts.service": { "name": "rhsm-facts.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhsm.service": { "name": "rhsm.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhsmcertd.service": { "name": "rhsmcertd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpc-gssd.service": { "name": "rpc-gssd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd-notify.service": { "name": "rpc-statd-notify.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd.service": { "name": "rpc-statd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpcbind.service": { "name": "rpcbind.service", "source": "systemd", "state": "running", "status": "enabled" }, "rsyslog.service": { "name": "rsyslog.service", "source": "systemd", "state": "running", "status": "enabled" }, "selinux-autorelabel-mark.service": { "name": "selinux-autorelabel-mark.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "selinux-autorelabel.service": { "name": "selinux-autorelabel.service", "source": "systemd", "state": "inactive", "status": "static" }, "serial-getty@.service": { "name": "serial-getty@.service", "source": "systemd", "state": "unknown", "status": "indirect" }, "serial-getty@ttyS0.service": { "name": "serial-getty@ttyS0.service", "source": "systemd", "state": "running", "status": "active" }, "snapd.seeded.service": { "name": "snapd.seeded.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sntp.service": { "name": "sntp.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sshd-keygen.service": { "name": "sshd-keygen.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sshd-keygen@.service": { "name": "sshd-keygen@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@ecdsa.service": { "name": "sshd-keygen@ecdsa.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd-keygen@ed25519.service": { "name": "sshd-keygen@ed25519.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd-keygen@rsa.service": { "name": "sshd-keygen@rsa.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd.service": { "name": "sshd.service", "source": "systemd", "state": "running", "status": "enabled" }, "sshd@.service": { "name": "sshd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "sssd-autofs.service": { "name": "sssd-autofs.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-kcm.service": { "name": "sssd-kcm.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "sssd-nss.service": { "name": "sssd-nss.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pac.service": { "name": "sssd-pac.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pam.service": { "name": "sssd-pam.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-ssh.service": { "name": "sssd-ssh.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-sudo.service": { "name": "sssd-sudo.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd.service": { "name": "sssd.service", "source": "systemd", "state": "running", "status": "enabled" }, "syslog.service": { "name": "syslog.service", "source": "systemd", "state": "active", "status": "enabled" }, "system-update-cleanup.service": { "name": "system-update-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-ask-password-console.service": { "name": "systemd-ask-password-console.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-wall.service": { "name": "systemd-ask-password-wall.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-backlight@.service": { "name": "systemd-backlight@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-binfmt.service": { "name": "systemd-binfmt.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-coredump@.service": { "name": "systemd-coredump@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-exit.service": { "name": "systemd-exit.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-firstboot.service": { "name": "systemd-firstboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck-root.service": { "name": "systemd-fsck-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck@.service": { "name": "systemd-fsck@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-fsck@dev-disk-by\\x2duuid-7B77\\x2d95E7.service": { "name": "systemd-fsck@dev-disk-by\\x2duuid-7B77\\x2d95E7.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-fsck@dev-vdb1.service": { "name": "systemd-fsck@dev-vdb1.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-fsck@dev-vdc1.service": { "name": "systemd-fsck@dev-vdc1.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-halt.service": { "name": "systemd-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hibernate-resume@.service": { "name": "systemd-hibernate-resume@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-hibernate.service": { "name": "systemd-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hostnamed.service": { "name": "systemd-hostnamed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hwdb-update.service": { "name": "systemd-hwdb-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-hybrid-sleep.service": { "name": "systemd-hybrid-sleep.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-initctl.service": { "name": "systemd-initctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-catalog-update.service": { "name": "systemd-journal-catalog-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-flush.service": { "name": "systemd-journal-flush.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journald.service": { "name": "systemd-journald.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-kexec.service": { "name": "systemd-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-localed.service": { "name": "systemd-localed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-logind.service": { "name": "systemd-logind.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-machine-id-commit.service": { "name": "systemd-machine-id-commit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-modules-load.service": { "name": "systemd-modules-load.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-networkd-wait-online.service": { "name": "systemd-networkd-wait-online.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "systemd-portabled.service": { "name": "systemd-portabled.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-poweroff.service": { "name": "systemd-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-quotacheck.service": { "name": "systemd-quotacheck.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-random-seed.service": { "name": "systemd-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-reboot.service": { "name": "systemd-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-remount-fs.service": { "name": "systemd-remount-fs.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-resolved.service": { "name": "systemd-resolved.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-rfkill.service": { "name": "systemd-rfkill.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend-then-hibernate.service": { "name": "systemd-suspend-then-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend.service": { "name": "systemd-suspend.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-sysctl.service": { "name": "systemd-sysctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-sysusers.service": { "name": "systemd-sysusers.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-timedated.service": { "name": "systemd-timedated.service", "source": "systemd", "state": "inactive", "status": "masked" }, "systemd-timesyncd.service": { "name": "systemd-timesyncd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "systemd-tmpfiles-clean.service": { "name": "systemd-tmpfiles-clean.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev.service": { "name": "systemd-tmpfiles-setup-dev.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup.service": { "name": "systemd-tmpfiles-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-settle.service": { "name": "systemd-udev-settle.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-trigger.service": { "name": "systemd-udev-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udevd.service": { "name": "systemd-udevd.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-update-done.service": { "name": "systemd-update-done.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp-runlevel.service": { "name": "systemd-update-utmp-runlevel.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp.service": { "name": "systemd-update-utmp.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-user-sessions.service": { "name": "systemd-user-sessions.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-vconsole-setup.service": { "name": "systemd-vconsole-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-volatile-root.service": { "name": "systemd-volatile-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "tcsd.service": { "name": "tcsd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "teamd@.service": { "name": "teamd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "timedatex.service": { "name": "timedatex.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "tuned.service": { "name": "tuned.service", "source": "systemd", "state": "running", "status": "enabled" }, "unbound-anchor.service": { "name": "unbound-anchor.service", "source": "systemd", "state": "stopped", "status": "static" }, "user-runtime-dir@.service": { "name": "user-runtime-dir@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user-runtime-dir@0.service": { "name": "user-runtime-dir@0.service", "source": "systemd", "state": "stopped", "status": "active" }, "user@.service": { "name": "user@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user@0.service": { "name": "user@0.service", "source": "systemd", "state": "running", "status": "active" }, "vdo-start-by-dev@.service": { "name": "vdo-start-by-dev@.service", "source": "systemd", "state": "unknown", "status": "static" }, "vdo.service": { "name": "vdo.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "ypbind.service": { "name": "ypbind.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "yppasswdd.service": { "name": "yppasswdd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ypserv.service": { "name": "ypserv.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ypxfrd.service": { "name": "ypxfrd.service", "source": "systemd", "state": "stopped", "status": "not-found" } } }, "changed": false } TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] ***** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:53 Thursday 21 July 2022 19:06:51 +0000 (0:00:01.727) 0:02:09.354 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:58 Thursday 21 July 2022 19:06:51 +0000 (0:00:00.062) 0:02:09.417 ********* TASK [fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64 Thursday 21 July 2022 19:06:51 +0000 (0:00:00.024) 0:02:09.441 ********* changed: [/cache/rhel-8-y.qcow2.snap] => { "actions": [ { "action": "destroy format", "device": "/dev/mapper/foo-test1", "fs_type": "xfs" }, { "action": "destroy device", "device": "/dev/mapper/foo-test1", "fs_type": null } ], "changed": true, "crypts": [], "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/vda2", "/dev/vda3", "/dev/mapper/foo-test2", "/dev/sdb", "/dev/sdc", "/dev/nvme0n1", "/dev/nvme1n1", "/dev/nvme2n1", "/dev/vdb1", "/dev/vdc1", "/dev/vdd", "/dev/vde", "/dev/vdf" ], "mounts": [ { "fstype": "xfs", "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "absent" }, { "dump": 0, "fstype": "ext4", "opts": "defaults", "passno": 0, "path": "/opt/test2", "src": "/dev/mapper/foo-test2", "state": "mounted" } ], "packages": [ "lvm2", "xfsprogs", "e2fsprogs", "dosfstools" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "60%", "state": "absent", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test2", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test2", "_raw_device": "/dev/mapper/foo-test2", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "25%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:78 Thursday 21 July 2022 19:06:54 +0000 (0:00:02.184) 0:02:11.626 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90 Thursday 21 July 2022 19:06:54 +0000 (0:00:00.039) 0:02:11.665 ********* TASK [fedora.linux_system_roles.storage : show blivet_output] ****************** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:96 Thursday 21 July 2022 19:06:54 +0000 (0:00:00.022) 0:02:11.687 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "blivet_output": { "actions": [ { "action": "destroy format", "device": "/dev/mapper/foo-test1", "fs_type": "xfs" }, { "action": "destroy device", "device": "/dev/mapper/foo-test1", "fs_type": null } ], "changed": true, "crypts": [], "failed": false, "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/vda2", "/dev/vda3", "/dev/mapper/foo-test2", "/dev/sdb", "/dev/sdc", "/dev/nvme0n1", "/dev/nvme1n1", "/dev/nvme2n1", "/dev/vdb1", "/dev/vdc1", "/dev/vdd", "/dev/vde", "/dev/vdf" ], "mounts": [ { "fstype": "xfs", "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "absent" }, { "dump": 0, "fstype": "ext4", "opts": "defaults", "passno": 0, "path": "/opt/test2", "src": "/dev/mapper/foo-test2", "state": "mounted" } ], "packages": [ "lvm2", "xfsprogs", "e2fsprogs", "dosfstools" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "60%", "state": "absent", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test2", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test2", "_raw_device": "/dev/mapper/foo-test2", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "25%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } } TASK [fedora.linux_system_roles.storage : set the list of pools for test verification] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:101 Thursday 21 July 2022 19:06:54 +0000 (0:00:00.045) 0:02:11.733 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "60%", "state": "absent", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test2", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test2", "_raw_device": "/dev/mapper/foo-test2", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "25%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] }, "changed": false } TASK [fedora.linux_system_roles.storage : set the list of volumes for test verification] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:105 Thursday 21 July 2022 19:06:54 +0000 (0:00:00.041) 0:02:11.775 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : remove obsolete mounts] ************** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:121 Thursday 21 July 2022 19:06:54 +0000 (0:00:00.041) 0:02:11.816 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount changed: [/cache/rhel-8-y.qcow2.snap] => (item={'src': '/dev/mapper/foo-test1', 'path': '/opt/test1', 'state': 'absent', 'fstype': 'xfs'}) => { "ansible_loop_var": "mount_info", "backup_file": "", "boot": "yes", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": { "fstype": "xfs", "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "absent" }, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test1" } TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132 Thursday 21 July 2022 19:06:54 +0000 (0:00:00.407) 0:02:12.223 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.storage : set up new/current mounts] *********** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:137 Thursday 21 July 2022 19:06:55 +0000 (0:00:00.625) 0:02:12.849 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount ok: [/cache/rhel-8-y.qcow2.snap] => (item={'src': '/dev/mapper/foo-test2', 'path': '/opt/test2', 'fstype': 'ext4', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted'}) => { "ansible_loop_var": "mount_info", "backup_file": "", "boot": "yes", "changed": false, "dump": "0", "fstab": "/etc/fstab", "fstype": "ext4", "mount_info": { "dump": 0, "fstype": "ext4", "opts": "defaults", "passno": 0, "path": "/opt/test2", "src": "/dev/mapper/foo-test2", "state": "mounted" }, "name": "/opt/test2", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test2" } TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148 Thursday 21 July 2022 19:06:55 +0000 (0:00:00.461) 0:02:13.310 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.storage : retrieve facts for the /etc/crypttab file] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:156 Thursday 21 July 2022 19:06:56 +0000 (0:00:00.654) 0:02:13.965 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "stat": { "exists": false } } TASK [fedora.linux_system_roles.storage : manage /etc/crypttab to account for changes we just made] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:161 Thursday 21 July 2022 19:06:56 +0000 (0:00:00.379) 0:02:14.345 ********* TASK [fedora.linux_system_roles.storage : Update facts] ************************ task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:183 Thursday 21 July 2022 19:06:56 +0000 (0:00:00.026) 0:02:14.371 ********* ok: [/cache/rhel-8-y.qcow2.snap] META: role_complete for /cache/rhel-8-y.qcow2.snap TASK [include_tasks] *********************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/tests_lvm_percent_size.yml:136 Thursday 21 July 2022 19:06:57 +0000 (0:00:01.000) 0:02:15.372 ********* included: /tmp/tmpmnqeja9w/tests/storage/verify-role-results.yml for /cache/rhel-8-y.qcow2.snap TASK [Print out pool information] ********************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-role-results.yml:1 Thursday 21 July 2022 19:06:57 +0000 (0:00:00.048) 0:02:15.420 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "60%", "state": "absent", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test2", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test2", "_raw_device": "/dev/mapper/foo-test2", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "25%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] } TASK [Print out volume information] ******************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-role-results.yml:6 Thursday 21 July 2022 19:06:57 +0000 (0:00:00.054) 0:02:15.474 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => {} TASK [Collect info about the volumes.] ***************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-role-results.yml:14 Thursday 21 July 2022 19:06:57 +0000 (0:00:00.039) 0:02:15.513 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "info": { "/dev/mapper/foo-test2": { "fstype": "ext4", "label": "", "name": "/dev/mapper/foo-test2", "size": "2.5G", "type": "lvm", "uuid": "64454f89-f2cd-4a8b-958f-4c900f72cf79" }, "/dev/nvme0n1": { "fstype": "", "label": "", "name": "/dev/nvme0n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme1n1": { "fstype": "", "label": "", "name": "/dev/nvme1n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme2n1": { "fstype": "", "label": "", "name": "/dev/nvme2n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sda": { "fstype": "LVM2_member", "label": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "HA17fj-P3dB-xyZl-9Zru-ZzXr-ULh6-93Sgui" }, "/dev/sdb": { "fstype": "", "label": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc": { "fstype": "", "label": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sr0": { "fstype": "iso9660", "label": "cidata", "name": "/dev/sr0", "size": "364K", "type": "rom", "uuid": "2022-07-21-19-04-29-00" }, "/dev/vda": { "fstype": "", "label": "", "name": "/dev/vda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vda1": { "fstype": "", "label": "", "name": "/dev/vda1", "size": "1M", "type": "partition", "uuid": "" }, "/dev/vda2": { "fstype": "vfat", "label": "", "name": "/dev/vda2", "size": "100M", "type": "partition", "uuid": "7B77-95E7" }, "/dev/vda3": { "fstype": "xfs", "label": "root", "name": "/dev/vda3", "size": "9.9G", "type": "partition", "uuid": "58dca2db-4ccd-4171-a017-dd53ad8c1c54" }, "/dev/vdb": { "fstype": "", "label": "", "name": "/dev/vdb", "size": "2G", "type": "disk", "uuid": "" }, "/dev/vdb1": { "fstype": "ext4", "label": "yumcache", "name": "/dev/vdb1", "size": "2G", "type": "partition", "uuid": "4d1ed9a1-5608-4ff8-ad89-21d95caa66fa" }, "/dev/vdc": { "fstype": "", "label": "", "name": "/dev/vdc", "size": "2G", "type": "disk", "uuid": "" }, "/dev/vdc1": { "fstype": "ext4", "label": "yumvarlib", "name": "/dev/vdc1", "size": "2G", "type": "partition", "uuid": "a7abcc97-edc5-476c-8ccd-694b4621fd28" }, "/dev/vdd": { "fstype": "", "label": "", "name": "/dev/vdd", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vde": { "fstype": "", "label": "", "name": "/dev/vde", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vdf": { "fstype": "", "label": "", "name": "/dev/vdf", "size": "10G", "type": "disk", "uuid": "" } } } TASK [Read the /etc/fstab file for volume existence] *************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-role-results.yml:19 Thursday 21 July 2022 19:06:58 +0000 (0:00:00.385) 0:02:15.899 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "cmd": [ "cat", "/etc/fstab" ], "delta": "0:00:00.002794", "end": "2022-07-21 15:06:58.792572", "rc": 0, "start": "2022-07-21 15:06:58.789778" } STDOUT: UUID=58dca2db-4ccd-4171-a017-dd53ad8c1c54 / xfs defaults 0 0 UUID=7B77-95E7 /boot/efi vfat defaults,uid=0,gid=0,umask=077,shortname=winnt 0 2 /dev/vdb1 /var/cache/dnf auto defaults,nofail,x-systemd.requires=cloud-init.service,_netdev,comment=cloudconfig 0 2 /dev/vdc1 /var/lib/dnf auto defaults,nofail,x-systemd.requires=cloud-init.service,_netdev,comment=cloudconfig 0 2 /dev/mapper/foo-test2 /opt/test2 ext4 defaults 0 0 TASK [Read the /etc/crypttab file] ********************************************* task path: /tmp/tmpmnqeja9w/tests/storage/verify-role-results.yml:24 Thursday 21 July 2022 19:06:58 +0000 (0:00:00.375) 0:02:16.274 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "cmd": [ "cat", "/etc/crypttab" ], "delta": "0:00:00.002910", "end": "2022-07-21 15:06:59.167158", "failed_when_result": false, "rc": 1, "start": "2022-07-21 15:06:59.164248" } STDERR: cat: /etc/crypttab: No such file or directory MSG: non-zero return code TASK [Verify the volumes listed in storage_pools were correctly managed] ******* task path: /tmp/tmpmnqeja9w/tests/storage/verify-role-results.yml:33 Thursday 21 July 2022 19:06:59 +0000 (0:00:00.375) 0:02:16.649 ********* included: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool.yml for /cache/rhel-8-y.qcow2.snap => (item={'disks': ['sda'], 'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'name': 'foo', 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_metadata_version': None, 'state': 'present', 'type': 'lvm', 'volumes': [{'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'name': 'test1', 'raid_level': None, 'size': '60%', 'state': 'absent', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test1', '_raw_device': '/dev/mapper/foo-test1', '_mount_id': '/dev/mapper/foo-test1'}, {'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'ext4', 'mount_options': 'defaults', 'mount_point': '/opt/test2', 'name': 'test2', 'raid_level': None, 'size': '25%', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test2', '_raw_device': '/dev/mapper/foo-test2', '_mount_id': '/dev/mapper/foo-test2', '_kernel_device': '/dev/dm-0', '_raw_kernel_device': '/dev/dm-0'}], 'raid_chunk_size': None}) TASK [Set _storage_pool_tests] ************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool.yml:5 Thursday 21 July 2022 19:06:59 +0000 (0:00:00.063) 0:02:16.713 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_pool_tests": [ "members", "volumes" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool.yml:18 Thursday 21 July 2022 19:06:59 +0000 (0:00:00.036) 0:02:16.750 ********* included: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml for /cache/rhel-8-y.qcow2.snap => (item=members) included: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-volumes.yml for /cache/rhel-8-y.qcow2.snap => (item=volumes) TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:1 Thursday 21 July 2022 19:06:59 +0000 (0:00:00.049) 0:02:16.800 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_pv_count": "1", "_storage_test_pool_pvs_lvm": [ "/dev/sda" ] }, "changed": false } TASK [Get the canonical device path for each member device] ******************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:6 Thursday 21 July 2022 19:06:59 +0000 (0:00:00.097) 0:02:16.898 ********* ok: [/cache/rhel-8-y.qcow2.snap] => (item=/dev/sda) => { "ansible_loop_var": "pv", "changed": false, "device": "/dev/sda", "pv": "/dev/sda" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:15 Thursday 21 July 2022 19:06:59 +0000 (0:00:00.434) 0:02:17.332 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "__pvs_lvm_len": "1" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:19 Thursday 21 July 2022 19:06:59 +0000 (0:00:00.084) 0:02:17.417 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_pool_pvs": [ "/dev/sda" ] }, "changed": false } TASK [Verify PV count] ********************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:23 Thursday 21 July 2022 19:06:59 +0000 (0:00:00.084) 0:02:17.501 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:29 Thursday 21 July 2022 19:07:00 +0000 (0:00:00.082) 0:02:17.584 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:33 Thursday 21 July 2022 19:07:00 +0000 (0:00:00.037) 0:02:17.621 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:37 Thursday 21 July 2022 19:07:00 +0000 (0:00:00.048) 0:02:17.670 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check the type of each PV] *********************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:41 Thursday 21 July 2022 19:07:00 +0000 (0:00:00.024) 0:02:17.695 ********* ok: [/cache/rhel-8-y.qcow2.snap] => (item=/dev/sda) => { "ansible_loop_var": "pv", "changed": false, "pv": "/dev/sda" } MSG: All assertions passed TASK [Check MD RAID] *********************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:50 Thursday 21 July 2022 19:07:00 +0000 (0:00:00.042) 0:02:17.737 ********* included: /tmp/tmpmnqeja9w/tests/storage/verify-pool-md.yml for /cache/rhel-8-y.qcow2.snap TASK [get information about RAID] ********************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-md.yml:6 Thursday 21 July 2022 19:07:00 +0000 (0:00:00.043) 0:02:17.781 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-md.yml:12 Thursday 21 July 2022 19:07:00 +0000 (0:00:00.026) 0:02:17.808 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-md.yml:16 Thursday 21 July 2022 19:07:00 +0000 (0:00:00.025) 0:02:17.834 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-md.yml:20 Thursday 21 July 2022 19:07:00 +0000 (0:00:00.026) 0:02:17.860 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-md.yml:24 Thursday 21 July 2022 19:07:00 +0000 (0:00:00.026) 0:02:17.886 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-md.yml:30 Thursday 21 July 2022 19:07:00 +0000 (0:00:00.027) 0:02:17.913 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-md.yml:36 Thursday 21 July 2022 19:07:00 +0000 (0:00:00.037) 0:02:17.951 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-md.yml:44 Thursday 21 July 2022 19:07:00 +0000 (0:00:00.038) 0:02:17.989 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_md_active_devices_re": null, "storage_test_md_metadata_version_re": null, "storage_test_md_spare_devices_re": null }, "changed": false } TASK [Check LVM RAID] ********************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:53 Thursday 21 July 2022 19:07:00 +0000 (0:00:00.041) 0:02:18.031 ********* included: /tmp/tmpmnqeja9w/tests/storage/verify-pool-members-lvmraid.yml for /cache/rhel-8-y.qcow2.snap TASK [Validate pool member LVM RAID settings] ********************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-members-lvmraid.yml:1 Thursday 21 July 2022 19:07:00 +0000 (0:00:00.044) 0:02:18.076 ********* included: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-lvmraid.yml for /cache/rhel-8-y.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'name': 'test1', 'raid_level': None, 'size': '60%', 'state': 'absent', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test1', '_raw_device': '/dev/mapper/foo-test1', '_mount_id': '/dev/mapper/foo-test1'}) included: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-lvmraid.yml for /cache/rhel-8-y.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'ext4', 'mount_options': 'defaults', 'mount_point': '/opt/test2', 'name': 'test2', 'raid_level': None, 'size': '25%', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test2', '_raw_device': '/dev/mapper/foo-test2', '_mount_id': '/dev/mapper/foo-test2', '_kernel_device': '/dev/dm-0', '_raw_kernel_device': '/dev/dm-0'}) TASK [Get information about LVM RAID] ****************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-lvmraid.yml:3 Thursday 21 July 2022 19:07:00 +0000 (0:00:00.055) 0:02:18.131 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is LVM RAID] ******************************************* task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-lvmraid.yml:8 Thursday 21 July 2022 19:07:00 +0000 (0:00:00.027) 0:02:18.158 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-lvmraid.yml:12 Thursday 21 July 2022 19:07:00 +0000 (0:00:00.029) 0:02:18.188 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about LVM RAID] ****************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-lvmraid.yml:3 Thursday 21 July 2022 19:07:00 +0000 (0:00:00.029) 0:02:18.218 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is LVM RAID] ******************************************* task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-lvmraid.yml:8 Thursday 21 July 2022 19:07:00 +0000 (0:00:00.031) 0:02:18.249 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-lvmraid.yml:12 Thursday 21 July 2022 19:07:00 +0000 (0:00:00.027) 0:02:18.277 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check Thin Pools] ******************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:56 Thursday 21 July 2022 19:07:00 +0000 (0:00:00.032) 0:02:18.309 ********* included: /tmp/tmpmnqeja9w/tests/storage/verify-pool-members-thin.yml for /cache/rhel-8-y.qcow2.snap TASK [Validate pool member thinpool settings] ********************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-members-thin.yml:1 Thursday 21 July 2022 19:07:00 +0000 (0:00:00.052) 0:02:18.362 ********* included: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-thin.yml for /cache/rhel-8-y.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'name': 'test1', 'raid_level': None, 'size': '60%', 'state': 'absent', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test1', '_raw_device': '/dev/mapper/foo-test1', '_mount_id': '/dev/mapper/foo-test1'}) included: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-thin.yml for /cache/rhel-8-y.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'ext4', 'mount_options': 'defaults', 'mount_point': '/opt/test2', 'name': 'test2', 'raid_level': None, 'size': '25%', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test2', '_raw_device': '/dev/mapper/foo-test2', '_mount_id': '/dev/mapper/foo-test2', '_kernel_device': '/dev/dm-0', '_raw_kernel_device': '/dev/dm-0'}) TASK [Get information about thinpool] ****************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-thin.yml:3 Thursday 21 July 2022 19:07:00 +0000 (0:00:00.049) 0:02:18.411 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-thin.yml:8 Thursday 21 July 2022 19:07:00 +0000 (0:00:00.023) 0:02:18.435 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-thin.yml:13 Thursday 21 July 2022 19:07:00 +0000 (0:00:00.023) 0:02:18.458 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-thin.yml:17 Thursday 21 July 2022 19:07:00 +0000 (0:00:00.022) 0:02:18.480 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about thinpool] ****************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-thin.yml:3 Thursday 21 July 2022 19:07:00 +0000 (0:00:00.022) 0:02:18.503 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-thin.yml:8 Thursday 21 July 2022 19:07:00 +0000 (0:00:00.025) 0:02:18.528 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-thin.yml:13 Thursday 21 July 2022 19:07:01 +0000 (0:00:00.024) 0:02:18.553 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-thin.yml:17 Thursday 21 July 2022 19:07:01 +0000 (0:00:00.022) 0:02:18.575 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check member encryption] ************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:59 Thursday 21 July 2022 19:07:01 +0000 (0:00:00.022) 0:02:18.598 ********* included: /tmp/tmpmnqeja9w/tests/storage/verify-pool-members-encryption.yml for /cache/rhel-8-y.qcow2.snap TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-members-encryption.yml:4 Thursday 21 July 2022 19:07:01 +0000 (0:00:00.050) 0:02:18.648 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Validate pool member LUKS settings] ************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-members-encryption.yml:8 Thursday 21 July 2022 19:07:01 +0000 (0:00:00.105) 0:02:18.754 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => (item=/dev/sda) => { "_storage_test_pool_member_path": "/dev/sda", "ansible_loop_var": "_storage_test_pool_member_path", "changed": false, "skip_reason": "Conditional result was False" } TASK [Validate pool member crypttab entries] *********************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-members-encryption.yml:15 Thursday 21 July 2022 19:07:01 +0000 (0:00:00.027) 0:02:18.782 ********* included: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-crypttab.yml for /cache/rhel-8-y.qcow2.snap => (item=/dev/sda) TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-crypttab.yml:1 Thursday 21 July 2022 19:07:01 +0000 (0:00:00.083) 0:02:18.865 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": [] }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-crypttab.yml:4 Thursday 21 July 2022 19:07:01 +0000 (0:00:00.052) 0:02:18.918 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-crypttab.yml:9 Thursday 21 July 2022 19:07:01 +0000 (0:00:00.053) 0:02:18.971 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-crypttab.yml:15 Thursday 21 July 2022 19:07:01 +0000 (0:00:00.037) 0:02:19.009 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-crypttab.yml:21 Thursday 21 July 2022 19:07:01 +0000 (0:00:00.037) 0:02:19.046 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-crypttab.yml:27 Thursday 21 July 2022 19:07:01 +0000 (0:00:00.038) 0:02:19.084 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-members-encryption.yml:22 Thursday 21 July 2022 19:07:01 +0000 (0:00:00.040) 0:02:19.125 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_crypttab_key_file": null }, "changed": false } TASK [Check VDO] *************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:62 Thursday 21 July 2022 19:07:01 +0000 (0:00:00.036) 0:02:19.161 ********* included: /tmp/tmpmnqeja9w/tests/storage/verify-pool-members-vdo.yml for /cache/rhel-8-y.qcow2.snap TASK [Validate pool member VDO settings] *************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-members-vdo.yml:1 Thursday 21 July 2022 19:07:01 +0000 (0:00:00.048) 0:02:19.210 ********* included: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-vdo.yml for /cache/rhel-8-y.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'name': 'test1', 'raid_level': None, 'size': '60%', 'state': 'absent', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test1', '_raw_device': '/dev/mapper/foo-test1', '_mount_id': '/dev/mapper/foo-test1'}) included: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-vdo.yml for /cache/rhel-8-y.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'ext4', 'mount_options': 'defaults', 'mount_point': '/opt/test2', 'name': 'test2', 'raid_level': None, 'size': '25%', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test2', '_raw_device': '/dev/mapper/foo-test2', '_mount_id': '/dev/mapper/foo-test2', '_kernel_device': '/dev/dm-0', '_raw_kernel_device': '/dev/dm-0'}) TASK [get information about VDO deduplication] ********************************* task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-vdo.yml:3 Thursday 21 July 2022 19:07:01 +0000 (0:00:00.054) 0:02:19.264 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-vdo.yml:8 Thursday 21 July 2022 19:07:01 +0000 (0:00:00.023) 0:02:19.288 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-vdo.yml:11 Thursday 21 July 2022 19:07:01 +0000 (0:00:00.026) 0:02:19.314 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-vdo.yml:16 Thursday 21 July 2022 19:07:01 +0000 (0:00:00.025) 0:02:19.340 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-vdo.yml:21 Thursday 21 July 2022 19:07:01 +0000 (0:00:00.026) 0:02:19.366 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-vdo.yml:24 Thursday 21 July 2022 19:07:01 +0000 (0:00:00.026) 0:02:19.392 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-vdo.yml:29 Thursday 21 July 2022 19:07:01 +0000 (0:00:00.025) 0:02:19.418 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-vdo.yml:39 Thursday 21 July 2022 19:07:01 +0000 (0:00:00.025) 0:02:19.443 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [get information about VDO deduplication] ********************************* task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-vdo.yml:3 Thursday 21 July 2022 19:07:01 +0000 (0:00:00.036) 0:02:19.480 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-vdo.yml:8 Thursday 21 July 2022 19:07:01 +0000 (0:00:00.023) 0:02:19.504 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-vdo.yml:11 Thursday 21 July 2022 19:07:01 +0000 (0:00:00.024) 0:02:19.529 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-vdo.yml:16 Thursday 21 July 2022 19:07:02 +0000 (0:00:00.023) 0:02:19.552 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-vdo.yml:21 Thursday 21 July 2022 19:07:02 +0000 (0:00:00.025) 0:02:19.578 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-vdo.yml:24 Thursday 21 July 2022 19:07:02 +0000 (0:00:00.026) 0:02:19.605 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-vdo.yml:29 Thursday 21 July 2022 19:07:02 +0000 (0:00:00.026) 0:02:19.632 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-vdo.yml:39 Thursday 21 July 2022 19:07:02 +0000 (0:00:00.024) 0:02:19.656 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:65 Thursday 21 July 2022 19:07:02 +0000 (0:00:00.039) 0:02:19.696 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "__pvs_lvm_len": null, "_storage_test_expected_pv_count": null, "_storage_test_expected_pv_type": null, "_storage_test_pool_pvs": [], "_storage_test_pool_pvs_lvm": [] }, "changed": false } TASK [verify the volumes] ****************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-volumes.yml:3 Thursday 21 July 2022 19:07:02 +0000 (0:00:00.035) 0:02:19.732 ********* included: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume.yml for /cache/rhel-8-y.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'name': 'test1', 'raid_level': None, 'size': '60%', 'state': 'absent', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test1', '_raw_device': '/dev/mapper/foo-test1', '_mount_id': '/dev/mapper/foo-test1'}) included: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume.yml for /cache/rhel-8-y.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'ext4', 'mount_options': 'defaults', 'mount_point': '/opt/test2', 'name': 'test2', 'raid_level': None, 'size': '25%', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test2', '_raw_device': '/dev/mapper/foo-test2', '_mount_id': '/dev/mapper/foo-test2', '_kernel_device': '/dev/dm-0', '_raw_kernel_device': '/dev/dm-0'}) TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume.yml:2 Thursday 21 July 2022 19:07:02 +0000 (0:00:00.047) 0:02:19.779 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_volume_present": false, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume.yml:10 Thursday 21 July 2022 19:07:02 +0000 (0:00:00.051) 0:02:19.830 ********* included: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml for /cache/rhel-8-y.qcow2.snap => (item=mount) included: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-fstab.yml for /cache/rhel-8-y.qcow2.snap => (item=fstab) included: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-fs.yml for /cache/rhel-8-y.qcow2.snap => (item=fs) included: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-device.yml for /cache/rhel-8-y.qcow2.snap => (item=device) included: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml for /cache/rhel-8-y.qcow2.snap => (item=encryption) included: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-md.yml for /cache/rhel-8-y.qcow2.snap => (item=md) included: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml for /cache/rhel-8-y.qcow2.snap => (item=size) included: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-cache.yml for /cache/rhel-8-y.qcow2.snap => (item=cache) TASK [Get expected mount device based on device type] ************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml:6 Thursday 21 July 2022 19:07:02 +0000 (0:00:00.089) 0:02:19.919 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/foo-test1" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml:10 Thursday 21 July 2022 19:07:02 +0000 (0:00:00.054) 0:02:19.974 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_mount_device_matches": [], "storage_test_mount_expected_match_count": "0", "storage_test_mount_point_matches": [], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Verify the current mount state by device] ******************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml:20 Thursday 21 July 2022 19:07:02 +0000 (0:00:00.060) 0:02:20.034 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the current mount state by mount point] *************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml:29 Thursday 21 July 2022 19:07:02 +0000 (0:00:00.025) 0:02:20.060 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify the mount fs type] ************************************************ task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml:37 Thursday 21 July 2022 19:07:02 +0000 (0:00:00.098) 0:02:20.158 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [command] ***************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml:46 Thursday 21 July 2022 19:07:02 +0000 (0:00:00.087) 0:02:20.246 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml:50 Thursday 21 July 2022 19:07:02 +0000 (0:00:00.024) 0:02:20.270 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml:55 Thursday 21 July 2022 19:07:02 +0000 (0:00:00.025) 0:02:20.296 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml:65 Thursday 21 July 2022 19:07:02 +0000 (0:00:00.027) 0:02:20.323 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-fstab.yml:2 Thursday 21 July 2022 19:07:02 +0000 (0:00:00.036) 0:02:20.359 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "0", "storage_test_fstab_expected_mount_options_matches": "0", "storage_test_fstab_expected_mount_point_matches": "0", "storage_test_fstab_id_matches": [], "storage_test_fstab_mount_options_matches": [], "storage_test_fstab_mount_point_matches": [] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-fstab.yml:12 Thursday 21 July 2022 19:07:02 +0000 (0:00:00.062) 0:02:20.422 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the fstab mount point] ******************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-fstab.yml:19 Thursday 21 July 2022 19:07:02 +0000 (0:00:00.026) 0:02:20.448 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-fstab.yml:25 Thursday 21 July 2022 19:07:02 +0000 (0:00:00.055) 0:02:20.504 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-fstab.yml:34 Thursday 21 July 2022 19:07:02 +0000 (0:00:00.039) 0:02:20.543 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-fs.yml:4 Thursday 21 July 2022 19:07:03 +0000 (0:00:00.037) 0:02:20.580 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify fs label] ********************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-fs.yml:10 Thursday 21 July 2022 19:07:03 +0000 (0:00:00.023) 0:02:20.603 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [See whether the device node is present] ********************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-device.yml:4 Thursday 21 July 2022 19:07:03 +0000 (0:00:00.026) 0:02:20.630 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "stat": { "exists": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-device.yml:10 Thursday 21 July 2022 19:07:03 +0000 (0:00:00.394) 0:02:21.025 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about this volume] ********************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-device.yml:15 Thursday 21 July 2022 19:07:03 +0000 (0:00:00.040) 0:02:21.065 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [(1/2) Process volume type (set initial value)] *************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-device.yml:21 Thursday 21 July 2022 19:07:03 +0000 (0:00:00.026) 0:02:21.092 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [(2/2) Process volume type (get RAID value)] ****************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-device.yml:25 Thursday 21 July 2022 19:07:03 +0000 (0:00:00.038) 0:02:21.130 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-device.yml:30 Thursday 21 July 2022 19:07:03 +0000 (0:00:00.026) 0:02:21.156 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Stat the LUKS device, if encrypted] ************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:3 Thursday 21 July 2022 19:07:03 +0000 (0:00:00.025) 0:02:21.182 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:10 Thursday 21 July 2022 19:07:03 +0000 (0:00:00.026) 0:02:21.209 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [Collect LUKS info for this volume] *************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:15 Thursday 21 July 2022 19:07:05 +0000 (0:00:01.348) 0:02:22.557 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:21 Thursday 21 July 2022 19:07:05 +0000 (0:00:00.026) 0:02:22.584 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:27 Thursday 21 July 2022 19:07:05 +0000 (0:00:00.025) 0:02:22.609 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:33 Thursday 21 July 2022 19:07:05 +0000 (0:00:00.024) 0:02:22.634 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:39 Thursday 21 July 2022 19:07:05 +0000 (0:00:00.024) 0:02:22.658 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:44 Thursday 21 July 2022 19:07:05 +0000 (0:00:00.024) 0:02:22.683 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:50 Thursday 21 July 2022 19:07:05 +0000 (0:00:00.027) 0:02:22.711 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:56 Thursday 21 July 2022 19:07:05 +0000 (0:00:00.025) 0:02:22.737 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:62 Thursday 21 July 2022 19:07:05 +0000 (0:00:00.024) 0:02:22.761 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:67 Thursday 21 July 2022 19:07:05 +0000 (0:00:00.055) 0:02:22.816 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:72 Thursday 21 July 2022 19:07:05 +0000 (0:00:00.051) 0:02:22.867 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:78 Thursday 21 July 2022 19:07:05 +0000 (0:00:00.035) 0:02:22.903 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:84 Thursday 21 July 2022 19:07:05 +0000 (0:00:00.034) 0:02:22.938 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:90 Thursday 21 July 2022 19:07:05 +0000 (0:00:00.043) 0:02:22.981 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [get information about RAID] ********************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-md.yml:7 Thursday 21 July 2022 19:07:05 +0000 (0:00:00.035) 0:02:23.017 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-md.yml:13 Thursday 21 July 2022 19:07:05 +0000 (0:00:00.037) 0:02:23.055 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-md.yml:17 Thursday 21 July 2022 19:07:05 +0000 (0:00:00.038) 0:02:23.093 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-md.yml:21 Thursday 21 July 2022 19:07:05 +0000 (0:00:00.075) 0:02:23.168 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-md.yml:25 Thursday 21 July 2022 19:07:05 +0000 (0:00:00.039) 0:02:23.208 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-md.yml:31 Thursday 21 July 2022 19:07:05 +0000 (0:00:00.038) 0:02:23.247 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-md.yml:37 Thursday 21 July 2022 19:07:05 +0000 (0:00:00.037) 0:02:23.284 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the actual size of the volume] ************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:3 Thursday 21 July 2022 19:07:05 +0000 (0:00:00.040) 0:02:23.325 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the requested size of the volume] ********************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:9 Thursday 21 July 2022 19:07:05 +0000 (0:00:00.026) 0:02:23.351 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Establish base value for expected size] ********************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:15 Thursday 21 July 2022 19:07:05 +0000 (0:00:00.037) 0:02:23.389 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:20 Thursday 21 July 2022 19:07:05 +0000 (0:00:00.038) 0:02:23.427 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_expected_size": "2684354560.0" } TASK [debug] ******************************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:25 Thursday 21 July 2022 19:07:05 +0000 (0:00:00.041) 0:02:23.468 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => {} TASK [debug] ******************************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:28 Thursday 21 July 2022 19:07:05 +0000 (0:00:00.038) 0:02:23.507 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => {} TASK [Get the size of parent/pool device] ************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:31 Thursday 21 July 2022 19:07:05 +0000 (0:00:00.039) 0:02:23.546 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:36 Thursday 21 July 2022 19:07:06 +0000 (0:00:00.037) 0:02:23.584 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => {} TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:39 Thursday 21 July 2022 19:07:06 +0000 (0:00:00.046) 0:02:23.631 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:44 Thursday 21 July 2022 19:07:06 +0000 (0:00:00.041) 0:02:23.673 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_actual_size": { "changed": false, "skip_reason": "Conditional result was False", "skipped": true } } TASK [debug] ******************************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:47 Thursday 21 July 2022 19:07:06 +0000 (0:00:00.044) 0:02:23.717 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_expected_size": "2684354560.0" } TASK [assert] ****************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:50 Thursday 21 July 2022 19:07:06 +0000 (0:00:00.040) 0:02:23.757 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about the LV] ******************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-cache.yml:6 Thursday 21 July 2022 19:07:06 +0000 (0:00:00.029) 0:02:23.786 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-cache.yml:14 Thursday 21 July 2022 19:07:06 +0000 (0:00:00.027) 0:02:23.813 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check segment type] ****************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-cache.yml:17 Thursday 21 July 2022 19:07:06 +0000 (0:00:00.025) 0:02:23.838 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-cache.yml:22 Thursday 21 July 2022 19:07:06 +0000 (0:00:00.026) 0:02:23.865 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the requested cache size] ****************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-cache.yml:26 Thursday 21 July 2022 19:07:06 +0000 (0:00:00.027) 0:02:23.893 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-cache.yml:32 Thursday 21 July 2022 19:07:06 +0000 (0:00:00.025) 0:02:23.918 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-cache.yml:36 Thursday 21 July 2022 19:07:06 +0000 (0:00:00.029) 0:02:23.947 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume.yml:16 Thursday 21 July 2022 19:07:06 +0000 (0:00:00.027) 0:02:23.975 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume.yml:2 Thursday 21 July 2022 19:07:06 +0000 (0:00:00.043) 0:02:24.019 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_volume_present": true, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume.yml:10 Thursday 21 July 2022 19:07:06 +0000 (0:00:00.055) 0:02:24.074 ********* included: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml for /cache/rhel-8-y.qcow2.snap => (item=mount) included: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-fstab.yml for /cache/rhel-8-y.qcow2.snap => (item=fstab) included: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-fs.yml for /cache/rhel-8-y.qcow2.snap => (item=fs) included: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-device.yml for /cache/rhel-8-y.qcow2.snap => (item=device) included: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml for /cache/rhel-8-y.qcow2.snap => (item=encryption) included: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-md.yml for /cache/rhel-8-y.qcow2.snap => (item=md) included: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml for /cache/rhel-8-y.qcow2.snap => (item=size) included: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-cache.yml for /cache/rhel-8-y.qcow2.snap => (item=cache) TASK [Get expected mount device based on device type] ************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml:6 Thursday 21 July 2022 19:07:06 +0000 (0:00:00.082) 0:02:24.157 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/foo-test2" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml:10 Thursday 21 July 2022 19:07:06 +0000 (0:00:00.041) 0:02:24.198 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_mount_device_matches": [ { "block_available": 588749, "block_size": 4096, "block_total": 625618, "block_used": 36869, "device": "/dev/mapper/foo-test2", "fstype": "ext4", "inode_available": 163829, "inode_total": 163840, "inode_used": 11, "mount": "/opt/test2", "options": "rw,seclabel,relatime", "size_available": 2411515904, "size_total": 2562531328, "uuid": "64454f89-f2cd-4a8b-958f-4c900f72cf79" } ], "storage_test_mount_expected_match_count": "1", "storage_test_mount_point_matches": [ { "block_available": 588749, "block_size": 4096, "block_total": 625618, "block_used": 36869, "device": "/dev/mapper/foo-test2", "fstype": "ext4", "inode_available": 163829, "inode_total": 163840, "inode_used": 11, "mount": "/opt/test2", "options": "rw,seclabel,relatime", "size_available": 2411515904, "size_total": 2562531328, "uuid": "64454f89-f2cd-4a8b-958f-4c900f72cf79" } ], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Verify the current mount state by device] ******************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml:20 Thursday 21 July 2022 19:07:06 +0000 (0:00:00.056) 0:02:24.254 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify the current mount state by mount point] *************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml:29 Thursday 21 July 2022 19:07:06 +0000 (0:00:00.047) 0:02:24.302 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify the mount fs type] ************************************************ task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml:37 Thursday 21 July 2022 19:07:06 +0000 (0:00:00.047) 0:02:24.349 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [command] ***************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml:46 Thursday 21 July 2022 19:07:06 +0000 (0:00:00.050) 0:02:24.399 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml:50 Thursday 21 July 2022 19:07:06 +0000 (0:00:00.023) 0:02:24.423 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml:55 Thursday 21 July 2022 19:07:06 +0000 (0:00:00.023) 0:02:24.446 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml:65 Thursday 21 July 2022 19:07:06 +0000 (0:00:00.022) 0:02:24.469 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-fstab.yml:2 Thursday 21 July 2022 19:07:06 +0000 (0:00:00.035) 0:02:24.504 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "1", "storage_test_fstab_expected_mount_options_matches": "1", "storage_test_fstab_expected_mount_point_matches": "1", "storage_test_fstab_id_matches": [ "/dev/mapper/foo-test2 " ], "storage_test_fstab_mount_options_matches": [ " /opt/test2 ext4 defaults " ], "storage_test_fstab_mount_point_matches": [ " /opt/test2 " ] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-fstab.yml:12 Thursday 21 July 2022 19:07:07 +0000 (0:00:00.058) 0:02:24.562 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify the fstab mount point] ******************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-fstab.yml:19 Thursday 21 July 2022 19:07:07 +0000 (0:00:00.088) 0:02:24.651 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-fstab.yml:25 Thursday 21 July 2022 19:07:07 +0000 (0:00:00.138) 0:02:24.790 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-fstab.yml:34 Thursday 21 July 2022 19:07:07 +0000 (0:00:00.038) 0:02:24.829 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-fs.yml:4 Thursday 21 July 2022 19:07:07 +0000 (0:00:00.041) 0:02:24.871 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify fs label] ********************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-fs.yml:10 Thursday 21 July 2022 19:07:07 +0000 (0:00:00.040) 0:02:24.912 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [See whether the device node is present] ********************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-device.yml:4 Thursday 21 July 2022 19:07:07 +0000 (0:00:00.045) 0:02:24.957 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "stat": { "atime": 1658430383.5851166, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1658430383.5851166, "dev": 6, "device_type": 64768, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 59192, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/symlink", "mode": "0660", "mtime": 1658430383.5851166, "nlink": 1, "path": "/dev/mapper/foo-test2", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-device.yml:10 Thursday 21 July 2022 19:07:07 +0000 (0:00:00.387) 0:02:25.345 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about this volume] ********************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-device.yml:15 Thursday 21 July 2022 19:07:07 +0000 (0:00:00.040) 0:02:25.385 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [(1/2) Process volume type (set initial value)] *************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-device.yml:21 Thursday 21 July 2022 19:07:07 +0000 (0:00:00.043) 0:02:25.428 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [(2/2) Process volume type (get RAID value)] ****************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-device.yml:25 Thursday 21 July 2022 19:07:07 +0000 (0:00:00.036) 0:02:25.465 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-device.yml:30 Thursday 21 July 2022 19:07:07 +0000 (0:00:00.022) 0:02:25.488 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Stat the LUKS device, if encrypted] ************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:3 Thursday 21 July 2022 19:07:07 +0000 (0:00:00.037) 0:02:25.525 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:10 Thursday 21 July 2022 19:07:07 +0000 (0:00:00.024) 0:02:25.550 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [Collect LUKS info for this volume] *************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:15 Thursday 21 July 2022 19:07:09 +0000 (0:00:01.265) 0:02:26.815 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:21 Thursday 21 July 2022 19:07:09 +0000 (0:00:00.024) 0:02:26.839 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:27 Thursday 21 July 2022 19:07:09 +0000 (0:00:00.024) 0:02:26.863 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:33 Thursday 21 July 2022 19:07:09 +0000 (0:00:00.054) 0:02:26.918 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:39 Thursday 21 July 2022 19:07:09 +0000 (0:00:00.024) 0:02:26.943 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:44 Thursday 21 July 2022 19:07:09 +0000 (0:00:00.024) 0:02:26.967 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:50 Thursday 21 July 2022 19:07:09 +0000 (0:00:00.024) 0:02:26.992 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:56 Thursday 21 July 2022 19:07:09 +0000 (0:00:00.025) 0:02:27.018 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:62 Thursday 21 July 2022 19:07:09 +0000 (0:00:00.024) 0:02:27.042 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:67 Thursday 21 July 2022 19:07:09 +0000 (0:00:00.052) 0:02:27.094 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:72 Thursday 21 July 2022 19:07:09 +0000 (0:00:00.062) 0:02:27.157 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:78 Thursday 21 July 2022 19:07:09 +0000 (0:00:00.039) 0:02:27.196 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:84 Thursday 21 July 2022 19:07:09 +0000 (0:00:00.039) 0:02:27.235 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:90 Thursday 21 July 2022 19:07:09 +0000 (0:00:00.039) 0:02:27.275 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [get information about RAID] ********************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-md.yml:7 Thursday 21 July 2022 19:07:09 +0000 (0:00:00.040) 0:02:27.315 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-md.yml:13 Thursday 21 July 2022 19:07:09 +0000 (0:00:00.038) 0:02:27.354 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-md.yml:17 Thursday 21 July 2022 19:07:09 +0000 (0:00:00.038) 0:02:27.392 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-md.yml:21 Thursday 21 July 2022 19:07:09 +0000 (0:00:00.037) 0:02:27.429 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-md.yml:25 Thursday 21 July 2022 19:07:09 +0000 (0:00:00.039) 0:02:27.469 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-md.yml:31 Thursday 21 July 2022 19:07:10 +0000 (0:00:00.096) 0:02:27.566 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-md.yml:37 Thursday 21 July 2022 19:07:10 +0000 (0:00:00.038) 0:02:27.604 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the actual size of the volume] ************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:3 Thursday 21 July 2022 19:07:10 +0000 (0:00:00.038) 0:02:27.642 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "bytes": 2684354560, "changed": false, "lvm": "2g", "parted": "2GiB", "size": "2 GiB" } TASK [parse the requested size of the volume] ********************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:9 Thursday 21 July 2022 19:07:10 +0000 (0:00:00.377) 0:02:28.019 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Establish base value for expected size] ********************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:15 Thursday 21 July 2022 19:07:10 +0000 (0:00:00.040) 0:02:28.060 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:20 Thursday 21 July 2022 19:07:10 +0000 (0:00:00.047) 0:02:28.107 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_expected_size": "2684354560.0" } TASK [debug] ******************************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:25 Thursday 21 July 2022 19:07:10 +0000 (0:00:00.040) 0:02:28.148 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_pool": { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "60%", "state": "absent", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test2", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test2", "_raw_device": "/dev/mapper/foo-test2", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "25%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } } TASK [debug] ******************************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:28 Thursday 21 July 2022 19:07:10 +0000 (0:00:00.053) 0:02:28.202 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_blkinfo": { "changed": false, "failed": false, "info": { "/dev/mapper/foo-test2": { "fstype": "ext4", "label": "", "name": "/dev/mapper/foo-test2", "size": "2.5G", "type": "lvm", "uuid": "64454f89-f2cd-4a8b-958f-4c900f72cf79" }, "/dev/nvme0n1": { "fstype": "", "label": "", "name": "/dev/nvme0n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme1n1": { "fstype": "", "label": "", "name": "/dev/nvme1n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme2n1": { "fstype": "", "label": "", "name": "/dev/nvme2n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sda": { "fstype": "LVM2_member", "label": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "HA17fj-P3dB-xyZl-9Zru-ZzXr-ULh6-93Sgui" }, "/dev/sdb": { "fstype": "", "label": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc": { "fstype": "", "label": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sr0": { "fstype": "iso9660", "label": "cidata", "name": "/dev/sr0", "size": "364K", "type": "rom", "uuid": "2022-07-21-19-04-29-00" }, "/dev/vda": { "fstype": "", "label": "", "name": "/dev/vda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vda1": { "fstype": "", "label": "", "name": "/dev/vda1", "size": "1M", "type": "partition", "uuid": "" }, "/dev/vda2": { "fstype": "vfat", "label": "", "name": "/dev/vda2", "size": "100M", "type": "partition", "uuid": "7B77-95E7" }, "/dev/vda3": { "fstype": "xfs", "label": "root", "name": "/dev/vda3", "size": "9.9G", "type": "partition", "uuid": "58dca2db-4ccd-4171-a017-dd53ad8c1c54" }, "/dev/vdb": { "fstype": "", "label": "", "name": "/dev/vdb", "size": "2G", "type": "disk", "uuid": "" }, "/dev/vdb1": { "fstype": "ext4", "label": "yumcache", "name": "/dev/vdb1", "size": "2G", "type": "partition", "uuid": "4d1ed9a1-5608-4ff8-ad89-21d95caa66fa" }, "/dev/vdc": { "fstype": "", "label": "", "name": "/dev/vdc", "size": "2G", "type": "disk", "uuid": "" }, "/dev/vdc1": { "fstype": "ext4", "label": "yumvarlib", "name": "/dev/vdc1", "size": "2G", "type": "partition", "uuid": "a7abcc97-edc5-476c-8ccd-694b4621fd28" }, "/dev/vdd": { "fstype": "", "label": "", "name": "/dev/vdd", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vde": { "fstype": "", "label": "", "name": "/dev/vde", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vdf": { "fstype": "", "label": "", "name": "/dev/vdf", "size": "10G", "type": "disk", "uuid": "" } } } } TASK [Get the size of parent/pool device] ************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:31 Thursday 21 July 2022 19:07:10 +0000 (0:00:00.054) 0:02:28.257 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "bytes": 10737418240, "changed": false, "lvm": "10g", "parted": "10GiB", "size": "10 GiB" } TASK [debug] ******************************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:36 Thursday 21 July 2022 19:07:11 +0000 (0:00:00.375) 0:02:28.633 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_pool_size": { "bytes": 10737418240, "changed": false, "failed": false, "lvm": "10g", "parted": "10GiB", "size": "10 GiB" } } TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:39 Thursday 21 July 2022 19:07:11 +0000 (0:00:00.052) 0:02:28.686 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_expected_size": "2684354560.0" }, "changed": false } TASK [debug] ******************************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:44 Thursday 21 July 2022 19:07:11 +0000 (0:00:00.052) 0:02:28.738 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_actual_size": { "bytes": 2684354560, "changed": false, "failed": false, "lvm": "2g", "parted": "2GiB", "size": "2 GiB" } } TASK [debug] ******************************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:47 Thursday 21 July 2022 19:07:11 +0000 (0:00:00.038) 0:02:28.776 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_expected_size": "2684354560.0" } TASK [assert] ****************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:50 Thursday 21 July 2022 19:07:11 +0000 (0:00:00.039) 0:02:28.816 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Get information about the LV] ******************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-cache.yml:6 Thursday 21 July 2022 19:07:11 +0000 (0:00:00.053) 0:02:28.869 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "cmd": [ "lvs", "--noheadings", "--nameprefixes", "--units=b", "--nosuffix", "--unquoted", "-o", "name,attr,cache_total_blocks,chunk_size,segtype", "foo/test2" ], "delta": "0:00:00.041521", "end": "2022-07-21 15:07:11.803871", "rc": 0, "start": "2022-07-21 15:07:11.762350" } STDOUT: LVM2_LV_NAME=test2 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-cache.yml:14 Thursday 21 July 2022 19:07:11 +0000 (0:00:00.421) 0:02:29.291 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_lv_segtype": [ "linear" ] }, "changed": false } TASK [check segment type] ****************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-cache.yml:17 Thursday 21 July 2022 19:07:11 +0000 (0:00:00.050) 0:02:29.341 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-cache.yml:22 Thursday 21 July 2022 19:07:11 +0000 (0:00:00.048) 0:02:29.390 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the requested cache size] ****************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-cache.yml:26 Thursday 21 July 2022 19:07:11 +0000 (0:00:00.036) 0:02:29.426 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-cache.yml:32 Thursday 21 July 2022 19:07:11 +0000 (0:00:00.041) 0:02:29.467 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-cache.yml:36 Thursday 21 July 2022 19:07:11 +0000 (0:00:00.039) 0:02:29.506 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume.yml:16 Thursday 21 July 2022 19:07:11 +0000 (0:00:00.036) 0:02:29.543 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [Verify the volumes with no pool were correctly managed] ****************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-role-results.yml:43 Thursday 21 July 2022 19:07:12 +0000 (0:00:00.033) 0:02:29.576 ********* TASK [Clean up variable namespace] ********************************************* task path: /tmp/tmpmnqeja9w/tests/storage/verify-role-results.yml:53 Thursday 21 July 2022 19:07:12 +0000 (0:00:00.021) 0:02:29.598 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_blkinfo": null, "storage_test_crypttab": null, "storage_test_fstab": null }, "changed": false } TASK [Get the size of test2 volume again] ************************************** task path: /tmp/tmpmnqeja9w/tests/storage/tests_lvm_percent_size.yml:138 Thursday 21 July 2022 19:07:12 +0000 (0:00:00.035) 0:02:29.633 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "cmd": [ "lsblk", "--noheadings", "-o", "SIZE", "/dev/mapper/foo-test2" ], "delta": "0:00:00.003672", "end": "2022-07-21 15:07:12.524276", "rc": 0, "start": "2022-07-21 15:07:12.520604" } STDOUT: 2.5G TASK [Verify that removing test1 didn't cause a change in test2 size] ********** task path: /tmp/tmpmnqeja9w/tests/storage/tests_lvm_percent_size.yml:143 Thursday 21 July 2022 19:07:12 +0000 (0:00:00.368) 0:02:30.002 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Grow test2 using a percentage-based size spec] *************************** task path: /tmp/tmpmnqeja9w/tests/storage/tests_lvm_percent_size.yml:147 Thursday 21 July 2022 19:07:12 +0000 (0:00:00.070) 0:02:30.072 ********* TASK [fedora.linux_system_roles.storage : set platform/version specific variables] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2 Thursday 21 July 2022 19:07:12 +0000 (0:00:00.063) 0:02:30.135 ********* included: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for /cache/rhel-8-y.qcow2.snap TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2 Thursday 21 July 2022 19:07:12 +0000 (0:00:00.063) 0:02:30.199 ********* ok: [/cache/rhel-8-y.qcow2.snap] TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7 Thursday 21 July 2022 19:07:13 +0000 (0:00:00.511) 0:02:30.711 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [/cache/rhel-8-y.qcow2.snap] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [/cache/rhel-8-y.qcow2.snap] => (item=RedHat_8.yml) => { "ansible_facts": { "blivet_package_list": [ "python3-blivet", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap", "vdo", "kmod-kvdo", "xfsprogs" ] }, "ansible_included_var_files": [ "/tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/vars/RedHat_8.yml" ], "ansible_loop_var": "item", "changed": false, "item": "RedHat_8.yml" } skipping: [/cache/rhel-8-y.qcow2.snap] => (item=RedHat_8.7.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat_8.7.yml", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : define an empty list of pools to be used in testing] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5 Thursday 21 July 2022 19:07:13 +0000 (0:00:00.061) 0:02:30.773 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : define an empty list of volumes to be used in testing] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9 Thursday 21 July 2022 19:07:13 +0000 (0:00:00.034) 0:02:30.807 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : include the appropriate provider tasks] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13 Thursday 21 July 2022 19:07:13 +0000 (0:00:00.035) 0:02:30.843 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount included: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for /cache/rhel-8-y.qcow2.snap TASK [fedora.linux_system_roles.storage : get a list of rpm packages installed on host machine] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2 Thursday 21 July 2022 19:07:13 +0000 (0:00:00.056) 0:02:30.900 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : make sure blivet is available] ******* task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:7 Thursday 21 July 2022 19:07:13 +0000 (0:00:00.021) 0:02:30.921 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [fedora.linux_system_roles.storage : show storage_pools] ****************** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:13 Thursday 21 July 2022 19:07:14 +0000 (0:00:01.326) 0:02:32.248 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_pools": [ { "disks": [ "sda" ], "name": "foo", "state": "present", "volumes": [ { "mount_point": "/opt/test2", "name": "test2", "size": "50%" } ] } ] } TASK [fedora.linux_system_roles.storage : show storage_volumes] **************** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:18 Thursday 21 July 2022 19:07:14 +0000 (0:00:00.040) 0:02:32.289 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [fedora.linux_system_roles.storage : get required packages] *************** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:23 Thursday 21 July 2022 19:07:14 +0000 (0:00:00.041) 0:02:32.331 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [ "lvm2" ], "pools": [], "volumes": [] } TASK [fedora.linux_system_roles.storage : enable copr repositories if needed] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:35 Thursday 21 July 2022 19:07:16 +0000 (0:00:01.537) 0:02:33.868 ********* included: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml for /cache/rhel-8-y.qcow2.snap TASK [fedora.linux_system_roles.storage : check if the COPR support packages should be installed] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:2 Thursday 21 July 2022 19:07:16 +0000 (0:00:00.055) 0:02:33.924 ********* TASK [fedora.linux_system_roles.storage : make sure COPR support packages are present] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:13 Thursday 21 July 2022 19:07:16 +0000 (0:00:00.040) 0:02:33.965 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : enable COPRs] ************************ task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:18 Thursday 21 July 2022 19:07:16 +0000 (0:00:00.048) 0:02:34.013 ********* TASK [fedora.linux_system_roles.storage : make sure required packages are installed] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:41 Thursday 21 July 2022 19:07:16 +0000 (0:00:00.037) 0:02:34.051 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [fedora.linux_system_roles.storage : get service facts] ******************* task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:47 Thursday 21 July 2022 19:07:17 +0000 (0:00:01.373) 0:02:35.424 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "services": { "NetworkManager-dispatcher.service": { "name": "NetworkManager-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "NetworkManager-wait-online.service": { "name": "NetworkManager-wait-online.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "NetworkManager.service": { "name": "NetworkManager.service", "source": "systemd", "state": "running", "status": "enabled" }, "arp-ethers.service": { "name": "arp-ethers.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "auditd.service": { "name": "auditd.service", "source": "systemd", "state": "running", "status": "enabled" }, "auth-rpcgss-module.service": { "name": "auth-rpcgss-module.service", "source": "systemd", "state": "stopped", "status": "static" }, "autovt@.service": { "name": "autovt@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "blivet.service": { "name": "blivet.service", "source": "systemd", "state": "inactive", "status": "static" }, "blk-availability.service": { "name": "blk-availability.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "chrony-dnssrv@.service": { "name": "chrony-dnssrv@.service", "source": "systemd", "state": "unknown", "status": "static" }, "chrony-wait.service": { "name": "chrony-wait.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd.service": { "name": "chronyd.service", "source": "systemd", "state": "running", "status": "enabled" }, "cloud-config.service": { "name": "cloud-config.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-final.service": { "name": "cloud-final.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init-hotplugd.service": { "name": "cloud-init-hotplugd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cloud-init-local.service": { "name": "cloud-init-local.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init.service": { "name": "cloud-init.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cockpit-motd.service": { "name": "cockpit-motd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cockpit-wsinstance-http.service": { "name": "cockpit-wsinstance-http.service", "source": "systemd", "state": "inactive", "status": "static" }, "cockpit-wsinstance-https-factory@.service": { "name": "cockpit-wsinstance-https-factory@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cockpit-wsinstance-https@.service": { "name": "cockpit-wsinstance-https@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cockpit.service": { "name": "cockpit.service", "source": "systemd", "state": "inactive", "status": "static" }, "console-getty.service": { "name": "console-getty.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "container-getty@.service": { "name": "container-getty@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cpupower.service": { "name": "cpupower.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "crond.service": { "name": "crond.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-org.freedesktop.hostname1.service": { "name": "dbus-org.freedesktop.hostname1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.locale1.service": { "name": "dbus-org.freedesktop.locale1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.login1.service": { "name": "dbus-org.freedesktop.login1.service", "source": "systemd", "state": "active", "status": "static" }, "dbus-org.freedesktop.nm-dispatcher.service": { "name": "dbus-org.freedesktop.nm-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "dbus-org.freedesktop.portable1.service": { "name": "dbus-org.freedesktop.portable1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.timedate1.service": { "name": "dbus-org.freedesktop.timedate1.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "dbus.service": { "name": "dbus.service", "source": "systemd", "state": "running", "status": "static" }, "dbxtool.service": { "name": "dbxtool.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "debug-shell.service": { "name": "debug-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "display-manager.service": { "name": "display-manager.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "dm-event.service": { "name": "dm-event.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-makecache.service": { "name": "dnf-makecache.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-cmdline.service": { "name": "dracut-cmdline.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-initqueue.service": { "name": "dracut-initqueue.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-mount.service": { "name": "dracut-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-mount.service": { "name": "dracut-pre-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-pivot.service": { "name": "dracut-pre-pivot.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-trigger.service": { "name": "dracut-pre-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-udev.service": { "name": "dracut-pre-udev.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown.service": { "name": "dracut-shutdown.service", "source": "systemd", "state": "stopped", "status": "static" }, "emergency.service": { "name": "emergency.service", "source": "systemd", "state": "stopped", "status": "static" }, "fcoe.service": { "name": "fcoe.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "fstrim.service": { "name": "fstrim.service", "source": "systemd", "state": "inactive", "status": "static" }, "getty@.service": { "name": "getty@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "getty@tty1.service": { "name": "getty@tty1.service", "source": "systemd", "state": "running", "status": "active" }, "grub-boot-indeterminate.service": { "name": "grub-boot-indeterminate.service", "source": "systemd", "state": "inactive", "status": "static" }, "gssproxy.service": { "name": "gssproxy.service", "source": "systemd", "state": "running", "status": "disabled" }, "halt-local.service": { "name": "halt-local.service", "source": "systemd", "state": "inactive", "status": "static" }, "hv_kvp_daemon.service": { "name": "hv_kvp_daemon.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "import-state.service": { "name": "import-state.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "initrd-cleanup.service": { "name": "initrd-cleanup.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-parse-etc.service": { "name": "initrd-parse-etc.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-switch-root.service": { "name": "initrd-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-udevadm-cleanup-db.service": { "name": "initrd-udevadm-cleanup-db.service", "source": "systemd", "state": "stopped", "status": "static" }, "insights-client-boot.service": { "name": "insights-client-boot.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "insights-client-results.service": { "name": "insights-client-results.service", "source": "systemd", "state": "inactive", "status": "static" }, "insights-client.service": { "name": "insights-client.service", "source": "systemd", "state": "inactive", "status": "static" }, "irqbalance.service": { "name": "irqbalance.service", "source": "systemd", "state": "running", "status": "enabled" }, "iscsi-shutdown.service": { "name": "iscsi-shutdown.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "iscsi.service": { "name": "iscsi.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "iscsid.service": { "name": "iscsid.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "kdump.service": { "name": "kdump.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "kmod-static-nodes.service": { "name": "kmod-static-nodes.service", "source": "systemd", "state": "stopped", "status": "static" }, "kvm_stat.service": { "name": "kvm_stat.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "ldconfig.service": { "name": "ldconfig.service", "source": "systemd", "state": "stopped", "status": "static" }, "loadmodules.service": { "name": "loadmodules.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "lvm2-activation-early.service": { "name": "lvm2-activation-early.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "lvm2-activation.service": { "name": "lvm2-activation.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "lvm2-lvmpolld.service": { "name": "lvm2-lvmpolld.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-monitor.service": { "name": "lvm2-monitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "lvm2-pvscan@.service": { "name": "lvm2-pvscan@.service", "source": "systemd", "state": "unknown", "status": "static" }, "lvm2-pvscan@8:0.service": { "name": "lvm2-pvscan@8:0.service", "source": "systemd", "state": "stopped", "status": "active" }, "man-db-cache-update.service": { "name": "man-db-cache-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "man-db-restart-cache-update.service": { "name": "man-db-restart-cache-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "mdadm-grow-continue@.service": { "name": "mdadm-grow-continue@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdadm-last-resort@.service": { "name": "mdadm-last-resort@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdcheck_continue.service": { "name": "mdcheck_continue.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdcheck_start.service": { "name": "mdcheck_start.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmon@.service": { "name": "mdmon@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdmonitor-oneshot.service": { "name": "mdmonitor-oneshot.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmonitor.service": { "name": "mdmonitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "messagebus.service": { "name": "messagebus.service", "source": "systemd", "state": "active", "status": "static" }, "microcode.service": { "name": "microcode.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "multipathd.service": { "name": "multipathd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "ndctl-monitor.service": { "name": "ndctl-monitor.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "network.service": { "name": "network.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "nfs-blkmap.service": { "name": "nfs-blkmap.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-convert.service": { "name": "nfs-convert.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-idmapd.service": { "name": "nfs-idmapd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-mountd.service": { "name": "nfs-mountd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-server.service": { "name": "nfs-server.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-utils.service": { "name": "nfs-utils.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfsdcld.service": { "name": "nfsdcld.service", "source": "systemd", "state": "stopped", "status": "static" }, "nis-domainname.service": { "name": "nis-domainname.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "ntpd.service": { "name": "ntpd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ntpdate.service": { "name": "ntpdate.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "oddjobd.service": { "name": "oddjobd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "packagekit-offline-update.service": { "name": "packagekit-offline-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "packagekit.service": { "name": "packagekit.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-quit-wait.service": { "name": "plymouth-quit-wait.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "plymouth-start.service": { "name": "plymouth-start.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "polkit.service": { "name": "polkit.service", "source": "systemd", "state": "running", "status": "static" }, "qemu-guest-agent.service": { "name": "qemu-guest-agent.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "quotaon.service": { "name": "quotaon.service", "source": "systemd", "state": "inactive", "status": "static" }, "rbdmap.service": { "name": "rbdmap.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "rc-local.service": { "name": "rc-local.service", "source": "systemd", "state": "stopped", "status": "static" }, "rdisc.service": { "name": "rdisc.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rescue.service": { "name": "rescue.service", "source": "systemd", "state": "stopped", "status": "static" }, "rhcd.service": { "name": "rhcd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhsm-facts.service": { "name": "rhsm-facts.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhsm.service": { "name": "rhsm.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhsmcertd.service": { "name": "rhsmcertd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpc-gssd.service": { "name": "rpc-gssd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd-notify.service": { "name": "rpc-statd-notify.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd.service": { "name": "rpc-statd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpcbind.service": { "name": "rpcbind.service", "source": "systemd", "state": "running", "status": "enabled" }, "rsyslog.service": { "name": "rsyslog.service", "source": "systemd", "state": "running", "status": "enabled" }, "selinux-autorelabel-mark.service": { "name": "selinux-autorelabel-mark.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "selinux-autorelabel.service": { "name": "selinux-autorelabel.service", "source": "systemd", "state": "inactive", "status": "static" }, "serial-getty@.service": { "name": "serial-getty@.service", "source": "systemd", "state": "unknown", "status": "indirect" }, "serial-getty@ttyS0.service": { "name": "serial-getty@ttyS0.service", "source": "systemd", "state": "running", "status": "active" }, "snapd.seeded.service": { "name": "snapd.seeded.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sntp.service": { "name": "sntp.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sshd-keygen.service": { "name": "sshd-keygen.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sshd-keygen@.service": { "name": "sshd-keygen@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@ecdsa.service": { "name": "sshd-keygen@ecdsa.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd-keygen@ed25519.service": { "name": "sshd-keygen@ed25519.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd-keygen@rsa.service": { "name": "sshd-keygen@rsa.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd.service": { "name": "sshd.service", "source": "systemd", "state": "running", "status": "enabled" }, "sshd@.service": { "name": "sshd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "sssd-autofs.service": { "name": "sssd-autofs.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-kcm.service": { "name": "sssd-kcm.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "sssd-nss.service": { "name": "sssd-nss.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pac.service": { "name": "sssd-pac.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pam.service": { "name": "sssd-pam.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-ssh.service": { "name": "sssd-ssh.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-sudo.service": { "name": "sssd-sudo.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd.service": { "name": "sssd.service", "source": "systemd", "state": "running", "status": "enabled" }, "syslog.service": { "name": "syslog.service", "source": "systemd", "state": "active", "status": "enabled" }, "system-update-cleanup.service": { "name": "system-update-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-ask-password-console.service": { "name": "systemd-ask-password-console.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-wall.service": { "name": "systemd-ask-password-wall.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-backlight@.service": { "name": "systemd-backlight@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-binfmt.service": { "name": "systemd-binfmt.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-coredump@.service": { "name": "systemd-coredump@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-exit.service": { "name": "systemd-exit.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-firstboot.service": { "name": "systemd-firstboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck-root.service": { "name": "systemd-fsck-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck@.service": { "name": "systemd-fsck@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-fsck@dev-disk-by\\x2duuid-7B77\\x2d95E7.service": { "name": "systemd-fsck@dev-disk-by\\x2duuid-7B77\\x2d95E7.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-fsck@dev-vdb1.service": { "name": "systemd-fsck@dev-vdb1.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-fsck@dev-vdc1.service": { "name": "systemd-fsck@dev-vdc1.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-halt.service": { "name": "systemd-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hibernate-resume@.service": { "name": "systemd-hibernate-resume@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-hibernate.service": { "name": "systemd-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hostnamed.service": { "name": "systemd-hostnamed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hwdb-update.service": { "name": "systemd-hwdb-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-hybrid-sleep.service": { "name": "systemd-hybrid-sleep.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-initctl.service": { "name": "systemd-initctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-catalog-update.service": { "name": "systemd-journal-catalog-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-flush.service": { "name": "systemd-journal-flush.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journald.service": { "name": "systemd-journald.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-kexec.service": { "name": "systemd-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-localed.service": { "name": "systemd-localed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-logind.service": { "name": "systemd-logind.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-machine-id-commit.service": { "name": "systemd-machine-id-commit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-modules-load.service": { "name": "systemd-modules-load.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-networkd-wait-online.service": { "name": "systemd-networkd-wait-online.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "systemd-portabled.service": { "name": "systemd-portabled.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-poweroff.service": { "name": "systemd-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-quotacheck.service": { "name": "systemd-quotacheck.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-random-seed.service": { "name": "systemd-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-reboot.service": { "name": "systemd-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-remount-fs.service": { "name": "systemd-remount-fs.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-resolved.service": { "name": "systemd-resolved.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-rfkill.service": { "name": "systemd-rfkill.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend-then-hibernate.service": { "name": "systemd-suspend-then-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend.service": { "name": "systemd-suspend.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-sysctl.service": { "name": "systemd-sysctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-sysusers.service": { "name": "systemd-sysusers.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-timedated.service": { "name": "systemd-timedated.service", "source": "systemd", "state": "inactive", "status": "masked" }, "systemd-timesyncd.service": { "name": "systemd-timesyncd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "systemd-tmpfiles-clean.service": { "name": "systemd-tmpfiles-clean.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev.service": { "name": "systemd-tmpfiles-setup-dev.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup.service": { "name": "systemd-tmpfiles-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-settle.service": { "name": "systemd-udev-settle.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-trigger.service": { "name": "systemd-udev-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udevd.service": { "name": "systemd-udevd.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-update-done.service": { "name": "systemd-update-done.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp-runlevel.service": { "name": "systemd-update-utmp-runlevel.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp.service": { "name": "systemd-update-utmp.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-user-sessions.service": { "name": "systemd-user-sessions.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-vconsole-setup.service": { "name": "systemd-vconsole-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-volatile-root.service": { "name": "systemd-volatile-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "tcsd.service": { "name": "tcsd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "teamd@.service": { "name": "teamd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "timedatex.service": { "name": "timedatex.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "tuned.service": { "name": "tuned.service", "source": "systemd", "state": "running", "status": "enabled" }, "unbound-anchor.service": { "name": "unbound-anchor.service", "source": "systemd", "state": "stopped", "status": "static" }, "user-runtime-dir@.service": { "name": "user-runtime-dir@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user-runtime-dir@0.service": { "name": "user-runtime-dir@0.service", "source": "systemd", "state": "stopped", "status": "active" }, "user@.service": { "name": "user@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user@0.service": { "name": "user@0.service", "source": "systemd", "state": "running", "status": "active" }, "vdo-start-by-dev@.service": { "name": "vdo-start-by-dev@.service", "source": "systemd", "state": "unknown", "status": "static" }, "vdo.service": { "name": "vdo.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "ypbind.service": { "name": "ypbind.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "yppasswdd.service": { "name": "yppasswdd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ypserv.service": { "name": "ypserv.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ypxfrd.service": { "name": "ypxfrd.service", "source": "systemd", "state": "stopped", "status": "not-found" } } }, "changed": false } TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] ***** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:53 Thursday 21 July 2022 19:07:19 +0000 (0:00:01.800) 0:02:37.224 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:58 Thursday 21 July 2022 19:07:19 +0000 (0:00:00.103) 0:02:37.328 ********* TASK [fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64 Thursday 21 July 2022 19:07:19 +0000 (0:00:00.068) 0:02:37.397 ********* changed: [/cache/rhel-8-y.qcow2.snap] => { "actions": [ { "action": "resize device", "device": "/dev/mapper/foo-test2", "fs_type": null }, { "action": "resize format", "device": "/dev/mapper/foo-test2", "fs_type": "ext4" } ], "changed": true, "crypts": [], "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/vda2", "/dev/vda3", "/dev/mapper/foo-test2", "/dev/sdb", "/dev/sdc", "/dev/nvme0n1", "/dev/nvme1n1", "/dev/nvme2n1", "/dev/vdb1", "/dev/vdc1", "/dev/vdd", "/dev/vde", "/dev/vdf" ], "mounts": [ { "dump": 0, "fstype": "ext4", "opts": "defaults", "passno": 0, "path": "/opt/test2", "src": "/dev/mapper/foo-test2", "state": "mounted" } ], "packages": [ "dosfstools", "e2fsprogs", "lvm2", "xfsprogs" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test2", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test2", "_raw_device": "/dev/mapper/foo-test2", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "50%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:78 Thursday 21 July 2022 19:07:21 +0000 (0:00:02.126) 0:02:39.523 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90 Thursday 21 July 2022 19:07:22 +0000 (0:00:00.038) 0:02:39.562 ********* TASK [fedora.linux_system_roles.storage : show blivet_output] ****************** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:96 Thursday 21 July 2022 19:07:22 +0000 (0:00:00.022) 0:02:39.584 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "blivet_output": { "actions": [ { "action": "resize device", "device": "/dev/mapper/foo-test2", "fs_type": null }, { "action": "resize format", "device": "/dev/mapper/foo-test2", "fs_type": "ext4" } ], "changed": true, "crypts": [], "failed": false, "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/vda2", "/dev/vda3", "/dev/mapper/foo-test2", "/dev/sdb", "/dev/sdc", "/dev/nvme0n1", "/dev/nvme1n1", "/dev/nvme2n1", "/dev/vdb1", "/dev/vdc1", "/dev/vdd", "/dev/vde", "/dev/vdf" ], "mounts": [ { "dump": 0, "fstype": "ext4", "opts": "defaults", "passno": 0, "path": "/opt/test2", "src": "/dev/mapper/foo-test2", "state": "mounted" } ], "packages": [ "dosfstools", "e2fsprogs", "lvm2", "xfsprogs" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test2", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test2", "_raw_device": "/dev/mapper/foo-test2", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "50%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } } TASK [fedora.linux_system_roles.storage : set the list of pools for test verification] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:101 Thursday 21 July 2022 19:07:22 +0000 (0:00:00.043) 0:02:39.628 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test2", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test2", "_raw_device": "/dev/mapper/foo-test2", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "50%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] }, "changed": false } TASK [fedora.linux_system_roles.storage : set the list of volumes for test verification] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:105 Thursday 21 July 2022 19:07:22 +0000 (0:00:00.041) 0:02:39.670 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : remove obsolete mounts] ************** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:121 Thursday 21 July 2022 19:07:22 +0000 (0:00:00.039) 0:02:39.709 ********* TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132 Thursday 21 July 2022 19:07:22 +0000 (0:00:00.039) 0:02:39.748 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.storage : set up new/current mounts] *********** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:137 Thursday 21 July 2022 19:07:22 +0000 (0:00:00.645) 0:02:40.394 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount changed: [/cache/rhel-8-y.qcow2.snap] => (item={'src': '/dev/mapper/foo-test2', 'path': '/opt/test2', 'fstype': 'ext4', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted'}) => { "ansible_loop_var": "mount_info", "backup_file": "", "boot": "yes", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "ext4", "mount_info": { "dump": 0, "fstype": "ext4", "opts": "defaults", "passno": 0, "path": "/opt/test2", "src": "/dev/mapper/foo-test2", "state": "mounted" }, "name": "/opt/test2", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test2" } TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148 Thursday 21 July 2022 19:07:23 +0000 (0:00:00.436) 0:02:40.831 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.storage : retrieve facts for the /etc/crypttab file] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:156 Thursday 21 July 2022 19:07:23 +0000 (0:00:00.651) 0:02:41.482 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "stat": { "exists": false } } TASK [fedora.linux_system_roles.storage : manage /etc/crypttab to account for changes we just made] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:161 Thursday 21 July 2022 19:07:24 +0000 (0:00:00.377) 0:02:41.860 ********* TASK [fedora.linux_system_roles.storage : Update facts] ************************ task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:183 Thursday 21 July 2022 19:07:24 +0000 (0:00:00.022) 0:02:41.883 ********* ok: [/cache/rhel-8-y.qcow2.snap] META: role_complete for /cache/rhel-8-y.qcow2.snap TASK [include_tasks] *********************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/tests_lvm_percent_size.yml:160 Thursday 21 July 2022 19:07:25 +0000 (0:00:01.009) 0:02:42.892 ********* included: /tmp/tmpmnqeja9w/tests/storage/verify-role-results.yml for /cache/rhel-8-y.qcow2.snap TASK [Print out pool information] ********************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-role-results.yml:1 Thursday 21 July 2022 19:07:25 +0000 (0:00:00.050) 0:02:42.943 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test2", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test2", "_raw_device": "/dev/mapper/foo-test2", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "50%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] } TASK [Print out volume information] ******************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-role-results.yml:6 Thursday 21 July 2022 19:07:25 +0000 (0:00:00.052) 0:02:42.996 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => {} TASK [Collect info about the volumes.] ***************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-role-results.yml:14 Thursday 21 July 2022 19:07:25 +0000 (0:00:00.037) 0:02:43.034 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "info": { "/dev/mapper/foo-test2": { "fstype": "ext4", "label": "", "name": "/dev/mapper/foo-test2", "size": "5G", "type": "lvm", "uuid": "64454f89-f2cd-4a8b-958f-4c900f72cf79" }, "/dev/nvme0n1": { "fstype": "", "label": "", "name": "/dev/nvme0n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme1n1": { "fstype": "", "label": "", "name": "/dev/nvme1n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme2n1": { "fstype": "", "label": "", "name": "/dev/nvme2n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sda": { "fstype": "LVM2_member", "label": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "HA17fj-P3dB-xyZl-9Zru-ZzXr-ULh6-93Sgui" }, "/dev/sdb": { "fstype": "", "label": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc": { "fstype": "", "label": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sr0": { "fstype": "iso9660", "label": "cidata", "name": "/dev/sr0", "size": "364K", "type": "rom", "uuid": "2022-07-21-19-04-29-00" }, "/dev/vda": { "fstype": "", "label": "", "name": "/dev/vda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vda1": { "fstype": "", "label": "", "name": "/dev/vda1", "size": "1M", "type": "partition", "uuid": "" }, "/dev/vda2": { "fstype": "vfat", "label": "", "name": "/dev/vda2", "size": "100M", "type": "partition", "uuid": "7B77-95E7" }, "/dev/vda3": { "fstype": "xfs", "label": "root", "name": "/dev/vda3", "size": "9.9G", "type": "partition", "uuid": "58dca2db-4ccd-4171-a017-dd53ad8c1c54" }, "/dev/vdb": { "fstype": "", "label": "", "name": "/dev/vdb", "size": "2G", "type": "disk", "uuid": "" }, "/dev/vdb1": { "fstype": "ext4", "label": "yumcache", "name": "/dev/vdb1", "size": "2G", "type": "partition", "uuid": "4d1ed9a1-5608-4ff8-ad89-21d95caa66fa" }, "/dev/vdc": { "fstype": "", "label": "", "name": "/dev/vdc", "size": "2G", "type": "disk", "uuid": "" }, "/dev/vdc1": { "fstype": "ext4", "label": "yumvarlib", "name": "/dev/vdc1", "size": "2G", "type": "partition", "uuid": "a7abcc97-edc5-476c-8ccd-694b4621fd28" }, "/dev/vdd": { "fstype": "", "label": "", "name": "/dev/vdd", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vde": { "fstype": "", "label": "", "name": "/dev/vde", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vdf": { "fstype": "", "label": "", "name": "/dev/vdf", "size": "10G", "type": "disk", "uuid": "" } } } TASK [Read the /etc/fstab file for volume existence] *************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-role-results.yml:19 Thursday 21 July 2022 19:07:25 +0000 (0:00:00.420) 0:02:43.454 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "cmd": [ "cat", "/etc/fstab" ], "delta": "0:00:00.003275", "end": "2022-07-21 15:07:26.399851", "rc": 0, "start": "2022-07-21 15:07:26.396576" } STDOUT: UUID=58dca2db-4ccd-4171-a017-dd53ad8c1c54 / xfs defaults 0 0 UUID=7B77-95E7 /boot/efi vfat defaults,uid=0,gid=0,umask=077,shortname=winnt 0 2 /dev/vdb1 /var/cache/dnf auto defaults,nofail,x-systemd.requires=cloud-init.service,_netdev,comment=cloudconfig 0 2 /dev/vdc1 /var/lib/dnf auto defaults,nofail,x-systemd.requires=cloud-init.service,_netdev,comment=cloudconfig 0 2 /dev/mapper/foo-test2 /opt/test2 ext4 defaults 0 0 TASK [Read the /etc/crypttab file] ********************************************* task path: /tmp/tmpmnqeja9w/tests/storage/verify-role-results.yml:24 Thursday 21 July 2022 19:07:26 +0000 (0:00:00.430) 0:02:43.885 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "cmd": [ "cat", "/etc/crypttab" ], "delta": "0:00:00.003107", "end": "2022-07-21 15:07:26.789301", "failed_when_result": false, "rc": 1, "start": "2022-07-21 15:07:26.786194" } STDERR: cat: /etc/crypttab: No such file or directory MSG: non-zero return code TASK [Verify the volumes listed in storage_pools were correctly managed] ******* task path: /tmp/tmpmnqeja9w/tests/storage/verify-role-results.yml:33 Thursday 21 July 2022 19:07:26 +0000 (0:00:00.388) 0:02:44.273 ********* included: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool.yml for /cache/rhel-8-y.qcow2.snap => (item={'disks': ['sda'], 'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'name': 'foo', 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_metadata_version': None, 'state': 'present', 'type': 'lvm', 'volumes': [{'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'ext4', 'mount_options': 'defaults', 'mount_point': '/opt/test2', 'name': 'test2', 'raid_level': None, 'size': '50%', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test2', '_raw_device': '/dev/mapper/foo-test2', '_mount_id': '/dev/mapper/foo-test2', '_kernel_device': '/dev/dm-0', '_raw_kernel_device': '/dev/dm-0'}], 'raid_chunk_size': None}) TASK [Set _storage_pool_tests] ************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool.yml:5 Thursday 21 July 2022 19:07:26 +0000 (0:00:00.069) 0:02:44.342 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_pool_tests": [ "members", "volumes" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool.yml:18 Thursday 21 July 2022 19:07:26 +0000 (0:00:00.036) 0:02:44.379 ********* included: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml for /cache/rhel-8-y.qcow2.snap => (item=members) included: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-volumes.yml for /cache/rhel-8-y.qcow2.snap => (item=volumes) TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:1 Thursday 21 July 2022 19:07:26 +0000 (0:00:00.052) 0:02:44.432 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_pv_count": "1", "_storage_test_pool_pvs_lvm": [ "/dev/sda" ] }, "changed": false } TASK [Get the canonical device path for each member device] ******************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:6 Thursday 21 July 2022 19:07:26 +0000 (0:00:00.056) 0:02:44.488 ********* ok: [/cache/rhel-8-y.qcow2.snap] => (item=/dev/sda) => { "ansible_loop_var": "pv", "changed": false, "device": "/dev/sda", "pv": "/dev/sda" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:15 Thursday 21 July 2022 19:07:27 +0000 (0:00:00.399) 0:02:44.888 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "__pvs_lvm_len": "1" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:19 Thursday 21 July 2022 19:07:27 +0000 (0:00:00.049) 0:02:44.938 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_pool_pvs": [ "/dev/sda" ] }, "changed": false } TASK [Verify PV count] ********************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:23 Thursday 21 July 2022 19:07:27 +0000 (0:00:00.050) 0:02:44.988 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:29 Thursday 21 July 2022 19:07:27 +0000 (0:00:00.055) 0:02:45.044 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:33 Thursday 21 July 2022 19:07:27 +0000 (0:00:00.036) 0:02:45.080 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:37 Thursday 21 July 2022 19:07:27 +0000 (0:00:00.050) 0:02:45.131 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check the type of each PV] *********************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:41 Thursday 21 July 2022 19:07:27 +0000 (0:00:00.025) 0:02:45.156 ********* ok: [/cache/rhel-8-y.qcow2.snap] => (item=/dev/sda) => { "ansible_loop_var": "pv", "changed": false, "pv": "/dev/sda" } MSG: All assertions passed TASK [Check MD RAID] *********************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:50 Thursday 21 July 2022 19:07:27 +0000 (0:00:00.046) 0:02:45.203 ********* included: /tmp/tmpmnqeja9w/tests/storage/verify-pool-md.yml for /cache/rhel-8-y.qcow2.snap TASK [get information about RAID] ********************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-md.yml:6 Thursday 21 July 2022 19:07:27 +0000 (0:00:00.044) 0:02:45.247 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-md.yml:12 Thursday 21 July 2022 19:07:27 +0000 (0:00:00.028) 0:02:45.275 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-md.yml:16 Thursday 21 July 2022 19:07:27 +0000 (0:00:00.025) 0:02:45.301 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-md.yml:20 Thursday 21 July 2022 19:07:27 +0000 (0:00:00.027) 0:02:45.328 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-md.yml:24 Thursday 21 July 2022 19:07:27 +0000 (0:00:00.026) 0:02:45.355 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-md.yml:30 Thursday 21 July 2022 19:07:27 +0000 (0:00:00.027) 0:02:45.383 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-md.yml:36 Thursday 21 July 2022 19:07:27 +0000 (0:00:00.024) 0:02:45.407 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-md.yml:44 Thursday 21 July 2022 19:07:27 +0000 (0:00:00.026) 0:02:45.434 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_md_active_devices_re": null, "storage_test_md_metadata_version_re": null, "storage_test_md_spare_devices_re": null }, "changed": false } TASK [Check LVM RAID] ********************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:53 Thursday 21 July 2022 19:07:27 +0000 (0:00:00.088) 0:02:45.522 ********* included: /tmp/tmpmnqeja9w/tests/storage/verify-pool-members-lvmraid.yml for /cache/rhel-8-y.qcow2.snap TASK [Validate pool member LVM RAID settings] ********************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-members-lvmraid.yml:1 Thursday 21 July 2022 19:07:28 +0000 (0:00:00.046) 0:02:45.569 ********* included: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-lvmraid.yml for /cache/rhel-8-y.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'ext4', 'mount_options': 'defaults', 'mount_point': '/opt/test2', 'name': 'test2', 'raid_level': None, 'size': '50%', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test2', '_raw_device': '/dev/mapper/foo-test2', '_mount_id': '/dev/mapper/foo-test2', '_kernel_device': '/dev/dm-0', '_raw_kernel_device': '/dev/dm-0'}) TASK [Get information about LVM RAID] ****************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-lvmraid.yml:3 Thursday 21 July 2022 19:07:28 +0000 (0:00:00.043) 0:02:45.612 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is LVM RAID] ******************************************* task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-lvmraid.yml:8 Thursday 21 July 2022 19:07:28 +0000 (0:00:00.070) 0:02:45.683 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-lvmraid.yml:12 Thursday 21 July 2022 19:07:28 +0000 (0:00:00.030) 0:02:45.713 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check Thin Pools] ******************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:56 Thursday 21 July 2022 19:07:28 +0000 (0:00:00.031) 0:02:45.744 ********* included: /tmp/tmpmnqeja9w/tests/storage/verify-pool-members-thin.yml for /cache/rhel-8-y.qcow2.snap TASK [Validate pool member thinpool settings] ********************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-members-thin.yml:1 Thursday 21 July 2022 19:07:28 +0000 (0:00:00.048) 0:02:45.793 ********* included: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-thin.yml for /cache/rhel-8-y.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'ext4', 'mount_options': 'defaults', 'mount_point': '/opt/test2', 'name': 'test2', 'raid_level': None, 'size': '50%', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test2', '_raw_device': '/dev/mapper/foo-test2', '_mount_id': '/dev/mapper/foo-test2', '_kernel_device': '/dev/dm-0', '_raw_kernel_device': '/dev/dm-0'}) TASK [Get information about thinpool] ****************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-thin.yml:3 Thursday 21 July 2022 19:07:28 +0000 (0:00:00.055) 0:02:45.849 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-thin.yml:8 Thursday 21 July 2022 19:07:28 +0000 (0:00:00.026) 0:02:45.875 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-thin.yml:13 Thursday 21 July 2022 19:07:28 +0000 (0:00:00.027) 0:02:45.902 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-thin.yml:17 Thursday 21 July 2022 19:07:28 +0000 (0:00:00.025) 0:02:45.928 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check member encryption] ************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:59 Thursday 21 July 2022 19:07:28 +0000 (0:00:00.025) 0:02:45.954 ********* included: /tmp/tmpmnqeja9w/tests/storage/verify-pool-members-encryption.yml for /cache/rhel-8-y.qcow2.snap TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-members-encryption.yml:4 Thursday 21 July 2022 19:07:28 +0000 (0:00:00.051) 0:02:46.006 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Validate pool member LUKS settings] ************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-members-encryption.yml:8 Thursday 21 July 2022 19:07:28 +0000 (0:00:00.055) 0:02:46.061 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => (item=/dev/sda) => { "_storage_test_pool_member_path": "/dev/sda", "ansible_loop_var": "_storage_test_pool_member_path", "changed": false, "skip_reason": "Conditional result was False" } TASK [Validate pool member crypttab entries] *********************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-members-encryption.yml:15 Thursday 21 July 2022 19:07:28 +0000 (0:00:00.028) 0:02:46.089 ********* included: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-crypttab.yml for /cache/rhel-8-y.qcow2.snap => (item=/dev/sda) TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-crypttab.yml:1 Thursday 21 July 2022 19:07:28 +0000 (0:00:00.045) 0:02:46.135 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": [] }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-crypttab.yml:4 Thursday 21 July 2022 19:07:28 +0000 (0:00:00.052) 0:02:46.187 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-crypttab.yml:9 Thursday 21 July 2022 19:07:28 +0000 (0:00:00.054) 0:02:46.242 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-crypttab.yml:15 Thursday 21 July 2022 19:07:28 +0000 (0:00:00.039) 0:02:46.281 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-crypttab.yml:21 Thursday 21 July 2022 19:07:28 +0000 (0:00:00.037) 0:02:46.319 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-crypttab.yml:27 Thursday 21 July 2022 19:07:28 +0000 (0:00:00.044) 0:02:46.363 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-members-encryption.yml:22 Thursday 21 July 2022 19:07:28 +0000 (0:00:00.047) 0:02:46.411 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_crypttab_key_file": null }, "changed": false } TASK [Check VDO] *************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:62 Thursday 21 July 2022 19:07:28 +0000 (0:00:00.037) 0:02:46.449 ********* included: /tmp/tmpmnqeja9w/tests/storage/verify-pool-members-vdo.yml for /cache/rhel-8-y.qcow2.snap TASK [Validate pool member VDO settings] *************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-members-vdo.yml:1 Thursday 21 July 2022 19:07:28 +0000 (0:00:00.051) 0:02:46.500 ********* included: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-vdo.yml for /cache/rhel-8-y.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'ext4', 'mount_options': 'defaults', 'mount_point': '/opt/test2', 'name': 'test2', 'raid_level': None, 'size': '50%', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test2', '_raw_device': '/dev/mapper/foo-test2', '_mount_id': '/dev/mapper/foo-test2', '_kernel_device': '/dev/dm-0', '_raw_kernel_device': '/dev/dm-0'}) TASK [get information about VDO deduplication] ********************************* task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-vdo.yml:3 Thursday 21 July 2022 19:07:28 +0000 (0:00:00.047) 0:02:46.548 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-vdo.yml:8 Thursday 21 July 2022 19:07:29 +0000 (0:00:00.026) 0:02:46.574 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-vdo.yml:11 Thursday 21 July 2022 19:07:29 +0000 (0:00:00.026) 0:02:46.601 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-vdo.yml:16 Thursday 21 July 2022 19:07:29 +0000 (0:00:00.027) 0:02:46.628 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-vdo.yml:21 Thursday 21 July 2022 19:07:29 +0000 (0:00:00.024) 0:02:46.652 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-vdo.yml:24 Thursday 21 July 2022 19:07:29 +0000 (0:00:00.023) 0:02:46.676 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-vdo.yml:29 Thursday 21 July 2022 19:07:29 +0000 (0:00:00.026) 0:02:46.702 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-member-vdo.yml:39 Thursday 21 July 2022 19:07:29 +0000 (0:00:00.027) 0:02:46.730 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:65 Thursday 21 July 2022 19:07:29 +0000 (0:00:00.036) 0:02:46.766 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "__pvs_lvm_len": null, "_storage_test_expected_pv_count": null, "_storage_test_expected_pv_type": null, "_storage_test_pool_pvs": [], "_storage_test_pool_pvs_lvm": [] }, "changed": false } TASK [verify the volumes] ****************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-volumes.yml:3 Thursday 21 July 2022 19:07:29 +0000 (0:00:00.038) 0:02:46.805 ********* included: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume.yml for /cache/rhel-8-y.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'ext4', 'mount_options': 'defaults', 'mount_point': '/opt/test2', 'name': 'test2', 'raid_level': None, 'size': '50%', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/foo-test2', '_raw_device': '/dev/mapper/foo-test2', '_mount_id': '/dev/mapper/foo-test2', '_kernel_device': '/dev/dm-0', '_raw_kernel_device': '/dev/dm-0'}) TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume.yml:2 Thursday 21 July 2022 19:07:29 +0000 (0:00:00.046) 0:02:46.851 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_volume_present": true, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume.yml:10 Thursday 21 July 2022 19:07:29 +0000 (0:00:00.096) 0:02:46.948 ********* included: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml for /cache/rhel-8-y.qcow2.snap => (item=mount) included: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-fstab.yml for /cache/rhel-8-y.qcow2.snap => (item=fstab) included: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-fs.yml for /cache/rhel-8-y.qcow2.snap => (item=fs) included: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-device.yml for /cache/rhel-8-y.qcow2.snap => (item=device) included: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml for /cache/rhel-8-y.qcow2.snap => (item=encryption) included: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-md.yml for /cache/rhel-8-y.qcow2.snap => (item=md) included: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml for /cache/rhel-8-y.qcow2.snap => (item=size) included: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-cache.yml for /cache/rhel-8-y.qcow2.snap => (item=cache) TASK [Get expected mount device based on device type] ************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml:6 Thursday 21 July 2022 19:07:29 +0000 (0:00:00.121) 0:02:47.070 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/foo-test2" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml:10 Thursday 21 July 2022 19:07:29 +0000 (0:00:00.041) 0:02:47.111 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_mount_device_matches": [ { "block_available": 1200036, "block_size": 4096, "block_total": 1269672, "block_used": 69636, "device": "/dev/mapper/foo-test2", "fstype": "ext4", "inode_available": 327669, "inode_total": 327680, "inode_used": 11, "mount": "/opt/test2", "options": "rw,seclabel,relatime", "size_available": 4915347456, "size_total": 5200576512, "uuid": "64454f89-f2cd-4a8b-958f-4c900f72cf79" } ], "storage_test_mount_expected_match_count": "1", "storage_test_mount_point_matches": [ { "block_available": 1200036, "block_size": 4096, "block_total": 1269672, "block_used": 69636, "device": "/dev/mapper/foo-test2", "fstype": "ext4", "inode_available": 327669, "inode_total": 327680, "inode_used": 11, "mount": "/opt/test2", "options": "rw,seclabel,relatime", "size_available": 4915347456, "size_total": 5200576512, "uuid": "64454f89-f2cd-4a8b-958f-4c900f72cf79" } ], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Verify the current mount state by device] ******************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml:20 Thursday 21 July 2022 19:07:29 +0000 (0:00:00.054) 0:02:47.165 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify the current mount state by mount point] *************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml:29 Thursday 21 July 2022 19:07:29 +0000 (0:00:00.050) 0:02:47.215 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify the mount fs type] ************************************************ task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml:37 Thursday 21 July 2022 19:07:29 +0000 (0:00:00.051) 0:02:47.266 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [command] ***************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml:46 Thursday 21 July 2022 19:07:29 +0000 (0:00:00.052) 0:02:47.319 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml:50 Thursday 21 July 2022 19:07:29 +0000 (0:00:00.026) 0:02:47.345 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml:55 Thursday 21 July 2022 19:07:29 +0000 (0:00:00.024) 0:02:47.370 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-mount.yml:65 Thursday 21 July 2022 19:07:29 +0000 (0:00:00.025) 0:02:47.395 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-fstab.yml:2 Thursday 21 July 2022 19:07:29 +0000 (0:00:00.035) 0:02:47.431 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "1", "storage_test_fstab_expected_mount_options_matches": "1", "storage_test_fstab_expected_mount_point_matches": "1", "storage_test_fstab_id_matches": [ "/dev/mapper/foo-test2 " ], "storage_test_fstab_mount_options_matches": [ " /opt/test2 ext4 defaults " ], "storage_test_fstab_mount_point_matches": [ " /opt/test2 " ] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-fstab.yml:12 Thursday 21 July 2022 19:07:29 +0000 (0:00:00.064) 0:02:47.495 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify the fstab mount point] ******************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-fstab.yml:19 Thursday 21 July 2022 19:07:29 +0000 (0:00:00.050) 0:02:47.546 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-fstab.yml:25 Thursday 21 July 2022 19:07:30 +0000 (0:00:00.048) 0:02:47.594 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-fstab.yml:34 Thursday 21 July 2022 19:07:30 +0000 (0:00:00.034) 0:02:47.629 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-fs.yml:4 Thursday 21 July 2022 19:07:30 +0000 (0:00:00.034) 0:02:47.663 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify fs label] ********************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-fs.yml:10 Thursday 21 July 2022 19:07:30 +0000 (0:00:00.042) 0:02:47.705 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [See whether the device node is present] ********************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-device.yml:4 Thursday 21 July 2022 19:07:30 +0000 (0:00:00.043) 0:02:47.749 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "stat": { "atime": 1658430441.9911165, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1658430441.9911165, "dev": 6, "device_type": 64768, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 59192, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/symlink", "mode": "0660", "mtime": 1658430441.9911165, "nlink": 1, "path": "/dev/mapper/foo-test2", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-device.yml:10 Thursday 21 July 2022 19:07:30 +0000 (0:00:00.382) 0:02:48.132 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about this volume] ********************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-device.yml:15 Thursday 21 July 2022 19:07:30 +0000 (0:00:00.040) 0:02:48.172 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [(1/2) Process volume type (set initial value)] *************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-device.yml:21 Thursday 21 July 2022 19:07:30 +0000 (0:00:00.040) 0:02:48.213 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [(2/2) Process volume type (get RAID value)] ****************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-device.yml:25 Thursday 21 July 2022 19:07:30 +0000 (0:00:00.039) 0:02:48.252 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-device.yml:30 Thursday 21 July 2022 19:07:30 +0000 (0:00:00.024) 0:02:48.276 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Stat the LUKS device, if encrypted] ************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:3 Thursday 21 July 2022 19:07:30 +0000 (0:00:00.039) 0:02:48.315 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:10 Thursday 21 July 2022 19:07:30 +0000 (0:00:00.025) 0:02:48.341 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [Collect LUKS info for this volume] *************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:15 Thursday 21 July 2022 19:07:32 +0000 (0:00:01.339) 0:02:49.681 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:21 Thursday 21 July 2022 19:07:32 +0000 (0:00:00.024) 0:02:49.705 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:27 Thursday 21 July 2022 19:07:32 +0000 (0:00:00.022) 0:02:49.728 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:33 Thursday 21 July 2022 19:07:32 +0000 (0:00:00.083) 0:02:49.811 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:39 Thursday 21 July 2022 19:07:32 +0000 (0:00:00.025) 0:02:49.836 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:44 Thursday 21 July 2022 19:07:32 +0000 (0:00:00.025) 0:02:49.862 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:50 Thursday 21 July 2022 19:07:32 +0000 (0:00:00.057) 0:02:49.919 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:56 Thursday 21 July 2022 19:07:32 +0000 (0:00:00.025) 0:02:49.945 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:62 Thursday 21 July 2022 19:07:32 +0000 (0:00:00.023) 0:02:49.969 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:67 Thursday 21 July 2022 19:07:32 +0000 (0:00:00.050) 0:02:50.019 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:72 Thursday 21 July 2022 19:07:32 +0000 (0:00:00.047) 0:02:50.067 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:78 Thursday 21 July 2022 19:07:32 +0000 (0:00:00.037) 0:02:50.105 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:84 Thursday 21 July 2022 19:07:32 +0000 (0:00:00.036) 0:02:50.142 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-encryption.yml:90 Thursday 21 July 2022 19:07:32 +0000 (0:00:00.041) 0:02:50.183 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [get information about RAID] ********************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-md.yml:7 Thursday 21 July 2022 19:07:32 +0000 (0:00:00.037) 0:02:50.220 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-md.yml:13 Thursday 21 July 2022 19:07:32 +0000 (0:00:00.038) 0:02:50.258 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-md.yml:17 Thursday 21 July 2022 19:07:32 +0000 (0:00:00.037) 0:02:50.295 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-md.yml:21 Thursday 21 July 2022 19:07:32 +0000 (0:00:00.041) 0:02:50.337 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-md.yml:25 Thursday 21 July 2022 19:07:32 +0000 (0:00:00.043) 0:02:50.381 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-md.yml:31 Thursday 21 July 2022 19:07:32 +0000 (0:00:00.043) 0:02:50.424 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-md.yml:37 Thursday 21 July 2022 19:07:32 +0000 (0:00:00.041) 0:02:50.466 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the actual size of the volume] ************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:3 Thursday 21 July 2022 19:07:32 +0000 (0:00:00.046) 0:02:50.513 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "bytes": 5368709120, "changed": false, "lvm": "5g", "parted": "5GiB", "size": "5 GiB" } TASK [parse the requested size of the volume] ********************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:9 Thursday 21 July 2022 19:07:33 +0000 (0:00:00.388) 0:02:50.901 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Establish base value for expected size] ********************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:15 Thursday 21 July 2022 19:07:33 +0000 (0:00:00.040) 0:02:50.941 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:20 Thursday 21 July 2022 19:07:33 +0000 (0:00:00.039) 0:02:50.980 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_expected_size": "2684354560.0" } TASK [debug] ******************************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:25 Thursday 21 July 2022 19:07:33 +0000 (0:00:00.035) 0:02:51.015 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_pool": { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test2", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test2", "_raw_device": "/dev/mapper/foo-test2", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "50%", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } } TASK [debug] ******************************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:28 Thursday 21 July 2022 19:07:33 +0000 (0:00:00.055) 0:02:51.071 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_blkinfo": { "changed": false, "failed": false, "info": { "/dev/mapper/foo-test2": { "fstype": "ext4", "label": "", "name": "/dev/mapper/foo-test2", "size": "5G", "type": "lvm", "uuid": "64454f89-f2cd-4a8b-958f-4c900f72cf79" }, "/dev/nvme0n1": { "fstype": "", "label": "", "name": "/dev/nvme0n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme1n1": { "fstype": "", "label": "", "name": "/dev/nvme1n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme2n1": { "fstype": "", "label": "", "name": "/dev/nvme2n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sda": { "fstype": "LVM2_member", "label": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "HA17fj-P3dB-xyZl-9Zru-ZzXr-ULh6-93Sgui" }, "/dev/sdb": { "fstype": "", "label": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc": { "fstype": "", "label": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sr0": { "fstype": "iso9660", "label": "cidata", "name": "/dev/sr0", "size": "364K", "type": "rom", "uuid": "2022-07-21-19-04-29-00" }, "/dev/vda": { "fstype": "", "label": "", "name": "/dev/vda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vda1": { "fstype": "", "label": "", "name": "/dev/vda1", "size": "1M", "type": "partition", "uuid": "" }, "/dev/vda2": { "fstype": "vfat", "label": "", "name": "/dev/vda2", "size": "100M", "type": "partition", "uuid": "7B77-95E7" }, "/dev/vda3": { "fstype": "xfs", "label": "root", "name": "/dev/vda3", "size": "9.9G", "type": "partition", "uuid": "58dca2db-4ccd-4171-a017-dd53ad8c1c54" }, "/dev/vdb": { "fstype": "", "label": "", "name": "/dev/vdb", "size": "2G", "type": "disk", "uuid": "" }, "/dev/vdb1": { "fstype": "ext4", "label": "yumcache", "name": "/dev/vdb1", "size": "2G", "type": "partition", "uuid": "4d1ed9a1-5608-4ff8-ad89-21d95caa66fa" }, "/dev/vdc": { "fstype": "", "label": "", "name": "/dev/vdc", "size": "2G", "type": "disk", "uuid": "" }, "/dev/vdc1": { "fstype": "ext4", "label": "yumvarlib", "name": "/dev/vdc1", "size": "2G", "type": "partition", "uuid": "a7abcc97-edc5-476c-8ccd-694b4621fd28" }, "/dev/vdd": { "fstype": "", "label": "", "name": "/dev/vdd", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vde": { "fstype": "", "label": "", "name": "/dev/vde", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vdf": { "fstype": "", "label": "", "name": "/dev/vdf", "size": "10G", "type": "disk", "uuid": "" } } } } TASK [Get the size of parent/pool device] ************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:31 Thursday 21 July 2022 19:07:33 +0000 (0:00:00.059) 0:02:51.130 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "bytes": 10737418240, "changed": false, "lvm": "10g", "parted": "10GiB", "size": "10 GiB" } TASK [debug] ******************************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:36 Thursday 21 July 2022 19:07:33 +0000 (0:00:00.403) 0:02:51.534 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_pool_size": { "bytes": 10737418240, "changed": false, "failed": false, "lvm": "10g", "parted": "10GiB", "size": "10 GiB" } } TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:39 Thursday 21 July 2022 19:07:34 +0000 (0:00:00.054) 0:02:51.588 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_expected_size": "5368709120.0" }, "changed": false } TASK [debug] ******************************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:44 Thursday 21 July 2022 19:07:34 +0000 (0:00:00.057) 0:02:51.646 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_actual_size": { "bytes": 5368709120, "changed": false, "failed": false, "lvm": "5g", "parted": "5GiB", "size": "5 GiB" } } TASK [debug] ******************************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:47 Thursday 21 July 2022 19:07:34 +0000 (0:00:00.039) 0:02:51.685 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_test_expected_size": "5368709120.0" } TASK [assert] ****************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-size.yml:50 Thursday 21 July 2022 19:07:34 +0000 (0:00:00.039) 0:02:51.724 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Get information about the LV] ******************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-cache.yml:6 Thursday 21 July 2022 19:07:34 +0000 (0:00:00.055) 0:02:51.780 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "cmd": [ "lvs", "--noheadings", "--nameprefixes", "--units=b", "--nosuffix", "--unquoted", "-o", "name,attr,cache_total_blocks,chunk_size,segtype", "foo/test2" ], "delta": "0:00:00.037837", "end": "2022-07-21 15:07:34.714076", "rc": 0, "start": "2022-07-21 15:07:34.676239" } STDOUT: LVM2_LV_NAME=test2 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-cache.yml:14 Thursday 21 July 2022 19:07:34 +0000 (0:00:00.419) 0:02:52.200 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_lv_segtype": [ "linear" ] }, "changed": false } TASK [check segment type] ****************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-cache.yml:17 Thursday 21 July 2022 19:07:34 +0000 (0:00:00.116) 0:02:52.316 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-cache.yml:22 Thursday 21 July 2022 19:07:34 +0000 (0:00:00.087) 0:02:52.403 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the requested cache size] ****************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-cache.yml:26 Thursday 21 July 2022 19:07:34 +0000 (0:00:00.040) 0:02:52.444 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-cache.yml:32 Thursday 21 July 2022 19:07:34 +0000 (0:00:00.040) 0:02:52.484 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume-cache.yml:36 Thursday 21 July 2022 19:07:34 +0000 (0:00:00.038) 0:02:52.523 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-volume.yml:16 Thursday 21 July 2022 19:07:35 +0000 (0:00:00.076) 0:02:52.600 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [Verify the volumes with no pool were correctly managed] ****************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-role-results.yml:43 Thursday 21 July 2022 19:07:35 +0000 (0:00:00.034) 0:02:52.635 ********* TASK [Clean up variable namespace] ********************************************* task path: /tmp/tmpmnqeja9w/tests/storage/verify-role-results.yml:53 Thursday 21 July 2022 19:07:35 +0000 (0:00:00.021) 0:02:52.656 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_blkinfo": null, "storage_test_crypttab": null, "storage_test_fstab": null }, "changed": false } TASK [Remove both of the LVM logical volumes in 'foo' created above] *********** task path: /tmp/tmpmnqeja9w/tests/storage/tests_lvm_percent_size.yml:162 Thursday 21 July 2022 19:07:35 +0000 (0:00:00.035) 0:02:52.691 ********* TASK [fedora.linux_system_roles.storage : set platform/version specific variables] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2 Thursday 21 July 2022 19:07:35 +0000 (0:00:00.068) 0:02:52.760 ********* included: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for /cache/rhel-8-y.qcow2.snap TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2 Thursday 21 July 2022 19:07:35 +0000 (0:00:00.037) 0:02:52.798 ********* ok: [/cache/rhel-8-y.qcow2.snap] TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7 Thursday 21 July 2022 19:07:35 +0000 (0:00:00.532) 0:02:53.331 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [/cache/rhel-8-y.qcow2.snap] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [/cache/rhel-8-y.qcow2.snap] => (item=RedHat_8.yml) => { "ansible_facts": { "blivet_package_list": [ "python3-blivet", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap", "vdo", "kmod-kvdo", "xfsprogs" ] }, "ansible_included_var_files": [ "/tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/vars/RedHat_8.yml" ], "ansible_loop_var": "item", "changed": false, "item": "RedHat_8.yml" } skipping: [/cache/rhel-8-y.qcow2.snap] => (item=RedHat_8.7.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat_8.7.yml", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : define an empty list of pools to be used in testing] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5 Thursday 21 July 2022 19:07:35 +0000 (0:00:00.066) 0:02:53.397 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : define an empty list of volumes to be used in testing] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9 Thursday 21 July 2022 19:07:35 +0000 (0:00:00.040) 0:02:53.438 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : include the appropriate provider tasks] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13 Thursday 21 July 2022 19:07:35 +0000 (0:00:00.035) 0:02:53.474 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount included: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for /cache/rhel-8-y.qcow2.snap TASK [fedora.linux_system_roles.storage : get a list of rpm packages installed on host machine] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2 Thursday 21 July 2022 19:07:35 +0000 (0:00:00.058) 0:02:53.533 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : make sure blivet is available] ******* task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:7 Thursday 21 July 2022 19:07:36 +0000 (0:00:00.022) 0:02:53.555 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [fedora.linux_system_roles.storage : show storage_pools] ****************** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:13 Thursday 21 July 2022 19:07:37 +0000 (0:00:01.349) 0:02:54.905 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_pools": [ { "disks": [ "sda" ], "name": "foo", "state": "absent" } ] } TASK [fedora.linux_system_roles.storage : show storage_volumes] **************** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:18 Thursday 21 July 2022 19:07:37 +0000 (0:00:00.041) 0:02:54.946 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [fedora.linux_system_roles.storage : get required packages] *************** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:23 Thursday 21 July 2022 19:07:37 +0000 (0:00:00.036) 0:02:54.982 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [], "pools": [], "volumes": [] } TASK [fedora.linux_system_roles.storage : enable copr repositories if needed] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:35 Thursday 21 July 2022 19:07:38 +0000 (0:00:01.557) 0:02:56.539 ********* included: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml for /cache/rhel-8-y.qcow2.snap TASK [fedora.linux_system_roles.storage : check if the COPR support packages should be installed] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:2 Thursday 21 July 2022 19:07:39 +0000 (0:00:00.054) 0:02:56.593 ********* TASK [fedora.linux_system_roles.storage : make sure COPR support packages are present] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:13 Thursday 21 July 2022 19:07:39 +0000 (0:00:00.040) 0:02:56.634 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : enable COPRs] ************************ task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:18 Thursday 21 July 2022 19:07:39 +0000 (0:00:00.041) 0:02:56.676 ********* TASK [fedora.linux_system_roles.storage : make sure required packages are installed] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:41 Thursday 21 July 2022 19:07:39 +0000 (0:00:00.038) 0:02:56.714 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [fedora.linux_system_roles.storage : get service facts] ******************* task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:47 Thursday 21 July 2022 19:07:40 +0000 (0:00:01.292) 0:02:58.007 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "services": { "NetworkManager-dispatcher.service": { "name": "NetworkManager-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "NetworkManager-wait-online.service": { "name": "NetworkManager-wait-online.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "NetworkManager.service": { "name": "NetworkManager.service", "source": "systemd", "state": "running", "status": "enabled" }, "arp-ethers.service": { "name": "arp-ethers.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "auditd.service": { "name": "auditd.service", "source": "systemd", "state": "running", "status": "enabled" }, "auth-rpcgss-module.service": { "name": "auth-rpcgss-module.service", "source": "systemd", "state": "stopped", "status": "static" }, "autovt@.service": { "name": "autovt@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "blivet.service": { "name": "blivet.service", "source": "systemd", "state": "inactive", "status": "static" }, "blk-availability.service": { "name": "blk-availability.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "chrony-dnssrv@.service": { "name": "chrony-dnssrv@.service", "source": "systemd", "state": "unknown", "status": "static" }, "chrony-wait.service": { "name": "chrony-wait.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd.service": { "name": "chronyd.service", "source": "systemd", "state": "running", "status": "enabled" }, "cloud-config.service": { "name": "cloud-config.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-final.service": { "name": "cloud-final.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init-hotplugd.service": { "name": "cloud-init-hotplugd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cloud-init-local.service": { "name": "cloud-init-local.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init.service": { "name": "cloud-init.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cockpit-motd.service": { "name": "cockpit-motd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cockpit-wsinstance-http.service": { "name": "cockpit-wsinstance-http.service", "source": "systemd", "state": "inactive", "status": "static" }, "cockpit-wsinstance-https-factory@.service": { "name": "cockpit-wsinstance-https-factory@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cockpit-wsinstance-https@.service": { "name": "cockpit-wsinstance-https@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cockpit.service": { "name": "cockpit.service", "source": "systemd", "state": "inactive", "status": "static" }, "console-getty.service": { "name": "console-getty.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "container-getty@.service": { "name": "container-getty@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cpupower.service": { "name": "cpupower.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "crond.service": { "name": "crond.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-org.freedesktop.hostname1.service": { "name": "dbus-org.freedesktop.hostname1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.locale1.service": { "name": "dbus-org.freedesktop.locale1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.login1.service": { "name": "dbus-org.freedesktop.login1.service", "source": "systemd", "state": "active", "status": "static" }, "dbus-org.freedesktop.nm-dispatcher.service": { "name": "dbus-org.freedesktop.nm-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "dbus-org.freedesktop.portable1.service": { "name": "dbus-org.freedesktop.portable1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.timedate1.service": { "name": "dbus-org.freedesktop.timedate1.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "dbus.service": { "name": "dbus.service", "source": "systemd", "state": "running", "status": "static" }, "dbxtool.service": { "name": "dbxtool.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "debug-shell.service": { "name": "debug-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "display-manager.service": { "name": "display-manager.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "dm-event.service": { "name": "dm-event.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-makecache.service": { "name": "dnf-makecache.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-cmdline.service": { "name": "dracut-cmdline.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-initqueue.service": { "name": "dracut-initqueue.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-mount.service": { "name": "dracut-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-mount.service": { "name": "dracut-pre-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-pivot.service": { "name": "dracut-pre-pivot.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-trigger.service": { "name": "dracut-pre-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-udev.service": { "name": "dracut-pre-udev.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown.service": { "name": "dracut-shutdown.service", "source": "systemd", "state": "stopped", "status": "static" }, "emergency.service": { "name": "emergency.service", "source": "systemd", "state": "stopped", "status": "static" }, "fcoe.service": { "name": "fcoe.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "fstrim.service": { "name": "fstrim.service", "source": "systemd", "state": "inactive", "status": "static" }, "getty@.service": { "name": "getty@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "getty@tty1.service": { "name": "getty@tty1.service", "source": "systemd", "state": "running", "status": "active" }, "grub-boot-indeterminate.service": { "name": "grub-boot-indeterminate.service", "source": "systemd", "state": "inactive", "status": "static" }, "gssproxy.service": { "name": "gssproxy.service", "source": "systemd", "state": "running", "status": "disabled" }, "halt-local.service": { "name": "halt-local.service", "source": "systemd", "state": "inactive", "status": "static" }, "hv_kvp_daemon.service": { "name": "hv_kvp_daemon.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "import-state.service": { "name": "import-state.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "initrd-cleanup.service": { "name": "initrd-cleanup.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-parse-etc.service": { "name": "initrd-parse-etc.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-switch-root.service": { "name": "initrd-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-udevadm-cleanup-db.service": { "name": "initrd-udevadm-cleanup-db.service", "source": "systemd", "state": "stopped", "status": "static" }, "insights-client-boot.service": { "name": "insights-client-boot.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "insights-client-results.service": { "name": "insights-client-results.service", "source": "systemd", "state": "inactive", "status": "static" }, "insights-client.service": { "name": "insights-client.service", "source": "systemd", "state": "inactive", "status": "static" }, "irqbalance.service": { "name": "irqbalance.service", "source": "systemd", "state": "running", "status": "enabled" }, "iscsi-shutdown.service": { "name": "iscsi-shutdown.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "iscsi.service": { "name": "iscsi.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "iscsid.service": { "name": "iscsid.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "kdump.service": { "name": "kdump.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "kmod-static-nodes.service": { "name": "kmod-static-nodes.service", "source": "systemd", "state": "stopped", "status": "static" }, "kvm_stat.service": { "name": "kvm_stat.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "ldconfig.service": { "name": "ldconfig.service", "source": "systemd", "state": "stopped", "status": "static" }, "loadmodules.service": { "name": "loadmodules.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "lvm2-activation-early.service": { "name": "lvm2-activation-early.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "lvm2-activation.service": { "name": "lvm2-activation.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "lvm2-lvmpolld.service": { "name": "lvm2-lvmpolld.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-monitor.service": { "name": "lvm2-monitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "lvm2-pvscan@.service": { "name": "lvm2-pvscan@.service", "source": "systemd", "state": "unknown", "status": "static" }, "lvm2-pvscan@8:0.service": { "name": "lvm2-pvscan@8:0.service", "source": "systemd", "state": "stopped", "status": "active" }, "man-db-cache-update.service": { "name": "man-db-cache-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "man-db-restart-cache-update.service": { "name": "man-db-restart-cache-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "mdadm-grow-continue@.service": { "name": "mdadm-grow-continue@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdadm-last-resort@.service": { "name": "mdadm-last-resort@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdcheck_continue.service": { "name": "mdcheck_continue.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdcheck_start.service": { "name": "mdcheck_start.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmon@.service": { "name": "mdmon@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdmonitor-oneshot.service": { "name": "mdmonitor-oneshot.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmonitor.service": { "name": "mdmonitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "messagebus.service": { "name": "messagebus.service", "source": "systemd", "state": "active", "status": "static" }, "microcode.service": { "name": "microcode.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "multipathd.service": { "name": "multipathd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "ndctl-monitor.service": { "name": "ndctl-monitor.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "network.service": { "name": "network.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "nfs-blkmap.service": { "name": "nfs-blkmap.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-convert.service": { "name": "nfs-convert.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-idmapd.service": { "name": "nfs-idmapd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-mountd.service": { "name": "nfs-mountd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-server.service": { "name": "nfs-server.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-utils.service": { "name": "nfs-utils.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfsdcld.service": { "name": "nfsdcld.service", "source": "systemd", "state": "stopped", "status": "static" }, "nis-domainname.service": { "name": "nis-domainname.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "ntpd.service": { "name": "ntpd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ntpdate.service": { "name": "ntpdate.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "oddjobd.service": { "name": "oddjobd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "packagekit-offline-update.service": { "name": "packagekit-offline-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "packagekit.service": { "name": "packagekit.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-quit-wait.service": { "name": "plymouth-quit-wait.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "plymouth-start.service": { "name": "plymouth-start.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "polkit.service": { "name": "polkit.service", "source": "systemd", "state": "running", "status": "static" }, "qemu-guest-agent.service": { "name": "qemu-guest-agent.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "quotaon.service": { "name": "quotaon.service", "source": "systemd", "state": "inactive", "status": "static" }, "rbdmap.service": { "name": "rbdmap.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "rc-local.service": { "name": "rc-local.service", "source": "systemd", "state": "stopped", "status": "static" }, "rdisc.service": { "name": "rdisc.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rescue.service": { "name": "rescue.service", "source": "systemd", "state": "stopped", "status": "static" }, "rhcd.service": { "name": "rhcd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhsm-facts.service": { "name": "rhsm-facts.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhsm.service": { "name": "rhsm.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhsmcertd.service": { "name": "rhsmcertd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpc-gssd.service": { "name": "rpc-gssd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd-notify.service": { "name": "rpc-statd-notify.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd.service": { "name": "rpc-statd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpcbind.service": { "name": "rpcbind.service", "source": "systemd", "state": "running", "status": "enabled" }, "rsyslog.service": { "name": "rsyslog.service", "source": "systemd", "state": "running", "status": "enabled" }, "selinux-autorelabel-mark.service": { "name": "selinux-autorelabel-mark.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "selinux-autorelabel.service": { "name": "selinux-autorelabel.service", "source": "systemd", "state": "inactive", "status": "static" }, "serial-getty@.service": { "name": "serial-getty@.service", "source": "systemd", "state": "unknown", "status": "indirect" }, "serial-getty@ttyS0.service": { "name": "serial-getty@ttyS0.service", "source": "systemd", "state": "running", "status": "active" }, "snapd.seeded.service": { "name": "snapd.seeded.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sntp.service": { "name": "sntp.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sshd-keygen.service": { "name": "sshd-keygen.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sshd-keygen@.service": { "name": "sshd-keygen@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@ecdsa.service": { "name": "sshd-keygen@ecdsa.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd-keygen@ed25519.service": { "name": "sshd-keygen@ed25519.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd-keygen@rsa.service": { "name": "sshd-keygen@rsa.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd.service": { "name": "sshd.service", "source": "systemd", "state": "running", "status": "enabled" }, "sshd@.service": { "name": "sshd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "sssd-autofs.service": { "name": "sssd-autofs.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-kcm.service": { "name": "sssd-kcm.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "sssd-nss.service": { "name": "sssd-nss.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pac.service": { "name": "sssd-pac.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pam.service": { "name": "sssd-pam.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-ssh.service": { "name": "sssd-ssh.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-sudo.service": { "name": "sssd-sudo.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd.service": { "name": "sssd.service", "source": "systemd", "state": "running", "status": "enabled" }, "syslog.service": { "name": "syslog.service", "source": "systemd", "state": "active", "status": "enabled" }, "system-update-cleanup.service": { "name": "system-update-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-ask-password-console.service": { "name": "systemd-ask-password-console.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-wall.service": { "name": "systemd-ask-password-wall.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-backlight@.service": { "name": "systemd-backlight@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-binfmt.service": { "name": "systemd-binfmt.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-coredump@.service": { "name": "systemd-coredump@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-exit.service": { "name": "systemd-exit.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-firstboot.service": { "name": "systemd-firstboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck-root.service": { "name": "systemd-fsck-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck@.service": { "name": "systemd-fsck@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-fsck@dev-disk-by\\x2duuid-7B77\\x2d95E7.service": { "name": "systemd-fsck@dev-disk-by\\x2duuid-7B77\\x2d95E7.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-fsck@dev-vdb1.service": { "name": "systemd-fsck@dev-vdb1.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-fsck@dev-vdc1.service": { "name": "systemd-fsck@dev-vdc1.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-halt.service": { "name": "systemd-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hibernate-resume@.service": { "name": "systemd-hibernate-resume@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-hibernate.service": { "name": "systemd-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hostnamed.service": { "name": "systemd-hostnamed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hwdb-update.service": { "name": "systemd-hwdb-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-hybrid-sleep.service": { "name": "systemd-hybrid-sleep.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-initctl.service": { "name": "systemd-initctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-catalog-update.service": { "name": "systemd-journal-catalog-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-flush.service": { "name": "systemd-journal-flush.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journald.service": { "name": "systemd-journald.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-kexec.service": { "name": "systemd-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-localed.service": { "name": "systemd-localed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-logind.service": { "name": "systemd-logind.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-machine-id-commit.service": { "name": "systemd-machine-id-commit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-modules-load.service": { "name": "systemd-modules-load.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-networkd-wait-online.service": { "name": "systemd-networkd-wait-online.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "systemd-portabled.service": { "name": "systemd-portabled.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-poweroff.service": { "name": "systemd-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-quotacheck.service": { "name": "systemd-quotacheck.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-random-seed.service": { "name": "systemd-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-reboot.service": { "name": "systemd-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-remount-fs.service": { "name": "systemd-remount-fs.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-resolved.service": { "name": "systemd-resolved.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-rfkill.service": { "name": "systemd-rfkill.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend-then-hibernate.service": { "name": "systemd-suspend-then-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend.service": { "name": "systemd-suspend.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-sysctl.service": { "name": "systemd-sysctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-sysusers.service": { "name": "systemd-sysusers.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-timedated.service": { "name": "systemd-timedated.service", "source": "systemd", "state": "inactive", "status": "masked" }, "systemd-timesyncd.service": { "name": "systemd-timesyncd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "systemd-tmpfiles-clean.service": { "name": "systemd-tmpfiles-clean.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev.service": { "name": "systemd-tmpfiles-setup-dev.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup.service": { "name": "systemd-tmpfiles-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-settle.service": { "name": "systemd-udev-settle.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-trigger.service": { "name": "systemd-udev-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udevd.service": { "name": "systemd-udevd.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-update-done.service": { "name": "systemd-update-done.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp-runlevel.service": { "name": "systemd-update-utmp-runlevel.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp.service": { "name": "systemd-update-utmp.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-user-sessions.service": { "name": "systemd-user-sessions.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-vconsole-setup.service": { "name": "systemd-vconsole-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-volatile-root.service": { "name": "systemd-volatile-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "tcsd.service": { "name": "tcsd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "teamd@.service": { "name": "teamd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "timedatex.service": { "name": "timedatex.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "tuned.service": { "name": "tuned.service", "source": "systemd", "state": "running", "status": "enabled" }, "unbound-anchor.service": { "name": "unbound-anchor.service", "source": "systemd", "state": "stopped", "status": "static" }, "user-runtime-dir@.service": { "name": "user-runtime-dir@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user-runtime-dir@0.service": { "name": "user-runtime-dir@0.service", "source": "systemd", "state": "stopped", "status": "active" }, "user@.service": { "name": "user@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user@0.service": { "name": "user@0.service", "source": "systemd", "state": "running", "status": "active" }, "vdo-start-by-dev@.service": { "name": "vdo-start-by-dev@.service", "source": "systemd", "state": "unknown", "status": "static" }, "vdo.service": { "name": "vdo.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "ypbind.service": { "name": "ypbind.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "yppasswdd.service": { "name": "yppasswdd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ypserv.service": { "name": "ypserv.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ypxfrd.service": { "name": "ypxfrd.service", "source": "systemd", "state": "stopped", "status": "not-found" } } }, "changed": false } TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] ***** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:53 Thursday 21 July 2022 19:07:42 +0000 (0:00:01.733) 0:02:59.741 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:58 Thursday 21 July 2022 19:07:42 +0000 (0:00:00.062) 0:02:59.804 ********* TASK [fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64 Thursday 21 July 2022 19:07:42 +0000 (0:00:00.024) 0:02:59.828 ********* changed: [/cache/rhel-8-y.qcow2.snap] => { "actions": [ { "action": "destroy format", "device": "/dev/mapper/foo-test2", "fs_type": "ext4" }, { "action": "destroy device", "device": "/dev/mapper/foo-test2", "fs_type": null }, { "action": "destroy device", "device": "/dev/foo", "fs_type": null }, { "action": "destroy format", "device": "/dev/sda", "fs_type": "lvmpv" } ], "changed": true, "crypts": [], "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/vda2", "/dev/vda3", "/dev/sda", "/dev/sdb", "/dev/sdc", "/dev/nvme0n1", "/dev/nvme1n1", "/dev/nvme2n1", "/dev/vdb1", "/dev/vdc1", "/dev/vdd", "/dev/vde", "/dev/vdf" ], "mounts": [ { "fstype": "ext4", "path": "/opt/test2", "src": "/dev/mapper/foo-test2", "state": "absent" } ], "packages": [ "xfsprogs", "dosfstools", "e2fsprogs" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "absent", "type": "lvm", "volumes": [] } ], "volumes": [] } TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:78 Thursday 21 July 2022 19:07:44 +0000 (0:00:02.176) 0:03:02.004 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90 Thursday 21 July 2022 19:07:44 +0000 (0:00:00.041) 0:03:02.046 ********* TASK [fedora.linux_system_roles.storage : show blivet_output] ****************** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:96 Thursday 21 July 2022 19:07:44 +0000 (0:00:00.023) 0:03:02.069 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "blivet_output": { "actions": [ { "action": "destroy format", "device": "/dev/mapper/foo-test2", "fs_type": "ext4" }, { "action": "destroy device", "device": "/dev/mapper/foo-test2", "fs_type": null }, { "action": "destroy device", "device": "/dev/foo", "fs_type": null }, { "action": "destroy format", "device": "/dev/sda", "fs_type": "lvmpv" } ], "changed": true, "crypts": [], "failed": false, "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/vda2", "/dev/vda3", "/dev/sda", "/dev/sdb", "/dev/sdc", "/dev/nvme0n1", "/dev/nvme1n1", "/dev/nvme2n1", "/dev/vdb1", "/dev/vdc1", "/dev/vdd", "/dev/vde", "/dev/vdf" ], "mounts": [ { "fstype": "ext4", "path": "/opt/test2", "src": "/dev/mapper/foo-test2", "state": "absent" } ], "packages": [ "xfsprogs", "dosfstools", "e2fsprogs" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "absent", "type": "lvm", "volumes": [] } ], "volumes": [] } } TASK [fedora.linux_system_roles.storage : set the list of pools for test verification] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:101 Thursday 21 July 2022 19:07:44 +0000 (0:00:00.039) 0:03:02.108 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "absent", "type": "lvm", "volumes": [] } ] }, "changed": false } TASK [fedora.linux_system_roles.storage : set the list of volumes for test verification] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:105 Thursday 21 July 2022 19:07:44 +0000 (0:00:00.042) 0:03:02.151 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : remove obsolete mounts] ************** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:121 Thursday 21 July 2022 19:07:44 +0000 (0:00:00.042) 0:03:02.193 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount changed: [/cache/rhel-8-y.qcow2.snap] => (item={'src': '/dev/mapper/foo-test2', 'path': '/opt/test2', 'state': 'absent', 'fstype': 'ext4'}) => { "ansible_loop_var": "mount_info", "backup_file": "", "boot": "yes", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "ext4", "mount_info": { "fstype": "ext4", "path": "/opt/test2", "src": "/dev/mapper/foo-test2", "state": "absent" }, "name": "/opt/test2", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test2" } TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132 Thursday 21 July 2022 19:07:45 +0000 (0:00:00.472) 0:03:02.666 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.storage : set up new/current mounts] *********** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:137 Thursday 21 July 2022 19:07:45 +0000 (0:00:00.639) 0:03:03.305 ********* TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148 Thursday 21 July 2022 19:07:45 +0000 (0:00:00.042) 0:03:03.347 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.storage : retrieve facts for the /etc/crypttab file] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:156 Thursday 21 July 2022 19:07:46 +0000 (0:00:00.654) 0:03:04.002 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "stat": { "exists": false } } TASK [fedora.linux_system_roles.storage : manage /etc/crypttab to account for changes we just made] *** task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:161 Thursday 21 July 2022 19:07:46 +0000 (0:00:00.470) 0:03:04.473 ********* TASK [fedora.linux_system_roles.storage : Update facts] ************************ task path: /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:183 Thursday 21 July 2022 19:07:46 +0000 (0:00:00.024) 0:03:04.498 ********* ok: [/cache/rhel-8-y.qcow2.snap] META: role_complete for /cache/rhel-8-y.qcow2.snap TASK [include_tasks] *********************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/tests_lvm_percent_size.yml:171 Thursday 21 July 2022 19:07:47 +0000 (0:00:00.990) 0:03:05.489 ********* included: /tmp/tmpmnqeja9w/tests/storage/verify-role-results.yml for /cache/rhel-8-y.qcow2.snap TASK [Print out pool information] ********************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-role-results.yml:1 Thursday 21 July 2022 19:07:47 +0000 (0:00:00.054) 0:03:05.544 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "absent", "type": "lvm", "volumes": [] } ] } TASK [Print out volume information] ******************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-role-results.yml:6 Thursday 21 July 2022 19:07:48 +0000 (0:00:00.054) 0:03:05.598 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => {} TASK [Collect info about the volumes.] ***************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-role-results.yml:14 Thursday 21 July 2022 19:07:48 +0000 (0:00:00.036) 0:03:05.635 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "info": { "/dev/nvme0n1": { "fstype": "", "label": "", "name": "/dev/nvme0n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme1n1": { "fstype": "", "label": "", "name": "/dev/nvme1n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme2n1": { "fstype": "", "label": "", "name": "/dev/nvme2n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sda": { "fstype": "", "label": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdb": { "fstype": "", "label": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc": { "fstype": "", "label": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sr0": { "fstype": "iso9660", "label": "cidata", "name": "/dev/sr0", "size": "364K", "type": "rom", "uuid": "2022-07-21-19-04-29-00" }, "/dev/vda": { "fstype": "", "label": "", "name": "/dev/vda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vda1": { "fstype": "", "label": "", "name": "/dev/vda1", "size": "1M", "type": "partition", "uuid": "" }, "/dev/vda2": { "fstype": "vfat", "label": "", "name": "/dev/vda2", "size": "100M", "type": "partition", "uuid": "7B77-95E7" }, "/dev/vda3": { "fstype": "xfs", "label": "root", "name": "/dev/vda3", "size": "9.9G", "type": "partition", "uuid": "58dca2db-4ccd-4171-a017-dd53ad8c1c54" }, "/dev/vdb": { "fstype": "", "label": "", "name": "/dev/vdb", "size": "2G", "type": "disk", "uuid": "" }, "/dev/vdb1": { "fstype": "ext4", "label": "yumcache", "name": "/dev/vdb1", "size": "2G", "type": "partition", "uuid": "4d1ed9a1-5608-4ff8-ad89-21d95caa66fa" }, "/dev/vdc": { "fstype": "", "label": "", "name": "/dev/vdc", "size": "2G", "type": "disk", "uuid": "" }, "/dev/vdc1": { "fstype": "ext4", "label": "yumvarlib", "name": "/dev/vdc1", "size": "2G", "type": "partition", "uuid": "a7abcc97-edc5-476c-8ccd-694b4621fd28" }, "/dev/vdd": { "fstype": "", "label": "", "name": "/dev/vdd", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vde": { "fstype": "", "label": "", "name": "/dev/vde", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vdf": { "fstype": "", "label": "", "name": "/dev/vdf", "size": "10G", "type": "disk", "uuid": "" } } } TASK [Read the /etc/fstab file for volume existence] *************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-role-results.yml:19 Thursday 21 July 2022 19:07:48 +0000 (0:00:00.382) 0:03:06.018 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "cmd": [ "cat", "/etc/fstab" ], "delta": "0:00:00.002599", "end": "2022-07-21 15:07:48.909072", "rc": 0, "start": "2022-07-21 15:07:48.906473" } STDOUT: UUID=58dca2db-4ccd-4171-a017-dd53ad8c1c54 / xfs defaults 0 0 UUID=7B77-95E7 /boot/efi vfat defaults,uid=0,gid=0,umask=077,shortname=winnt 0 2 /dev/vdb1 /var/cache/dnf auto defaults,nofail,x-systemd.requires=cloud-init.service,_netdev,comment=cloudconfig 0 2 /dev/vdc1 /var/lib/dnf auto defaults,nofail,x-systemd.requires=cloud-init.service,_netdev,comment=cloudconfig 0 2 TASK [Read the /etc/crypttab file] ********************************************* task path: /tmp/tmpmnqeja9w/tests/storage/verify-role-results.yml:24 Thursday 21 July 2022 19:07:48 +0000 (0:00:00.374) 0:03:06.392 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "cmd": [ "cat", "/etc/crypttab" ], "delta": "0:00:00.002486", "end": "2022-07-21 15:07:49.283938", "failed_when_result": false, "rc": 1, "start": "2022-07-21 15:07:49.281452" } STDERR: cat: /etc/crypttab: No such file or directory MSG: non-zero return code TASK [Verify the volumes listed in storage_pools were correctly managed] ******* task path: /tmp/tmpmnqeja9w/tests/storage/verify-role-results.yml:33 Thursday 21 July 2022 19:07:49 +0000 (0:00:00.376) 0:03:06.769 ********* included: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool.yml for /cache/rhel-8-y.qcow2.snap => (item={'disks': ['sda'], 'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'name': 'foo', 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_metadata_version': None, 'state': 'absent', 'type': 'lvm', 'volumes': [], 'raid_chunk_size': None}) TASK [Set _storage_pool_tests] ************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool.yml:5 Thursday 21 July 2022 19:07:49 +0000 (0:00:00.065) 0:03:06.834 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_pool_tests": [ "members", "volumes" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool.yml:18 Thursday 21 July 2022 19:07:49 +0000 (0:00:00.036) 0:03:06.871 ********* included: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml for /cache/rhel-8-y.qcow2.snap => (item=members) included: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-volumes.yml for /cache/rhel-8-y.qcow2.snap => (item=volumes) TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:1 Thursday 21 July 2022 19:07:49 +0000 (0:00:00.050) 0:03:06.921 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_pv_count": "0", "_storage_test_pool_pvs_lvm": [] }, "changed": false } TASK [Get the canonical device path for each member device] ******************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:6 Thursday 21 July 2022 19:07:49 +0000 (0:00:00.053) 0:03:06.975 ********* TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:15 Thursday 21 July 2022 19:07:49 +0000 (0:00:00.023) 0:03:06.998 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "__pvs_lvm_len": "0" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:19 Thursday 21 July 2022 19:07:49 +0000 (0:00:00.092) 0:03:07.091 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_pool_pvs": [] }, "changed": false } TASK [Verify PV count] ********************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:23 Thursday 21 July 2022 19:07:49 +0000 (0:00:00.088) 0:03:07.180 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:29 Thursday 21 July 2022 19:07:49 +0000 (0:00:00.087) 0:03:07.267 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:33 Thursday 21 July 2022 19:07:49 +0000 (0:00:00.122) 0:03:07.389 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:37 Thursday 21 July 2022 19:07:49 +0000 (0:00:00.051) 0:03:07.441 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check the type of each PV] *********************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:41 Thursday 21 July 2022 19:07:49 +0000 (0:00:00.023) 0:03:07.464 ********* TASK [Check MD RAID] *********************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:50 Thursday 21 July 2022 19:07:49 +0000 (0:00:00.020) 0:03:07.484 ********* included: /tmp/tmpmnqeja9w/tests/storage/verify-pool-md.yml for /cache/rhel-8-y.qcow2.snap TASK [get information about RAID] ********************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-md.yml:6 Thursday 21 July 2022 19:07:49 +0000 (0:00:00.045) 0:03:07.530 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-md.yml:12 Thursday 21 July 2022 19:07:50 +0000 (0:00:00.026) 0:03:07.556 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-md.yml:16 Thursday 21 July 2022 19:07:50 +0000 (0:00:00.025) 0:03:07.581 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-md.yml:20 Thursday 21 July 2022 19:07:50 +0000 (0:00:00.025) 0:03:07.607 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-md.yml:24 Thursday 21 July 2022 19:07:50 +0000 (0:00:00.025) 0:03:07.632 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-md.yml:30 Thursday 21 July 2022 19:07:50 +0000 (0:00:00.025) 0:03:07.657 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-md.yml:36 Thursday 21 July 2022 19:07:50 +0000 (0:00:00.024) 0:03:07.682 ********* skipping: [/cache/rhel-8-y.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-md.yml:44 Thursday 21 July 2022 19:07:50 +0000 (0:00:00.025) 0:03:07.708 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_md_active_devices_re": null, "storage_test_md_metadata_version_re": null, "storage_test_md_spare_devices_re": null }, "changed": false } TASK [Check LVM RAID] ********************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:53 Thursday 21 July 2022 19:07:50 +0000 (0:00:00.035) 0:03:07.743 ********* included: /tmp/tmpmnqeja9w/tests/storage/verify-pool-members-lvmraid.yml for /cache/rhel-8-y.qcow2.snap TASK [Validate pool member LVM RAID settings] ********************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-members-lvmraid.yml:1 Thursday 21 July 2022 19:07:50 +0000 (0:00:00.045) 0:03:07.788 ********* TASK [Check Thin Pools] ******************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:56 Thursday 21 July 2022 19:07:50 +0000 (0:00:00.025) 0:03:07.813 ********* included: /tmp/tmpmnqeja9w/tests/storage/verify-pool-members-thin.yml for /cache/rhel-8-y.qcow2.snap TASK [Validate pool member thinpool settings] ********************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-members-thin.yml:1 Thursday 21 July 2022 19:07:50 +0000 (0:00:00.048) 0:03:07.862 ********* TASK [Check member encryption] ************************************************* task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:59 Thursday 21 July 2022 19:07:50 +0000 (0:00:00.025) 0:03:07.887 ********* included: /tmp/tmpmnqeja9w/tests/storage/verify-pool-members-encryption.yml for /cache/rhel-8-y.qcow2.snap TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-members-encryption.yml:4 Thursday 21 July 2022 19:07:50 +0000 (0:00:00.046) 0:03:07.934 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Validate pool member LUKS settings] ************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-members-encryption.yml:8 Thursday 21 July 2022 19:07:50 +0000 (0:00:00.048) 0:03:07.982 ********* TASK [Validate pool member crypttab entries] *********************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-members-encryption.yml:15 Thursday 21 July 2022 19:07:50 +0000 (0:00:00.022) 0:03:08.005 ********* TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-members-encryption.yml:22 Thursday 21 July 2022 19:07:50 +0000 (0:00:00.021) 0:03:08.026 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_crypttab_key_file": null }, "changed": false } TASK [Check VDO] *************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:62 Thursday 21 July 2022 19:07:50 +0000 (0:00:00.035) 0:03:08.062 ********* included: /tmp/tmpmnqeja9w/tests/storage/verify-pool-members-vdo.yml for /cache/rhel-8-y.qcow2.snap TASK [Validate pool member VDO settings] *************************************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-pool-members-vdo.yml:1 Thursday 21 July 2022 19:07:50 +0000 (0:00:00.050) 0:03:08.113 ********* TASK [set_fact] **************************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-members.yml:65 Thursday 21 July 2022 19:07:50 +0000 (0:00:00.023) 0:03:08.136 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "__pvs_lvm_len": null, "_storage_test_expected_pv_count": null, "_storage_test_expected_pv_type": null, "_storage_test_pool_pvs": [], "_storage_test_pool_pvs_lvm": [] }, "changed": false } TASK [verify the volumes] ****************************************************** task path: /tmp/tmpmnqeja9w/tests/storage/test-verify-pool-volumes.yml:3 Thursday 21 July 2022 19:07:50 +0000 (0:00:00.037) 0:03:08.174 ********* TASK [Verify the volumes with no pool were correctly managed] ****************** task path: /tmp/tmpmnqeja9w/tests/storage/verify-role-results.yml:43 Thursday 21 July 2022 19:07:50 +0000 (0:00:00.025) 0:03:08.199 ********* TASK [Clean up variable namespace] ********************************************* task path: /tmp/tmpmnqeja9w/tests/storage/verify-role-results.yml:53 Thursday 21 July 2022 19:07:50 +0000 (0:00:00.023) 0:03:08.223 ********* ok: [/cache/rhel-8-y.qcow2.snap] => { "ansible_facts": { "storage_test_blkinfo": null, "storage_test_crypttab": null, "storage_test_fstab": null }, "changed": false } META: ran handlers META: ran handlers PLAY RECAP ********************************************************************* /cache/rhel-8-y.qcow2.snap : ok=795 changed=10 unreachable=0 failed=1 skipped=565 rescued=1 ignored=0 Thursday 21 July 2022 19:07:50 +0000 (0:00:00.051) 0:03:08.274 ********* =============================================================================== fedora.linux_system_roles.storage : get required packages -------------- 11.63s /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:23 fedora.linux_system_roles.storage : get service facts ------------------- 2.77s /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:47 fedora.linux_system_roles.storage : get service facts ------------------- 2.76s /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:47 fedora.linux_system_roles.storage : get service facts ------------------- 2.73s /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:47 fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state --- 2.43s /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64 fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state --- 2.30s /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64 fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state --- 2.18s /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64 fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state --- 2.18s /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64 fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state --- 2.13s /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64 fedora.linux_system_roles.storage : get service facts ------------------- 1.91s /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:47 fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state --- 1.81s /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64 fedora.linux_system_roles.storage : get service facts ------------------- 1.80s /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:47 fedora.linux_system_roles.storage : make sure blivet is available ------- 1.77s /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:7 fedora.linux_system_roles.storage : get service facts ------------------- 1.75s /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:47 fedora.linux_system_roles.storage : get service facts ------------------- 1.73s /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:47 fedora.linux_system_roles.storage : get service facts ------------------- 1.73s /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:47 fedora.linux_system_roles.storage : get required packages --------------- 1.71s /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:23 fedora.linux_system_roles.storage : get required packages --------------- 1.69s /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:23 fedora.linux_system_roles.storage : get required packages --------------- 1.63s /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:23 fedora.linux_system_roles.storage : get required packages --------------- 1.56s /tmp/tmpp_vyhrdp/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:23