# STDOUT: ---v---v---v---v---v--- ansible-playbook [core 2.14.2] config file = /etc/ansible/ansible.cfg configured module search path = ['/home/jenkins/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules'] ansible python module location = /opt/ansible-2.14/lib/python3.9/site-packages/ansible ansible collection location = /WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/.collection executable location = /opt/ansible-2.14/bin/ansible-playbook python version = 3.9.13 (main, Jun 24 2022, 15:32:51) [GCC 8.5.0 20210514 (Red Hat 8.5.0-13)] (/opt/ansible-2.14/bin/python3.9) jinja version = 3.1.2 libyaml = True Using /etc/ansible/ansible.cfg as config file Skipping callback 'debug', as we already have a stdout callback. Skipping callback 'default', as we already have a stdout callback. Skipping callback 'minimal', as we already have a stdout callback. Skipping callback 'oneline', as we already have a stdout callback. PLAYBOOK: tests_cluster_destroy.yml ******************************************** 2 plays in /WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/tests/tests_cluster_destroy.yml PLAY [all] ********************************************************************* TASK [Include vault variables] ************************************************* task path: /WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/tests/tests_cluster_destroy.yml:4 Saturday 11 February 2023 19:47:21 +0000 (0:00:00.778) 0:00:00.778 ***** ok: [sut] => { "ansible_facts": { "ha_cluster_hacluster_password": { "__ansible_vault": "$ANSIBLE_VAULT;1.1;AES256\n31303833633366333561656439323930303361333161363239346166656537323933313436\n3432386236656563343237306335323637396239616230353561330a313731623238393238\n62343064666336643930663239383936616465643134646536656532323461356237646133\n3761616633323839633232353637366266350a313163633236376666653238633435306565\n3264623032333736393535663833\n" } }, "ansible_included_var_files": [ "/WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/tests/vars/vault-variables.yml" ], "changed": false } PLAY [Deconfigure cluster] ***************************************************** TASK [Gathering Facts] ********************************************************* task path: /WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/tests/tests_cluster_destroy.yml:9 Saturday 11 February 2023 19:47:22 +0000 (0:00:01.170) 0:00:01.948 ***** ok: [sut] TASK [Set up test environment] ************************************************* task path: /WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/tests/tests_cluster_destroy.yml:18 Saturday 11 February 2023 19:47:43 +0000 (0:00:20.547) 0:00:22.496 ***** TASK [fedora.linux_system_roles.ha_cluster : Set node name to 'localhost' for single-node clusters] *** task path: /WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/.collection/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/test_setup.yml:9 Saturday 11 February 2023 19:47:45 +0000 (0:00:02.229) 0:00:24.726 ***** ok: [sut] => { "ansible_facts": { "inventory_hostname": "localhost" }, "changed": false } TASK [fedora.linux_system_roles.ha_cluster : Ensure facts used by tests] ******* task path: /WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/.collection/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/test_setup.yml:14 Saturday 11 February 2023 19:47:48 +0000 (0:00:02.605) 0:00:27.332 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.ha_cluster : Do not try to enable RHEL repositories] *** task path: /WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/.collection/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/test_setup.yml:20 Saturday 11 February 2023 19:47:49 +0000 (0:00:00.976) 0:00:28.308 ***** ok: [sut] => { "ansible_facts": { "ha_cluster_enable_repos": false }, "changed": false } TASK [Run HA Cluster role] ***************************************************** task path: /WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/tests/tests_cluster_destroy.yml:23 Saturday 11 February 2023 19:47:51 +0000 (0:00:02.292) 0:00:30.601 ***** TASK [fedora.linux_system_roles.ha_cluster : Set platform/version specific variables] *** task path: /WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/.collection/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:3 Saturday 11 February 2023 19:47:57 +0000 (0:00:05.412) 0:00:36.013 ***** included: /WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/.collection/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/set_vars.yml for sut TASK [fedora.linux_system_roles.ha_cluster : Ensure ansible_facts used by role] *** task path: /WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/.collection/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/set_vars.yml:2 Saturday 11 February 2023 19:47:58 +0000 (0:00:01.381) 0:00:37.394 ***** ok: [sut] TASK [fedora.linux_system_roles.ha_cluster : Set platform/version specific variables] *** task path: /WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/.collection/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/set_vars.yml:7 Saturday 11 February 2023 19:48:01 +0000 (0:00:03.594) 0:00:40.989 ***** skipping: [sut] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [sut] => (item=RedHat_9.yml) => { "ansible_facts": { "__ha_cluster_repos": [ { "id": "rhel-9-for-{{ ansible_architecture }}-highavailability-rpms", "name": "High Availability" } ] }, "ansible_included_var_files": [ "/WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/.collection/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/vars/RedHat_9.yml" ], "ansible_loop_var": "item", "changed": false, "item": "RedHat_9.yml" } skipping: [sut] => (item=RedHat_9.2.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat_9.2.yml", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.ha_cluster : Check and prepare role variables] *** task path: /WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/.collection/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:6 Saturday 11 February 2023 19:48:05 +0000 (0:00:03.987) 0:00:44.976 ***** included: /WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/.collection/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/check-and-prepare-role-variables.yml for sut TASK [fedora.linux_system_roles.ha_cluster : Fail if passwords are not specified] *** task path: /WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/.collection/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/check-and-prepare-role-variables.yml:5 Saturday 11 February 2023 19:48:07 +0000 (0:00:01.433) 0:00:46.409 ***** skipping: [sut] => (item=ha_cluster_hacluster_password) => { "ansible_loop_var": "item", "changed": false, "item": "ha_cluster_hacluster_password", "skip_reason": "Conditional result was False" } skipping: [sut] => { "changed": false } MSG: All items skipped TASK [fedora.linux_system_roles.ha_cluster : Fail if nodes do not have the same number of SBD devices specified] *** task path: /WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/.collection/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/check-and-prepare-role-variables.yml:15 Saturday 11 February 2023 19:48:15 +0000 (0:00:08.154) 0:00:54.564 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.ha_cluster : Fail if configuring qnetd on a cluster node] *** task path: /WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/.collection/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/check-and-prepare-role-variables.yml:30 Saturday 11 February 2023 19:48:17 +0000 (0:00:01.832) 0:00:56.396 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.ha_cluster : Discover cluster node names] ****** task path: /WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/.collection/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/check-and-prepare-role-variables.yml:40 Saturday 11 February 2023 19:48:18 +0000 (0:00:01.254) 0:00:57.651 ***** ok: [sut] => { "ansible_facts": { "__ha_cluster_node_name": "localhost" }, "changed": false } TASK [fedora.linux_system_roles.ha_cluster : Collect cluster node names] ******* task path: /WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/.collection/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/check-and-prepare-role-variables.yml:44 Saturday 11 February 2023 19:48:21 +0000 (0:00:02.527) 0:01:00.179 ***** ok: [sut] => { "ansible_facts": { "__ha_cluster_all_node_names": [ "localhost" ] }, "changed": false } TASK [fedora.linux_system_roles.ha_cluster : Extract qdevice settings] ********* task path: /WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/.collection/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/check-and-prepare-role-variables.yml:48 Saturday 11 February 2023 19:48:24 +0000 (0:00:03.661) 0:01:03.840 ***** ok: [sut] => { "ansible_facts": { "__ha_cluster_qdevice_host": "", "__ha_cluster_qdevice_in_use": false, "__ha_cluster_qdevice_model": "", "__ha_cluster_qdevice_pcs_address": "" }, "changed": false } TASK [fedora.linux_system_roles.ha_cluster : Figure out if ATB needs to be enabled for SBD] *** task path: /WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/.collection/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/check-and-prepare-role-variables.yml:57 Saturday 11 February 2023 19:48:36 +0000 (0:00:11.362) 0:01:15.202 ***** ok: [sut] => { "ansible_facts": { "__ha_cluster_sbd_needs_atb": false }, "changed": false } TASK [fedora.linux_system_roles.ha_cluster : Fail if SBD needs ATB enabled and the user configured ATB to be disabled] *** task path: /WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/.collection/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/check-and-prepare-role-variables.yml:67 Saturday 11 February 2023 19:48:41 +0000 (0:00:05.030) 0:01:20.233 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.ha_cluster : Fail if ha_cluster_pcsd_public_key_src and ha_cluster_pcsd_private_key_src are set along with ha_cluster_pcsd_certificates] *** task path: /WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/.collection/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/check-and-prepare-role-variables.yml:74 Saturday 11 February 2023 19:48:43 +0000 (0:00:02.542) 0:01:22.776 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.ha_cluster : Enable package repositories] ****** task path: /WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/.collection/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:9 Saturday 11 February 2023 19:48:45 +0000 (0:00:02.011) 0:01:24.787 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.ha_cluster : Install role essential packages] *** task path: /WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/.collection/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:14 Saturday 11 February 2023 19:48:47 +0000 (0:00:01.721) 0:01:26.509 ***** changed: [sut] => { "changed": true, "rc": 0, "results": [ "Installed: libknet1-1.25-2.el9.x86_64", "Installed: libknet1-compress-bzip2-plugin-1.25-2.el9.x86_64", "Installed: libknet1-compress-lz4-plugin-1.25-2.el9.x86_64", "Installed: libknet1-compress-lzma-plugin-1.25-2.el9.x86_64", "Installed: libknet1-compress-lzo2-plugin-1.25-2.el9.x86_64", "Installed: corosynclib-3.1.7-1.el9.x86_64", "Installed: libknet1-compress-plugins-all-1.25-2.el9.x86_64", "Installed: libknet1-compress-zlib-plugin-1.25-2.el9.x86_64", "Installed: libknet1-compress-zstd-plugin-1.25-2.el9.x86_64", "Installed: libknet1-crypto-nss-plugin-1.25-2.el9.x86_64", "Installed: libknet1-crypto-openssl-plugin-1.25-2.el9.x86_64", "Installed: libknet1-crypto-plugins-all-1.25-2.el9.x86_64", "Installed: libknet1-plugins-all-1.25-2.el9.x86_64", "Installed: libnozzle1-1.25-2.el9.x86_64", "Installed: ruby-3.0.4-160.el9_0.x86_64", "Installed: pacemaker-2.1.5-5.el9.x86_64", "Installed: pacemaker-cli-2.1.5-5.el9.x86_64", "Installed: pacemaker-cluster-libs-2.1.5-5.el9.x86_64", "Installed: nspr-4.34.0-14.el9_0.x86_64", "Installed: ruby-default-gems-3.0.4-160.el9_0.noarch", "Installed: pacemaker-libs-2.1.5-5.el9.x86_64", "Installed: nss-3.79.0-14.el9_0.x86_64", "Installed: device-mapper-event-9:1.02.187-5.el9.x86_64", "Installed: device-mapper-event-libs-9:1.02.187-5.el9.x86_64", "Installed: ruby-libs-3.0.4-160.el9_0.x86_64", "Installed: pacemaker-schemas-2.1.5-5.el9.noarch", "Installed: libqb-2.0.6-2.el9.x86_64", "Installed: nss-softokn-3.79.0-14.el9_0.x86_64", "Installed: rubygem-bigdecimal-3.0.0-160.el9_0.x86_64", "Installed: device-mapper-persistent-data-0.9.0-13.el9.x86_64", "Installed: nss-softokn-freebl-3.79.0-14.el9_0.x86_64", "Installed: rubygem-bundler-2.2.33-160.el9_0.noarch", "Installed: pcs-0.11.4-4.el9.x86_64", "Installed: nss-sysinit-3.79.0-14.el9_0.x86_64", "Installed: nss-tools-3.79.0-14.el9_0.x86_64", "Installed: pkgconf-1.7.3-10.el9.x86_64", "Installed: pkgconf-m4-1.7.3-10.el9.noarch", "Installed: pkgconf-pkg-config-1.7.3-10.el9.x86_64", "Installed: nss-util-3.79.0-14.el9_0.x86_64", "Installed: perl-TimeDate-1:2.33-6.el9.noarch", "Installed: rubygem-io-console-0.5.7-160.el9_0.x86_64", "Installed: rubygem-json-2.5.1-160.el9_0.x86_64", "Installed: libaio-0.3.111-13.el9.x86_64", "Installed: rubygem-psych-3.3.2-160.el9_0.x86_64", "Installed: resource-agents-4.10.0-34.el9.x86_64", "Installed: python3-cffi-1.14.5-5.el9.x86_64", "Installed: lvm2-9:2.03.17-5.el9.x86_64", "Installed: rubygem-rdoc-6.3.3-160.el9_0.noarch", "Installed: lvm2-libs-9:2.03.17-5.el9.x86_64", "Installed: libpkgconf-1.7.3-10.el9.x86_64", "Installed: python3-cryptography-36.0.1-2.el9.x86_64", "Installed: rubygem-rexml-3.2.5-160.el9_0.noarch", "Installed: rubygems-3.2.33-160.el9_0.noarch", "Installed: redhat-logos-90.4-1.el9.x86_64", "Installed: corosync-3.1.7-1.el9.x86_64", "Installed: python3-pycurl-7.43.0.6-8.el9.x86_64", "Installed: corosync-qnetd-3.0.2-1.el9.x86_64", "Installed: net-snmp-libs-1:5.9.1-7.el9_0.1.x86_64", "Installed: bzip2-1.0.8-8.el9.x86_64", "Installed: net-tools-2.0-0.62.20160912git.el9.x86_64", "Installed: python3-ply-3.11-14.el9.noarch", "Installed: python3-pycparser-2.20-6.el9.noarch" ] } TASK [fedora.linux_system_roles.ha_cluster : Set hacluster password] *********** task path: /WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/.collection/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:21 Saturday 11 February 2023 19:49:20 +0000 (0:00:33.434) 0:01:59.943 ***** changed: [sut] => { "append": false, "changed": true, "comment": "cluster user", "group": 189, "home": "/home/hacluster", "move_home": false, "name": "hacluster", "password": "NOT_LOGGING_PASSWORD", "shell": "/sbin/nologin", "state": "present", "uid": 189 } TASK [fedora.linux_system_roles.ha_cluster : Configure pcs / pcsd] ************* task path: /WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/.collection/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:28 Saturday 11 February 2023 19:49:43 +0000 (0:00:22.720) 0:02:22.663 ***** included: /WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/.collection/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/pcs-configure-pcs-pcsd.yml for sut TASK [fedora.linux_system_roles.ha_cluster : Stop pcsd] ************************ task path: /WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/.collection/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/pcs-configure-pcs-pcsd.yml:6 Saturday 11 February 2023 19:49:47 +0000 (0:00:03.496) 0:02:26.160 ***** ok: [sut] => { "changed": false, "name": "pcsd", "state": "stopped", "status": { "AccessSELinuxContext": "system_u:object_r:cluster_unit_file_t:s0", "ActiveEnterTimestampMonotonic": "0", "ActiveExitTimestampMonotonic": "0", "ActiveState": "inactive", "After": "sysinit.target basic.target pcsd-ruby.service systemd-journald.socket network-online.target system.slice", "AllowIsolate": "no", "AssertResult": "no", "AssertTimestampMonotonic": "0", "Before": "shutdown.target", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "CPUAccounting": "yes", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "[not set]", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "no", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore", "CleanResult": "success", "CollectMode": "inactive", "ConditionResult": "no", "ConditionTimestampMonotonic": "0", "ConfigurationDirectoryMode": "0755", "Conflicts": "shutdown.target", "ConsistsOf": "pcsd-ruby.service", "ControlGroupId": "0", "ControlPID": "0", "CoredumpFilter": "0x33", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "Delegate": "no", "Description": "PCS GUI and remote configuration interface", "DevicePolicy": "auto", "Documentation": "\"man:pcsd(8)\" \"man:pcs(8)\"", "DynamicUser": "no", "EnvironmentFiles": "/etc/sysconfig/pcsd (ignore_errors=no)", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "0", "ExecMainStartTimestampMonotonic": "0", "ExecMainStatus": "0", "ExecStart": "{ path=/usr/sbin/pcsd ; argv[]=/usr/sbin/pcsd ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStartEx": "{ path=/usr/sbin/pcsd ; argv[]=/usr/sbin/pcsd ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExitType": "main", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FinalKillSignal": "9", "FragmentPath": "/usr/lib/systemd/system/pcsd.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOReadBytes": "18446744073709551615", "IOReadOperations": "18446744073709551615", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "IOWeight": "[not set]", "IOWriteBytes": "18446744073709551615", "IOWriteOperations": "18446744073709551615", "IPAccounting": "no", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "Id": "pcsd.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestampMonotonic": "0", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "mixed", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "0", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "8388608", "LimitMEMLOCKSoft": "8388608", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "524288", "LimitNOFILESoft": "1024", "LimitNPROC": "13988", "LimitNPROCSoft": "13988", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "13988", "LimitSIGPENDINGSoft": "13988", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "0", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "ManagedOOMSwap": "auto", "MemoryAccounting": "yes", "MemoryAvailable": "infinity", "MemoryCurrent": "[not set]", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemorySwapMax": "infinity", "MountAPIVFS": "no", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAPolicy": "n/a", "Names": "pcsd.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "main", "OOMPolicy": "stop", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "OnSuccessJobMode": "fail", "Perpetual": "no", "PrivateDevices": "no", "PrivateIPC": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProcSubset": "all", "ProtectClock": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectHostname": "no", "ProtectKernelLogs": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectProc": "default", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "ReloadResult": "success", "RemainAfterExit": "no", "RemoveIPC": "no", "Requires": "sysinit.target pcsd-ruby.service network-online.target system.slice", "Restart": "no", "RestartKillSignal": "15", "RestartUSec": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "Slice": "system.slice", "StandardError": "inherit", "StandardInput": "null", "StandardOutput": "journal", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StateChangeTimestampMonotonic": "0", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "dead", "SuccessAction": "none", "SyslogFacility": "3", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "2147483646", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "[not set]", "TasksMax": "22380", "TimeoutAbortUSec": "1min 30s", "TimeoutCleanUSec": "infinity", "TimeoutStartFailureMode": "terminate", "TimeoutStartUSec": "1min 30s", "TimeoutStopFailureMode": "terminate", "TimeoutStopUSec": "1min 30s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "notify", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "disabled", "UnitFileState": "disabled", "UtmpMode": "init", "WatchdogSignal": "6", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "infinity" } } TASK [fedora.linux_system_roles.ha_cluster : Regenerate pcsd TLS certificate and key] *** task path: /WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/.collection/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/pcs-configure-pcs-pcsd.yml:11 Saturday 11 February 2023 19:50:07 +0000 (0:00:20.095) 0:02:46.256 ***** skipping: [sut] => (item=/var/lib/pcsd/pcsd.key) => { "ansible_loop_var": "item", "changed": false, "item": "/var/lib/pcsd/pcsd.key", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=/var/lib/pcsd/pcsd.crt) => { "ansible_loop_var": "item", "changed": false, "item": "/var/lib/pcsd/pcsd.crt", "skip_reason": "Conditional result was False" } skipping: [sut] => { "changed": false } MSG: All items skipped TASK [fedora.linux_system_roles.ha_cluster : Get the stat of /var/lib/pcsd] **** task path: /WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/.collection/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/pcs-configure-pcs-pcsd.yml:25 Saturday 11 February 2023 19:50:08 +0000 (0:00:01.626) 0:02:47.882 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.ha_cluster : Allow certmonger to write into pcsd's certificate directory] *** task path: /WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/.collection/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/pcs-configure-pcs-pcsd.yml:30 Saturday 11 February 2023 19:50:09 +0000 (0:00:00.979) 0:02:48.862 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure the name of ha_cluster_pcsd_certificates is /var/lib/pcsd/pcsd; Create certificates using the certificate role] *** task path: /WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/.collection/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/pcs-configure-pcs-pcsd.yml:37 Saturday 11 February 2023 19:50:10 +0000 (0:00:00.455) 0:02:49.317 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.ha_cluster : Set pcsd's certificate directory back to cluster_var_lib_t] *** task path: /WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/.collection/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/pcs-configure-pcs-pcsd.yml:49 Saturday 11 February 2023 19:50:10 +0000 (0:00:00.447) 0:02:49.764 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.ha_cluster : Distribute pcsd TLS private key] *** task path: /WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/.collection/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/pcs-configure-pcs-pcsd.yml:64 Saturday 11 February 2023 19:50:11 +0000 (0:00:00.419) 0:02:50.184 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.ha_cluster : Distribute pcsd TLS certificate] *** task path: /WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/.collection/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/pcs-configure-pcs-pcsd.yml:71 Saturday 11 February 2023 19:50:11 +0000 (0:00:00.237) 0:02:50.421 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.ha_cluster : Distribute pcs_settings.conf] ***** task path: /WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/.collection/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/pcs-configure-pcs-pcsd.yml:79 Saturday 11 February 2023 19:50:11 +0000 (0:00:00.325) 0:02:50.747 ***** changed: [sut] => { "changed": true, "checksum": "b504e1b9c9aa23803dd6f95e66c757088b08551d", "dest": "/var/lib/pcsd/pcs_settings.conf", "gid": 0, "group": "root", "md5sum": "087ff556d850518c8fff5ad1179d8817", "mode": "0644", "owner": "root", "secontext": "system_u:object_r:cluster_var_lib_t:s0", "size": 359, "src": "/root/.ansible/tmp/ansible-tmp-1676145011.8582773-19254-149985950659723/source", "state": "file", "uid": 0 } TASK [fedora.linux_system_roles.ha_cluster : Start pcsd with updated config files and configure it to start on boot] *** task path: /WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/.collection/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/pcs-configure-pcs-pcsd.yml:88 Saturday 11 February 2023 19:50:14 +0000 (0:00:02.996) 0:02:53.743 ***** changed: [sut] => { "changed": true, "enabled": true, "name": "pcsd", "state": "started", "status": { "AccessSELinuxContext": "system_u:object_r:cluster_unit_file_t:s0", "ActiveEnterTimestampMonotonic": "0", "ActiveExitTimestampMonotonic": "0", "ActiveState": "inactive", "After": "pcsd-ruby.service basic.target system.slice network-online.target sysinit.target systemd-journald.socket", "AllowIsolate": "no", "AssertResult": "no", "AssertTimestampMonotonic": "0", "Before": "shutdown.target", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "CPUAccounting": "yes", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "[not set]", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "no", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore", "CleanResult": "success", "CollectMode": "inactive", "ConditionResult": "no", "ConditionTimestampMonotonic": "0", "ConfigurationDirectoryMode": "0755", "Conflicts": "shutdown.target", "ConsistsOf": "pcsd-ruby.service", "ControlGroupId": "0", "ControlPID": "0", "CoredumpFilter": "0x33", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "Delegate": "no", "Description": "PCS GUI and remote configuration interface", "DevicePolicy": "auto", "Documentation": "\"man:pcsd(8)\" \"man:pcs(8)\"", "DynamicUser": "no", "EnvironmentFiles": "/etc/sysconfig/pcsd (ignore_errors=no)", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "0", "ExecMainStartTimestampMonotonic": "0", "ExecMainStatus": "0", "ExecStart": "{ path=/usr/sbin/pcsd ; argv[]=/usr/sbin/pcsd ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStartEx": "{ path=/usr/sbin/pcsd ; argv[]=/usr/sbin/pcsd ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExitType": "main", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FinalKillSignal": "9", "FragmentPath": "/usr/lib/systemd/system/pcsd.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOReadBytes": "18446744073709551615", "IOReadOperations": "18446744073709551615", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "IOWeight": "[not set]", "IOWriteBytes": "18446744073709551615", "IOWriteOperations": "18446744073709551615", "IPAccounting": "no", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "Id": "pcsd.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestampMonotonic": "0", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "mixed", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "0", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "8388608", "LimitMEMLOCKSoft": "8388608", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "524288", "LimitNOFILESoft": "1024", "LimitNPROC": "13988", "LimitNPROCSoft": "13988", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "13988", "LimitSIGPENDINGSoft": "13988", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "0", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "ManagedOOMSwap": "auto", "MemoryAccounting": "yes", "MemoryAvailable": "infinity", "MemoryCurrent": "[not set]", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemorySwapMax": "infinity", "MountAPIVFS": "no", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAPolicy": "n/a", "Names": "pcsd.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "main", "OOMPolicy": "stop", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "OnSuccessJobMode": "fail", "Perpetual": "no", "PrivateDevices": "no", "PrivateIPC": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProcSubset": "all", "ProtectClock": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectHostname": "no", "ProtectKernelLogs": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectProc": "default", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "ReloadResult": "success", "RemainAfterExit": "no", "RemoveIPC": "no", "Requires": "pcsd-ruby.service system.slice network-online.target sysinit.target", "Restart": "no", "RestartKillSignal": "15", "RestartUSec": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "Slice": "system.slice", "StandardError": "inherit", "StandardInput": "null", "StandardOutput": "journal", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StateChangeTimestampMonotonic": "0", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "dead", "SuccessAction": "none", "SyslogFacility": "3", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "2147483646", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "[not set]", "TasksMax": "22380", "TimeoutAbortUSec": "1min 30s", "TimeoutCleanUSec": "infinity", "TimeoutStartFailureMode": "terminate", "TimeoutStartUSec": "1min 30s", "TimeoutStopFailureMode": "terminate", "TimeoutStopUSec": "1min 30s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "notify", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "disabled", "UnitFileState": "disabled", "UtmpMode": "init", "WatchdogSignal": "6", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "infinity" } } TASK [fedora.linux_system_roles.ha_cluster : Fetch pcs capabilities] *********** task path: /WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/.collection/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/pcs-configure-pcs-pcsd.yml:94 Saturday 11 February 2023 19:50:18 +0000 (0:00:04.135) 0:02:57.878 ***** ok: [sut] => { "changed": false, "cmd": [ "pcs", "--version", "--full" ], "delta": "0:00:00.635130", "end": "2023-02-11 19:50:20.311968", "rc": 0, "start": "2023-02-11 19:50:19.676838" } STDOUT: 0.11.4 booth booth.enable-authfile.set booth.enable-authfile.unset cluster.config.backup-local cluster.config.restore-cluster cluster.config.restore-local cluster.config.uuid cluster.create cluster.create.enable cluster.create.local cluster.create.no-keys-sync cluster.create.separated-name-and-address cluster.create.start cluster.create.start.wait cluster.create.transport.knet cluster.create.transport.udp-udpu cluster.create.transport.udp-udpu.no-rrp cluster.destroy cluster.destroy.all cluster.report cluster.verify corosync.authkey.update corosync.config.get corosync.config.get.struct corosync.config.reload corosync.config.sync-to-local-cluster corosync.config.update corosync.link.add corosync.link.remove corosync.link.remove.list corosync.link.update corosync.qdevice corosync.qdevice.model.net corosync.quorum corosync.quorum.device corosync.quorum.device.heuristics corosync.quorum.device.model.net corosync.quorum.set-expected-votes-runtime corosync.quorum.status corosync.quorum.unblock corosync.totem.block_unlisted_ips corosync.uidgid node.add node.add.enable node.add.separated-name-and-address node.add.start node.add.start.wait node.attributes node.attributes.set-list-for-node node.confirm-off node.fence node.guest node.kill node.maintenance node.maintenance.all node.maintenance.list node.maintenance.wait node.remote node.remote.onfail-demote node.remove node.remove-from-caches node.remove.list node.standby node.standby.all node.standby.list node.standby.wait node.start-stop-enable-disable node.start-stop-enable-disable.all node.start-stop-enable-disable.list node.start-stop-enable-disable.start-wait node.utilization node.utilization.set-list-for-node pcmk.acl.enable-disable pcmk.acl.group pcmk.acl.role pcmk.acl.role.create-with-permissions pcmk.acl.role.delete-with-users-groups pcmk.acl.user pcmk.alert pcmk.cib.checkpoints pcmk.cib.checkpoints.diff pcmk.cib.edit pcmk.cib.get pcmk.cib.get.scope pcmk.cib.roles.promoted-unpromoted pcmk.cib.set pcmk.constraint.colocation.set pcmk.constraint.colocation.set.options pcmk.constraint.colocation.simple pcmk.constraint.colocation.simple.options pcmk.constraint.hide-expired pcmk.constraint.location.simple pcmk.constraint.location.simple.options pcmk.constraint.location.simple.resource-regexp pcmk.constraint.location.simple.rule pcmk.constraint.location.simple.rule.node-attr-type-number pcmk.constraint.location.simple.rule.options pcmk.constraint.location.simple.rule.rule-add-remove pcmk.constraint.no-autocorrect pcmk.constraint.order.set pcmk.constraint.order.set.options pcmk.constraint.order.simple pcmk.constraint.order.simple.options pcmk.constraint.ticket.set pcmk.constraint.ticket.set.options pcmk.constraint.ticket.simple pcmk.constraint.ticket.simple.constraint-id pcmk.properties.cluster pcmk.properties.operation-defaults pcmk.properties.operation-defaults.multiple pcmk.properties.operation-defaults.rule pcmk.properties.operation-defaults.rule-rsc-op pcmk.properties.operation-defaults.rule.hide-expired pcmk.properties.operation-defaults.rule.node-attr-type-number pcmk.properties.resource-defaults pcmk.properties.resource-defaults.multiple pcmk.properties.resource-defaults.rule pcmk.properties.resource-defaults.rule-rsc-op pcmk.properties.resource-defaults.rule.hide-expired pcmk.properties.resource-defaults.rule.node-attr-type-number pcmk.resource.ban-move-clear pcmk.resource.ban-move-clear.clear-expired pcmk.resource.bundle pcmk.resource.bundle.container-docker pcmk.resource.bundle.container-docker.promoted-max pcmk.resource.bundle.container-podman pcmk.resource.bundle.container-podman.promoted-max pcmk.resource.bundle.container-rkt pcmk.resource.bundle.container-rkt.promoted-max pcmk.resource.bundle.reset pcmk.resource.bundle.wait pcmk.resource.cleanup pcmk.resource.cleanup.one-resource pcmk.resource.cleanup.strict pcmk.resource.clone pcmk.resource.clone.custom-id pcmk.resource.clone.meta-in-create pcmk.resource.clone.wait pcmk.resource.config.output-formats pcmk.resource.create pcmk.resource.create.clone.custom-id pcmk.resource.create.in-existing-bundle pcmk.resource.create.meta pcmk.resource.create.no-master pcmk.resource.create.operations pcmk.resource.create.operations.onfail-demote pcmk.resource.create.promotable pcmk.resource.create.promotable.custom-id pcmk.resource.create.wait pcmk.resource.debug pcmk.resource.delete pcmk.resource.disable.safe pcmk.resource.disable.safe.brief pcmk.resource.disable.safe.tag pcmk.resource.disable.simulate pcmk.resource.disable.simulate.brief pcmk.resource.disable.simulate.tag pcmk.resource.enable-disable pcmk.resource.enable-disable.list pcmk.resource.enable-disable.tag pcmk.resource.enable-disable.wait pcmk.resource.failcount pcmk.resource.group pcmk.resource.group.add-remove-list pcmk.resource.group.wait pcmk.resource.manage-unmanage pcmk.resource.manage-unmanage.list pcmk.resource.manage-unmanage.tag pcmk.resource.manage-unmanage.with-monitor pcmk.resource.move.autoclean pcmk.resource.move.autoclean.default pcmk.resource.promotable pcmk.resource.promotable.custom-id pcmk.resource.promotable.meta-in-create pcmk.resource.promotable.wait pcmk.resource.refresh pcmk.resource.refresh.one-resource pcmk.resource.refresh.strict pcmk.resource.relations pcmk.resource.relocate pcmk.resource.restart pcmk.resource.update pcmk.resource.update-meta pcmk.resource.update-meta.list pcmk.resource.update-meta.wait pcmk.resource.update-operations pcmk.resource.update-operations.onfail-demote pcmk.resource.update.meta pcmk.resource.update.operations pcmk.resource.update.operations.onfail-demote pcmk.resource.update.wait pcmk.resource.utilization pcmk.resource.utilization-set-list-for-resource pcmk.stonith.cleanup pcmk.stonith.cleanup.one-resource pcmk.stonith.cleanup.strict pcmk.stonith.create pcmk.stonith.create.in-group pcmk.stonith.create.meta pcmk.stonith.create.operations pcmk.stonith.create.operations.onfail-demote pcmk.stonith.create.wait pcmk.stonith.delete pcmk.stonith.enable-disable pcmk.stonith.enable-disable.list pcmk.stonith.enable-disable.wait pcmk.stonith.history.cleanup pcmk.stonith.history.show pcmk.stonith.history.update pcmk.stonith.levels pcmk.stonith.levels.add-remove-devices-list pcmk.stonith.levels.clear pcmk.stonith.levels.node-attr pcmk.stonith.levels.node-regexp pcmk.stonith.levels.verify pcmk.stonith.refresh pcmk.stonith.refresh.one-resource pcmk.stonith.refresh.strict pcmk.stonith.update pcmk.stonith.update.scsi-devices pcmk.stonith.update.scsi-devices.add-remove pcmk.stonith.update.scsi-devices.mpath pcmk.tag pcmk.tag.resources pcs.auth.client pcs.auth.client.cluster pcs.auth.client.token pcs.auth.deauth-client pcs.auth.deauth-server pcs.auth.no-bidirectional pcs.auth.separated-name-and-address pcs.auth.server.token pcs.cfg-in-file.cib pcs.daemon-ssl-cert.set pcs.daemon-ssl-cert.sync-to-local-cluster pcs.disaster-recovery.essentials pcs.reports.severity.deprecation pcs.request-timeout resource-agents.describe resource-agents.list resource-agents.list.detailed resource-agents.ocf.version-1-0 resource-agents.ocf.version-1-1 resource-agents.self-validation sbd sbd.option-timeout-action sbd.shared-block-device status.corosync.membership status.pcmk.resources.hide-inactive status.pcmk.resources.id status.pcmk.resources.node status.pcmk.resources.orphaned status.pcmk.xml stonith-agents.describe stonith-agents.list stonith-agents.list.detailed stonith-agents.ocf.version-1-0 stonith-agents.ocf.version-1-1 stonith-agents.self-validation TASK [fedora.linux_system_roles.ha_cluster : Parse pcs capabilities] *********** task path: /WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/.collection/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/pcs-configure-pcs-pcsd.yml:101 Saturday 11 February 2023 19:50:20 +0000 (0:00:01.491) 0:02:59.370 ***** ok: [sut] => { "ansible_facts": { "__ha_cluster_pcs_capabilities": [ "booth", "booth.enable-authfile.set", "booth.enable-authfile.unset", "cluster.config.backup-local", "cluster.config.restore-cluster", "cluster.config.restore-local", "cluster.config.uuid", "cluster.create", "cluster.create.enable", "cluster.create.local", "cluster.create.no-keys-sync", "cluster.create.separated-name-and-address", "cluster.create.start", "cluster.create.start.wait", "cluster.create.transport.knet", "cluster.create.transport.udp-udpu", "cluster.create.transport.udp-udpu.no-rrp", "cluster.destroy", "cluster.destroy.all", "cluster.report", "cluster.verify", "corosync.authkey.update", "corosync.config.get", "corosync.config.get.struct", "corosync.config.reload", "corosync.config.sync-to-local-cluster", "corosync.config.update", "corosync.link.add", "corosync.link.remove", "corosync.link.remove.list", "corosync.link.update", "corosync.qdevice", "corosync.qdevice.model.net", "corosync.quorum", "corosync.quorum.device", "corosync.quorum.device.heuristics", "corosync.quorum.device.model.net", "corosync.quorum.set-expected-votes-runtime", "corosync.quorum.status", "corosync.quorum.unblock", "corosync.totem.block_unlisted_ips", "corosync.uidgid", "node.add", "node.add.enable", "node.add.separated-name-and-address", "node.add.start", "node.add.start.wait", "node.attributes", "node.attributes.set-list-for-node", "node.confirm-off", "node.fence", "node.guest", "node.kill", "node.maintenance", "node.maintenance.all", "node.maintenance.list", "node.maintenance.wait", "node.remote", "node.remote.onfail-demote", "node.remove", "node.remove-from-caches", "node.remove.list", "node.standby", "node.standby.all", "node.standby.list", "node.standby.wait", "node.start-stop-enable-disable", "node.start-stop-enable-disable.all", "node.start-stop-enable-disable.list", "node.start-stop-enable-disable.start-wait", "node.utilization", "node.utilization.set-list-for-node", "pcmk.acl.enable-disable", "pcmk.acl.group", "pcmk.acl.role", "pcmk.acl.role.create-with-permissions", "pcmk.acl.role.delete-with-users-groups", "pcmk.acl.user", "pcmk.alert", "pcmk.cib.checkpoints", "pcmk.cib.checkpoints.diff", "pcmk.cib.edit", "pcmk.cib.get", "pcmk.cib.get.scope", "pcmk.cib.roles.promoted-unpromoted", "pcmk.cib.set", "pcmk.constraint.colocation.set", "pcmk.constraint.colocation.set.options", "pcmk.constraint.colocation.simple", "pcmk.constraint.colocation.simple.options", "pcmk.constraint.hide-expired", "pcmk.constraint.location.simple", "pcmk.constraint.location.simple.options", "pcmk.constraint.location.simple.resource-regexp", "pcmk.constraint.location.simple.rule", "pcmk.constraint.location.simple.rule.node-attr-type-number", "pcmk.constraint.location.simple.rule.options", "pcmk.constraint.location.simple.rule.rule-add-remove", "pcmk.constraint.no-autocorrect", "pcmk.constraint.order.set", "pcmk.constraint.order.set.options", "pcmk.constraint.order.simple", "pcmk.constraint.order.simple.options", "pcmk.constraint.ticket.set", "pcmk.constraint.ticket.set.options", "pcmk.constraint.ticket.simple", "pcmk.constraint.ticket.simple.constraint-id", "pcmk.properties.cluster", "pcmk.properties.operation-defaults", "pcmk.properties.operation-defaults.multiple", "pcmk.properties.operation-defaults.rule", "pcmk.properties.operation-defaults.rule-rsc-op", "pcmk.properties.operation-defaults.rule.hide-expired", "pcmk.properties.operation-defaults.rule.node-attr-type-number", "pcmk.properties.resource-defaults", "pcmk.properties.resource-defaults.multiple", "pcmk.properties.resource-defaults.rule", "pcmk.properties.resource-defaults.rule-rsc-op", "pcmk.properties.resource-defaults.rule.hide-expired", "pcmk.properties.resource-defaults.rule.node-attr-type-number", "pcmk.resource.ban-move-clear", "pcmk.resource.ban-move-clear.clear-expired", "pcmk.resource.bundle", "pcmk.resource.bundle.container-docker", "pcmk.resource.bundle.container-docker.promoted-max", "pcmk.resource.bundle.container-podman", "pcmk.resource.bundle.container-podman.promoted-max", "pcmk.resource.bundle.container-rkt", "pcmk.resource.bundle.container-rkt.promoted-max", "pcmk.resource.bundle.reset", "pcmk.resource.bundle.wait", "pcmk.resource.cleanup", "pcmk.resource.cleanup.one-resource", "pcmk.resource.cleanup.strict", "pcmk.resource.clone", "pcmk.resource.clone.custom-id", "pcmk.resource.clone.meta-in-create", "pcmk.resource.clone.wait", "pcmk.resource.config.output-formats", "pcmk.resource.create", "pcmk.resource.create.clone.custom-id", "pcmk.resource.create.in-existing-bundle", "pcmk.resource.create.meta", "pcmk.resource.create.no-master", "pcmk.resource.create.operations", "pcmk.resource.create.operations.onfail-demote", "pcmk.resource.create.promotable", "pcmk.resource.create.promotable.custom-id", "pcmk.resource.create.wait", "pcmk.resource.debug", "pcmk.resource.delete", "pcmk.resource.disable.safe", "pcmk.resource.disable.safe.brief", "pcmk.resource.disable.safe.tag", "pcmk.resource.disable.simulate", "pcmk.resource.disable.simulate.brief", "pcmk.resource.disable.simulate.tag", "pcmk.resource.enable-disable", "pcmk.resource.enable-disable.list", "pcmk.resource.enable-disable.tag", "pcmk.resource.enable-disable.wait", "pcmk.resource.failcount", "pcmk.resource.group", "pcmk.resource.group.add-remove-list", "pcmk.resource.group.wait", "pcmk.resource.manage-unmanage", "pcmk.resource.manage-unmanage.list", "pcmk.resource.manage-unmanage.tag", "pcmk.resource.manage-unmanage.with-monitor", "pcmk.resource.move.autoclean", "pcmk.resource.move.autoclean.default", "pcmk.resource.promotable", "pcmk.resource.promotable.custom-id", "pcmk.resource.promotable.meta-in-create", "pcmk.resource.promotable.wait", "pcmk.resource.refresh", "pcmk.resource.refresh.one-resource", "pcmk.resource.refresh.strict", "pcmk.resource.relations", "pcmk.resource.relocate", "pcmk.resource.restart", "pcmk.resource.update", "pcmk.resource.update-meta", "pcmk.resource.update-meta.list", "pcmk.resource.update-meta.wait", "pcmk.resource.update-operations", "pcmk.resource.update-operations.onfail-demote", "pcmk.resource.update.meta", "pcmk.resource.update.operations", "pcmk.resource.update.operations.onfail-demote", "pcmk.resource.update.wait", "pcmk.resource.utilization", "pcmk.resource.utilization-set-list-for-resource", "pcmk.stonith.cleanup", "pcmk.stonith.cleanup.one-resource", "pcmk.stonith.cleanup.strict", "pcmk.stonith.create", "pcmk.stonith.create.in-group", "pcmk.stonith.create.meta", "pcmk.stonith.create.operations", "pcmk.stonith.create.operations.onfail-demote", "pcmk.stonith.create.wait", "pcmk.stonith.delete", "pcmk.stonith.enable-disable", "pcmk.stonith.enable-disable.list", "pcmk.stonith.enable-disable.wait", "pcmk.stonith.history.cleanup", "pcmk.stonith.history.show", "pcmk.stonith.history.update", "pcmk.stonith.levels", "pcmk.stonith.levels.add-remove-devices-list", "pcmk.stonith.levels.clear", "pcmk.stonith.levels.node-attr", "pcmk.stonith.levels.node-regexp", "pcmk.stonith.levels.verify", "pcmk.stonith.refresh", "pcmk.stonith.refresh.one-resource", "pcmk.stonith.refresh.strict", "pcmk.stonith.update", "pcmk.stonith.update.scsi-devices", "pcmk.stonith.update.scsi-devices.add-remove", "pcmk.stonith.update.scsi-devices.mpath", "pcmk.tag", "pcmk.tag.resources", "pcs.auth.client", "pcs.auth.client.cluster", "pcs.auth.client.token", "pcs.auth.deauth-client", "pcs.auth.deauth-server", "pcs.auth.no-bidirectional", "pcs.auth.separated-name-and-address", "pcs.auth.server.token", "pcs.cfg-in-file.cib", "pcs.daemon-ssl-cert.set", "pcs.daemon-ssl-cert.sync-to-local-cluster", "pcs.disaster-recovery.essentials", "pcs.reports.severity.deprecation", "pcs.request-timeout", "resource-agents.describe", "resource-agents.list", "resource-agents.list.detailed", "resource-agents.ocf.version-1-0", "resource-agents.ocf.version-1-1", "resource-agents.self-validation", "sbd", "sbd.option-timeout-action", "sbd.shared-block-device", "status.corosync.membership", "status.pcmk.resources.hide-inactive", "status.pcmk.resources.id", "status.pcmk.resources.node", "status.pcmk.resources.orphaned", "status.pcmk.xml", "stonith-agents.describe", "stonith-agents.list", "stonith-agents.list.detailed", "stonith-agents.ocf.version-1-0", "stonith-agents.ocf.version-1-1", "stonith-agents.self-validation" ] }, "changed": false } TASK [fedora.linux_system_roles.ha_cluster : Install cluster packages] ********* task path: /WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/.collection/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:34 Saturday 11 February 2023 19:50:20 +0000 (0:00:00.090) 0:02:59.460 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.ha_cluster : Configure firewall] *************** task path: /WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/.collection/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:39 Saturday 11 February 2023 19:50:20 +0000 (0:00:00.099) 0:02:59.560 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.ha_cluster : Configure selinux] **************** task path: /WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/.collection/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:42 Saturday 11 February 2023 19:50:20 +0000 (0:00:00.123) 0:02:59.684 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.ha_cluster : Distribute fence-virt authkey] **** task path: /WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/.collection/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:45 Saturday 11 February 2023 19:50:20 +0000 (0:00:00.216) 0:02:59.901 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.ha_cluster : Configure SBD] ******************** task path: /WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/.collection/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:50 Saturday 11 February 2023 19:50:21 +0000 (0:00:00.119) 0:03:00.020 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.ha_cluster : Configure corosync] *************** task path: /WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/.collection/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:53 Saturday 11 February 2023 19:50:21 +0000 (0:00:00.219) 0:03:00.240 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.ha_cluster : Pcs auth] ************************* task path: /WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/.collection/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:56 Saturday 11 February 2023 19:50:21 +0000 (0:00:00.199) 0:03:00.439 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.ha_cluster : Distribute cluster shared keys] *** task path: /WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/.collection/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:61 Saturday 11 February 2023 19:50:21 +0000 (0:00:00.130) 0:03:00.570 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.ha_cluster : Enable or disable cluster services on boot] *** task path: /WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/.collection/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:67 Saturday 11 February 2023 19:50:21 +0000 (0:00:00.070) 0:03:00.641 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.ha_cluster : Start the cluster and reload corosync.conf] *** task path: /WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/.collection/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:70 Saturday 11 February 2023 19:50:21 +0000 (0:00:00.126) 0:03:00.767 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.ha_cluster : Create and push CIB] ************** task path: /WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/.collection/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:73 Saturday 11 February 2023 19:50:21 +0000 (0:00:00.158) 0:03:00.926 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.ha_cluster : Remove cluster configuration] ***** task path: /WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/.collection/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:79 Saturday 11 February 2023 19:50:22 +0000 (0:00:00.125) 0:03:01.051 ***** included: /WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/.collection/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/cluster-destroy-pcs-0.10.yml for sut TASK [fedora.linux_system_roles.ha_cluster : Remove cluster configuration] ***** task path: /WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/.collection/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/cluster-destroy-pcs-0.10.yml:9 Saturday 11 February 2023 19:50:22 +0000 (0:00:00.111) 0:03:01.163 ***** ok: [sut] => (item=/etc/corosync/corosync.conf) => { "ansible_loop_var": "item", "changed": false, "cmd": [ "pcs", "cluster", "destroy" ], "delta": null, "end": null, "item": "/etc/corosync/corosync.conf", "rc": 0, "start": null } STDOUT: skipped, since /etc/corosync/corosync.conf does not exist MSG: Did not run command since '/etc/corosync/corosync.conf' does not exist ok: [sut] => (item=/var/lib/pacemaker/cib/cib.xml) => { "ansible_loop_var": "item", "changed": false, "cmd": [ "pcs", "cluster", "destroy" ], "delta": null, "end": null, "item": "/var/lib/pacemaker/cib/cib.xml", "rc": 0, "start": null } STDOUT: skipped, since /var/lib/pacemaker/cib/cib.xml does not exist MSG: Did not run command since '/var/lib/pacemaker/cib/cib.xml' does not exist TASK [fedora.linux_system_roles.ha_cluster : Remove fence-virt authkey] ******** task path: /WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/.collection/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:82 Saturday 11 February 2023 19:50:22 +0000 (0:00:00.653) 0:03:01.816 ***** ok: [sut] => { "changed": false, "path": "/etc/cluster/fence_xvm.key", "state": "absent" } TASK [fedora.linux_system_roles.ha_cluster : Configure qnetd] ****************** task path: /WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/.collection/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:87 Saturday 11 February 2023 19:50:23 +0000 (0:00:00.775) 0:03:02.592 ***** included: /WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/.collection/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/pcs-qnetd.yml for sut TASK [fedora.linux_system_roles.ha_cluster : Remove qnetd configuration] ******* task path: /WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/.collection/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/pcs-qnetd.yml:3 Saturday 11 February 2023 19:50:23 +0000 (0:00:00.137) 0:03:02.730 ***** changed: [sut] => { "changed": true, "cmd": [ "pcs", "--force", "--", "qdevice", "destroy", "net" ], "delta": "0:00:01.425243", "end": "2023-02-11 19:50:25.488228", "rc": 0, "start": "2023-02-11 19:50:24.062985" } STDERR: Stopping quorum device... quorum device stopped quorum device disabled Quorum device 'net' configuration files removed TASK [fedora.linux_system_roles.ha_cluster : Setup qnetd] ********************** task path: /WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/.collection/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/pcs-qnetd.yml:15 Saturday 11 February 2023 19:50:25 +0000 (0:00:01.813) 0:03:04.543 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.ha_cluster : Enable or disable qnetd service on boot] *** task path: /WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/.collection/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/pcs-qnetd.yml:25 Saturday 11 February 2023 19:50:25 +0000 (0:00:00.078) 0:03:04.622 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Stat corosync.conf] ****************************************************** task path: /WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/tests/tests_cluster_destroy.yml:28 Saturday 11 February 2023 19:50:25 +0000 (0:00:00.081) 0:03:04.703 ***** ok: [sut] => { "changed": false, "stat": { "exists": false } } TASK [Stat cib.xml] ************************************************************ task path: /WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/tests/tests_cluster_destroy.yml:33 Saturday 11 February 2023 19:50:25 +0000 (0:00:00.296) 0:03:04.999 ***** ok: [sut] => { "changed": false, "stat": { "exists": false } } TASK [Stat fence_xvm.key] ****************************************************** task path: /WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/tests/tests_cluster_destroy.yml:38 Saturday 11 February 2023 19:50:26 +0000 (0:00:00.374) 0:03:05.374 ***** ok: [sut] => { "changed": false, "stat": { "exists": false } } TASK [Check the files do not exist] ******************************************** task path: /WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/tests/tests_cluster_destroy.yml:43 Saturday 11 February 2023 19:50:26 +0000 (0:00:00.294) 0:03:05.669 ***** ok: [sut] => { "changed": false } MSG: All assertions passed PLAY RECAP ********************************************************************* sut : ok=29 changed=5 unreachable=0 failed=0 skipped=27 rescued=0 ignored=0 Saturday 11 February 2023 19:50:26 +0000 (0:00:00.132) 0:03:05.802 ***** =============================================================================== fedora.linux_system_roles.ha_cluster : Install role essential packages -- 33.43s /WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/.collection/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:14 fedora.linux_system_roles.ha_cluster : Set hacluster password ---------- 22.72s /WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/.collection/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:21 Gathering Facts -------------------------------------------------------- 20.55s /WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/tests/tests_cluster_destroy.yml:9 fedora.linux_system_roles.ha_cluster : Stop pcsd ----------------------- 20.10s /WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/.collection/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/pcs-configure-pcs-pcsd.yml:6 fedora.linux_system_roles.ha_cluster : Extract qdevice settings -------- 11.36s /WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/.collection/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/check-and-prepare-role-variables.yml:48 fedora.linux_system_roles.ha_cluster : Fail if passwords are not specified --- 8.15s /WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/.collection/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/check-and-prepare-role-variables.yml:5 Run HA Cluster role ----------------------------------------------------- 5.41s /WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/tests/tests_cluster_destroy.yml:23 fedora.linux_system_roles.ha_cluster : Figure out if ATB needs to be enabled for SBD --- 5.03s /WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/.collection/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/check-and-prepare-role-variables.yml:57 fedora.linux_system_roles.ha_cluster : Start pcsd with updated config files and configure it to start on boot --- 4.14s /WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/.collection/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/pcs-configure-pcs-pcsd.yml:88 fedora.linux_system_roles.ha_cluster : Set platform/version specific variables --- 3.99s /WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/.collection/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/set_vars.yml:7 fedora.linux_system_roles.ha_cluster : Collect cluster node names ------- 3.66s /WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/.collection/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/check-and-prepare-role-variables.yml:44 fedora.linux_system_roles.ha_cluster : Ensure ansible_facts used by role --- 3.59s /WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/.collection/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/set_vars.yml:2 fedora.linux_system_roles.ha_cluster : Configure pcs / pcsd ------------- 3.50s /WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/.collection/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:28 fedora.linux_system_roles.ha_cluster : Distribute pcs_settings.conf ----- 3.00s /WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/.collection/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/pcs-configure-pcs-pcsd.yml:79 fedora.linux_system_roles.ha_cluster : Set node name to 'localhost' for single-node clusters --- 2.61s /WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/.collection/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/test_setup.yml:9 fedora.linux_system_roles.ha_cluster : Fail if SBD needs ATB enabled and the user configured ATB to be disabled --- 2.54s /WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/.collection/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/check-and-prepare-role-variables.yml:67 fedora.linux_system_roles.ha_cluster : Discover cluster node names ------ 2.53s /WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/.collection/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/check-and-prepare-role-variables.yml:40 fedora.linux_system_roles.ha_cluster : Do not try to enable RHEL repositories --- 2.29s /WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/.collection/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/test_setup.yml:20 Set up test environment ------------------------------------------------- 2.23s /WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/tests/tests_cluster_destroy.yml:18 fedora.linux_system_roles.ha_cluster : Fail if ha_cluster_pcsd_public_key_src and ha_cluster_pcsd_private_key_src are set along with ha_cluster_pcsd_certificates --- 2.01s /WORKDIR/dist-git-ha_cluster-weekly-ci-ar7vq6vu/.collection/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/check-and-prepare-role-variables.yml:74 ---^---^---^---^---^--- # STDERR: ---v---v---v---v---v--- [DEPRECATION WARNING]: [defaults]callback_whitelist option, normalizing names to new standard, use callbacks_enabled instead. This feature will be removed from ansible-core in version 2.15. Deprecation warnings can be disabled by setting deprecation_warnings=False in ansible.cfg. [DEPRECATION WARNING]: Encryption using the Python crypt module is deprecated. The Python crypt module is deprecated and will be removed from Python 3.13. Install the passlib library for continued encryption functionality. This feature will be removed in version 2.17. Deprecation warnings can be disabled by setting deprecation_warnings=False in ansible.cfg. ---^---^---^---^---^---