# STDOUT: ---v---v---v---v---v--- ansible-playbook [core 2.16.0] config file = /etc/ansible/ansible.cfg configured module search path = ['/home/jenkins/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules'] ansible python module location = /opt/ansible-2.16/lib/python3.11/site-packages/ansible ansible collection location = /WORKDIR/git-fix-linger-handlingciq7eztj/.collection executable location = /opt/ansible-2.16/bin/ansible-playbook python version = 3.11.5 (main, Sep 7 2023, 00:00:00) [GCC 11.4.1 20230605 (Red Hat 11.4.1-2)] (/opt/ansible-2.16/bin/python) jinja version = 3.1.2 libyaml = True Using /etc/ansible/ansible.cfg as config file Skipping callback 'debug', as we already have a stdout callback. Skipping callback 'default', as we already have a stdout callback. Skipping callback 'minimal', as we already have a stdout callback. Skipping callback 'oneline', as we already have a stdout callback. PLAYBOOK: tests_quadlet_demo.yml *********************************************** 2 plays in /WORKDIR/git-fix-linger-handlingciq7eztj/tests/tests_quadlet_demo.yml PLAY [all] ********************************************************************* TASK [Include vault variables] ************************************************* task path: /WORKDIR/git-fix-linger-handlingciq7eztj/tests/tests_quadlet_demo.yml:4 Sunday 04 February 2024 19:38:28 +0000 (0:00:00.014) 0:00:00.014 ******* ok: [sut] => { "ansible_facts": { "mysql_container_root_password": { "__ansible_vault": "$ANSIBLE_VAULT;1.1;AES256\n61333932373230333539663035366431326163363166363036323963623131363530326231303634\n6635326161643165363366323062333334363730376631660a393566366139353861656364656661\n38653463363837336639363032646433666361646535366137303464623261313663643336306465\n6264663730656337310a343962353137386238383064646533366433333437303566656433386233\n34343235326665646661623131643335313236313131353661386338343366316261643634653633\n3832313034366536616531323963333234326461353130303532\n" } }, "ansible_included_var_files": [ "/WORKDIR/git-fix-linger-handlingciq7eztj/tests/vars/vault-variables.yml" ], "changed": false } PLAY [Deploy the quadlet demo app] ********************************************* TASK [Gathering Facts] ********************************************************* task path: /WORKDIR/git-fix-linger-handlingciq7eztj/tests/tests_quadlet_demo.yml:9 Sunday 04 February 2024 19:38:28 +0000 (0:00:00.014) 0:00:00.029 ******* ok: [sut] TASK [Generate certificates] *************************************************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/tests/tests_quadlet_demo.yml:39 Sunday 04 February 2024 19:38:29 +0000 (0:00:00.940) 0:00:00.969 ******* TASK [fedora.linux_system_roles.certificate : Set version specific variables] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:2 Sunday 04 February 2024 19:38:29 +0000 (0:00:00.027) 0:00:00.997 ******* included: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/set_vars.yml for sut TASK [fedora.linux_system_roles.certificate : Ensure ansible_facts used by role] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/set_vars.yml:2 Sunday 04 February 2024 19:38:29 +0000 (0:00:00.014) 0:00:01.012 ******* skipping: [sut] => { "changed": false, "false_condition": "__certificate_required_facts | difference(ansible_facts.keys() | list) | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.certificate : Check if system is ostree] ******* task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/set_vars.yml:10 Sunday 04 February 2024 19:38:29 +0000 (0:00:00.020) 0:00:01.032 ******* ok: [sut] => { "changed": false, "stat": { "exists": false } } TASK [fedora.linux_system_roles.certificate : Set flag to indicate system is ostree] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/set_vars.yml:15 Sunday 04 February 2024 19:38:29 +0000 (0:00:00.307) 0:00:01.339 ******* ok: [sut] => { "ansible_facts": { "__certificate_is_ostree": false }, "changed": false } TASK [fedora.linux_system_roles.certificate : Set platform/version specific variables] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/set_vars.yml:19 Sunday 04 February 2024 19:38:29 +0000 (0:00:00.017) 0:00:01.357 ******* skipping: [sut] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "false_condition": "__vars_file is file", "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [sut] => (item=Fedora.yml) => { "ansible_facts": { "__certificate_certmonger_packages": [ "certmonger", "python3-packaging" ] }, "ansible_included_var_files": [ "/WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/vars/Fedora.yml" ], "ansible_loop_var": "item", "changed": false, "item": "Fedora.yml" } skipping: [sut] => (item=Fedora_39.yml) => { "ansible_loop_var": "item", "changed": false, "false_condition": "__vars_file is file", "item": "Fedora_39.yml", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=Fedora_39.yml) => { "ansible_loop_var": "item", "changed": false, "false_condition": "__vars_file is file", "item": "Fedora_39.yml", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.certificate : Ensure certificate role dependencies are installed] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:5 Sunday 04 February 2024 19:38:29 +0000 (0:00:00.028) 0:00:01.386 ******* changed: [sut] => { "changed": true, "rc": 0, "results": [ "Installed: python3-ply-3.11-20.fc39.noarch", "Installed: python3-pycparser-2.20-11.fc39.noarch", "Installed: python3-cffi-1.15.1-6.fc39.x86_64", "Installed: python3-pyasn1-0.5.1-1.fc39.noarch", "Installed: python3-cryptography-41.0.3-1.fc39.x86_64" ] } TASK [fedora.linux_system_roles.certificate : Ensure provider packages are installed] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:23 Sunday 04 February 2024 19:38:35 +0000 (0:00:06.156) 0:00:07.542 ******* changed: [sut] => (item=certmonger) => { "__certificate_provider": "certmonger", "ansible_loop_var": "__certificate_provider", "changed": true, "rc": 0, "results": [ "Installed: certmonger-0.79.18-2.fc39.x86_64", "Installed: dbus-tools-1:1.14.10-1.fc39.x86_64", "Installed: python3-packaging-23.1-4.fc39.noarch" ] } TASK [fedora.linux_system_roles.certificate : Ensure pre-scripts hooks directory exists] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:35 Sunday 04 February 2024 19:38:40 +0000 (0:00:04.820) 0:00:12.362 ******* changed: [sut] => (item=certmonger) => { "__certificate_provider": "certmonger", "ansible_loop_var": "__certificate_provider", "changed": true, "gid": 0, "group": "root", "mode": "0700", "owner": "root", "path": "/etc/certmonger//pre-scripts", "secontext": "unconfined_u:object_r:etc_t:s0", "size": 4096, "state": "directory", "uid": 0 } TASK [fedora.linux_system_roles.certificate : Ensure post-scripts hooks directory exists] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:61 Sunday 04 February 2024 19:38:40 +0000 (0:00:00.354) 0:00:12.717 ******* changed: [sut] => (item=certmonger) => { "__certificate_provider": "certmonger", "ansible_loop_var": "__certificate_provider", "changed": true, "gid": 0, "group": "root", "mode": "0700", "owner": "root", "path": "/etc/certmonger//post-scripts", "secontext": "unconfined_u:object_r:etc_t:s0", "size": 4096, "state": "directory", "uid": 0 } TASK [fedora.linux_system_roles.certificate : Ensure provider service is running] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:90 Sunday 04 February 2024 19:38:41 +0000 (0:00:00.267) 0:00:12.984 ******* changed: [sut] => (item=certmonger) => { "__certificate_provider": "certmonger", "ansible_loop_var": "__certificate_provider", "changed": true, "enabled": true, "name": "certmonger", "state": "started", "status": { "AccessSELinuxContext": "system_u:object_r:certmonger_unit_file_t:s0", "ActiveEnterTimestampMonotonic": "0", "ActiveExitTimestampMonotonic": "0", "ActiveState": "inactive", "After": "dbus.socket systemd-journald.socket sysinit.target syslog.target dbus-broker.service basic.target system.slice network.target", "AllowIsolate": "no", "AssertResult": "no", "AssertTimestampMonotonic": "0", "Before": "shutdown.target", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "BusName": "org.fedorahosted.certmonger", "CPUAccounting": "yes", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "[not set]", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "no", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore", "CleanResult": "success", "CollectMode": "inactive", "ConditionResult": "no", "ConditionTimestampMonotonic": "0", "ConfigurationDirectoryMode": "0755", "Conflicts": "shutdown.target", "ControlGroupId": "0", "ControlPID": "0", "CoredumpFilter": "0x33", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "DefaultStartupMemoryLow": "0", "Delegate": "no", "Description": "Certificate monitoring and PKI enrollment", "DevicePolicy": "auto", "DropInPaths": "/usr/lib/systemd/system/service.d/10-timeout-abort.conf", "DynamicUser": "no", "EnvironmentFiles": "/etc/sysconfig/certmonger (ignore_errors=yes)", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "0", "ExecMainStartTimestampMonotonic": "0", "ExecMainStatus": "0", "ExecStart": "{ path=/usr/sbin/certmonger ; argv[]=/usr/sbin/certmonger -S -p /run/certmonger.pid -n $OPTS ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStartEx": "{ path=/usr/sbin/certmonger ; argv[]=/usr/sbin/certmonger -S -p /run/certmonger.pid -n $OPTS ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExitType": "main", "ExtensionImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FileDescriptorStorePreserve": "restart", "FinalKillSignal": "9", "FragmentPath": "/usr/lib/systemd/system/certmonger.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOReadBytes": "18446744073709551615", "IOReadOperations": "18446744073709551615", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "IOWeight": "[not set]", "IOWriteBytes": "18446744073709551615", "IOWriteOperations": "18446744073709551615", "IPAccounting": "no", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "Id": "certmonger.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestampMonotonic": "0", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "control-group", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "infinity", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "8388608", "LimitMEMLOCKSoft": "8388608", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "524288", "LimitNOFILESoft": "1024", "LimitNPROC": "14778", "LimitNPROCSoft": "14778", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "14778", "LimitSIGPENDINGSoft": "14778", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "0", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "ManagedOOMSwap": "auto", "MemoryAccounting": "yes", "MemoryAvailable": "infinity", "MemoryCurrent": "[not set]", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryKSM": "no", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemoryPressureThresholdUSec": "200ms", "MemoryPressureWatch": "auto", "MemorySwapMax": "infinity", "MemoryZSwapMax": "infinity", "MountAPIVFS": "no", "MountImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAPolicy": "n/a", "Names": "certmonger.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMPolicy": "stop", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "OnSuccessJobMode": "fail", "PIDFile": "/run/certmonger.pid", "PartOf": "dbus-broker.service", "Perpetual": "no", "PrivateDevices": "no", "PrivateIPC": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProcSubset": "all", "ProtectClock": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectHostname": "no", "ProtectKernelLogs": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectProc": "default", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "ReloadResult": "success", "ReloadSignal": "1", "RemainAfterExit": "no", "RemoveIPC": "no", "Requires": "sysinit.target dbus.socket system.slice", "Restart": "no", "RestartKillSignal": "15", "RestartMaxDelayUSec": "infinity", "RestartMode": "normal", "RestartSteps": "0", "RestartUSec": "100ms", "RestartUSecNext": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RootEphemeral": "no", "RootImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "Slice": "system.slice", "StandardError": "inherit", "StandardInput": "null", "StandardOutput": "journal", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StartupMemoryHigh": "18446744073709551615", "StartupMemoryLow": "0", "StartupMemoryMax": "18446744073709551615", "StartupMemorySwapMax": "18446744073709551615", "StartupMemoryZSwapMax": "18446744073709551615", "StateChangeTimestampMonotonic": "0", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "dead", "SuccessAction": "none", "SyslogFacility": "3", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "2147483646", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "[not set]", "TasksMax": "4433", "TimeoutAbortUSec": "45s", "TimeoutCleanUSec": "infinity", "TimeoutStartFailureMode": "terminate", "TimeoutStartUSec": "45s", "TimeoutStopFailureMode": "abort", "TimeoutStopUSec": "45s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "dbus", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "disabled", "UnitFileState": "disabled", "UtmpMode": "init", "WatchdogSignal": "6", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "infinity" } } TASK [fedora.linux_system_roles.certificate : Ensure certificate requests] ***** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:101 Sunday 04 February 2024 19:38:42 +0000 (0:00:01.120) 0:00:14.105 ******* changed: [sut] => (item={'name': 'quadlet_demo', 'dns': ['localhost'], 'ca': 'self-sign'}) => { "ansible_loop_var": "item", "changed": true, "item": { "ca": "self-sign", "dns": [ "localhost" ], "name": "quadlet_demo" } } MSG: Certificate requested (new). TASK [fedora.linux_system_roles.certificate : Slurp the contents of the files] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:152 Sunday 04 February 2024 19:38:43 +0000 (0:00:00.838) 0:00:14.944 ******* ok: [sut] => (item=['cert', {'name': 'quadlet_demo', 'dns': ['localhost'], 'ca': 'self-sign'}]) => { "ansible_loop_var": "item", "changed": false, "content": "LS0tLS1CRUdJTiBDRVJUSUZJQ0FURS0tLS0tCk1JSURnekNDQW11Z0F3SUJBZ0lSQUpyZjhoVjVDMGhGcG9BQVZYMXZkdEl3RFFZSktvWklodmNOQVFFTEJRQXcKVURFZ01CNEdBMVVFQXd3WFRHOWpZV3dnVTJsbmJtbHVaeUJCZFhSb2IzSnBkSGt4TERBcUJnTlZCQU1NSXpsaApaR1ptTWpFMUxUYzVNR0kwT0RRMUxXRTJPREF3TURVMUxUZGtObVkzTm1ReE1CNFhEVEkwTURJd05ERTVNemcwCk1sb1hEVEkxTURJd05ERTVNemcwTWxvd0ZERVNNQkFHQTFVRUF4TUpiRzlqWVd4b2IzTjBNSUlCSWpBTkJna3EKaGtpRzl3MEJBUUVGQUFPQ0FROEFNSUlCQ2dLQ0FRRUF2UEhFQjBNMW1lNHowT3ZFSHl5SVFWeERWaUI5dVhLWQpHV0llUkJmaTNweTE2STh0WlRBZUs3eXZHVXpXYXVkd0JKc3NPSnVqUE5UaDRmTklpUm82QjZmOXRDaExSSll4Ckpzdm5lUHFuZEdCamNmdE1kUGtHSHVDZks0enZrd1I5S2d0TXN3b094amc0d2hFaW0yZStKRFU2dlZiNHZLVWUKeVRnWURqVk1aMzlvSFlySmlHUG5oZWgvT0JyUGpXNWYvM0ppb1MzcVpOdVdLazBkUEQrOHVTOHJRMkpldVFxYwpvZUJHSExua0ZTeWxrN0s2eGFMbnIyS2U5R2RrVFNTdUNFWGZ2YktEc0JST200eDNoMmV2RVZxY3ljdEg0d3FNCnhpb2IyVUZZQXk3aHZ3Yy9ka0t5ZXExWWVOQmxVRXg3MlJHbUNwSkZ3Qi90VGgzRktyMFZTUUlEQVFBQm80R1QKTUlHUU1Bc0dBMVVkRHdRRUF3SUZvREFVQmdOVkhSRUVEVEFMZ2dsc2IyTmhiR2h2YzNRd0hRWURWUjBsQkJZdwpGQVlJS3dZQkJRVUhBd0VHQ0NzR0FRVUZCd01DTUF3R0ExVWRFd0VCL3dRQ01BQXdIUVlEVlIwT0JCWUVGTlZWCnJLdWVWS0JGRUdQSWZESDVVdjA0RG1vVk1COEdBMVVkSXdRWU1CYUFGQlpPcWkzWWxRNExhZ1ZlU0NyaVFLY3cKOWV2UE1BMEdDU3FHU0liM0RRRUJDd1VBQTRJQkFRQmw5VzQ4UlRhSXdsaS9sS08xbVl0NlJnTEU4ZllFR2hTWQo5YU02RVZrRTlSem1uamZ4UmkwNkZTMXZMdUlPUEQycU1kZFZVQjVPdGp4bTBKQ042RTRwbS92M3VRM3NWbStOCnQ0ekxRNFZzV0RHcGpFL2REd1pTZjUwVzJOYUpQcUs4ZVFuVFR1aGc3N1c2Ykl3NnJRYWZFWVRzdWdTT2x5VGMKM1U2WjYwWSt3ZjBjdTVhd3pWME5HSHpWUVluVjNmQU1jcm1iNXlVUWFhYTFKb2FkUzRHbUMzaU93ZGRJRTRPeApsVXdta25ta0p3L2d2UWZYYlVXWEhZSFZZcFZBdjR4dTNvZHVJWGs1NnhvS1praU5ZbG9zRUVDNzd5RjhQY3lVClgzWlhSeWY5WFZkNmNLVll2VGp3WlRPMnE0SkNidWpQelc0MFNNMlZFWjlPWWEzY1p2TSsKLS0tLS1FTkQgQ0VSVElGSUNBVEUtLS0tLQo=", "encoding": "base64", "item": [ "cert", { "ca": "self-sign", "dns": [ "localhost" ], "name": "quadlet_demo" } ], "source": "/etc/pki/tls/certs/quadlet_demo.crt" } ok: [sut] => (item=['key', {'name': 'quadlet_demo', 'dns': ['localhost'], 'ca': 'self-sign'}]) => { "ansible_loop_var": "item", "changed": false, "content": "LS0tLS1CRUdJTiBQUklWQVRFIEtFWS0tLS0tCk1JSUV2UUlCQURBTkJna3Foa2lHOXcwQkFRRUZBQVNDQktjd2dnU2pBZ0VBQW9JQkFRQzg4Y1FIUXpXWjdqUFEKNjhRZkxJaEJYRU5XSUgyNWNwZ1pZaDVFRitMZW5MWG9qeTFsTUI0cnZLOFpUTlpxNTNBRW15dzRtNk04MU9IaAo4MGlKR2pvSHAvMjBLRXRFbGpFbXkrZDQrcWQwWUdOeCsweDArUVllNEo4cmpPK1RCSDBxQzB5ekNnN0dPRGpDCkVTS2JaNzRrTlRxOVZ2aThwUjdKT0JnT05VeG5mMmdkaXNtSVkrZUY2SDg0R3MrTmJsLy9jbUtoTGVwazI1WXEKVFIwOFA3eTVMeXREWWw2NUNweWg0RVljdWVRVkxLV1RzcnJGb3VldllwNzBaMlJOSks0SVJkKzlzb093RkU2YgpqSGVIWjY4Uldwekp5MGZqQ296R0todlpRVmdETHVHL0J6OTJRcko2clZoNDBHVlFUSHZaRWFZS2trWEFIKzFPCkhjVXF2UlZKQWdNQkFBRUNnZ0VBSU52MDZRUnR2bnI0SlBYVHNrNDIxaE1iSE0vaXZZMHAyWVJBV2p4RE5zN1AKTGdZMGNWd1lheWZSMWRjUWxSL3gwY05Bd2dtRExwR2hSUC9mMWdRckltaTdUUWtTNmZjcXppSDQvQWV2ZHFSSwp0eWhxaUE5RTJycmF3NktyaVlobTBoaUwrcmtkajZoNGFudTdab1o0d1doR2NtWTlOT1dNZ0pjU1MrcEdBdHVQCllxdDArOHVZZUM2MGVsSDM3aURrRndNaWRtZW9ETDhuSndDdTJYM0QrK3VSbE1pSHNkRlNYMnIvZzIzTmt4R0QKOHNwRHYwbkNKMlNydzdubXQrakR0KzV2enVTK2cxQVRzRWhUMzNRaDJMdEVTem5pZ1I4K0xtSmRNcFlPbDZkLwphNy8ySjdGTlpQWEZRNXdTRkhQcDkydGFRMno3Y3lTbGtwd2JqemRoV1FLQmdRRG5LblhYc25FRUVlRSs3dFBwClRLbS9yOEJkWDF1N0VuM294RDBmenZ5RzVYTVFJOWpQdnpuTGNCaGVsUzA1TUo1QU0vYmlwTmFxMCt5eE5PQSsKOWJGR1VrOVJGVHYxVDJ5TGdjaG5XSnNHZ2JMekRjK1E3SVBsVGZ1M2tHWGhkZk5hVDN3QlEyemx5QjRvcjdaYgpHWlhIcXV0MkNYdFpVTmFsQUxYVVRtdXdMUUtCZ1FEUlBpQzErS2tiOHFNdTk5dFd4K0FhT0k0eUdZalJHWWp0CnF1TUpsZWM3bFl6V0s4bVAwbEZLU2JSdXBVckVSdFVsNFlKUE5Fczc0SFpSZDJQMk85aTJmL08xNXI4WXBZZTQKcTY4UEFpNHdXa011MFBnWDRlRHp1Ni9iQVBGRXpwSGJRVDQ1NGZGTkhOMGszR1pFdkVvdlhFbkxvWGVmQ3RPMApnNWJidE15UERRS0JnQnhZVklMU1dwRkpML2xyaDRoLzRnQ2pLRTlLdGhER1A3SWUvZSs4amJKc3lDTWVnaU5YCksxc0tyN0pKVUhaR3crRFRmbDVrSC81cVpnZUVtRXFnRGRDMW53d3VNbC91RVQwNStMelRYSTc3c0ZtbFErSmQKWGY3c0JRV0pLelR1Slp6WlNqZE5jVGpyeE9id1JNZis1eUxJam42Mnd1MDh3RlVUTlRpSE5TVXhBb0dBSGJmUgpOWWNubll5TW1MTWpOK1FUZENYc0xEbHNNU2FMU0JZcFFNZEZuRFRWVitYSUx1RVVJcFR0UjNtNmNMNE9yOTZqCmYxUlNDeU1LS3UvbUJxYmhWS28vbHVvNjRlbEFxWDNZMHl6Z0dvWDNYTUxaT2FzZkkvN1lMeWFuSGJxcEJKYnoKbTR6M2xKZ3c0SGRkSHc0UEMwTDBTRE9hcGNBSDZXOTNFYVZRR0ZFQ2dZRUEzWDZubzJXcG1sTmZDaUxDeXlXcQphY2ZZRHFBb1VvOG82TmNpVU5sYlh3RXpwMllPNUxxRUJIL21rRkc2MWlJVmtTaCt2VjFjVGgwZ2cxS1E4V2dLClRyOFVqNXptRFFWQmtaN29hVSthSSttY2V6TVpVMzlSYTAwc3IvRGlmTmo4Mm82bC95ZDhvSVVsSXQxMFM3bFQKcFliQ0VOaGtjL0FqQXM2NUxTekp2QmM9Ci0tLS0tRU5EIFBSSVZBVEUgS0VZLS0tLS0K", "encoding": "base64", "item": [ "key", { "ca": "self-sign", "dns": [ "localhost" ], "name": "quadlet_demo" } ], "source": "/etc/pki/tls/private/quadlet_demo.key" } ok: [sut] => (item=['ca', {'name': 'quadlet_demo', 'dns': ['localhost'], 'ca': 'self-sign'}]) => { "ansible_loop_var": "item", "changed": false, "content": "LS0tLS1CRUdJTiBDRVJUSUZJQ0FURS0tLS0tCk1JSURnekNDQW11Z0F3SUJBZ0lSQUpyZjhoVjVDMGhGcG9BQVZYMXZkdEl3RFFZSktvWklodmNOQVFFTEJRQXcKVURFZ01CNEdBMVVFQXd3WFRHOWpZV3dnVTJsbmJtbHVaeUJCZFhSb2IzSnBkSGt4TERBcUJnTlZCQU1NSXpsaApaR1ptTWpFMUxUYzVNR0kwT0RRMUxXRTJPREF3TURVMUxUZGtObVkzTm1ReE1CNFhEVEkwTURJd05ERTVNemcwCk1sb1hEVEkxTURJd05ERTVNemcwTWxvd0ZERVNNQkFHQTFVRUF4TUpiRzlqWVd4b2IzTjBNSUlCSWpBTkJna3EKaGtpRzl3MEJBUUVGQUFPQ0FROEFNSUlCQ2dLQ0FRRUF2UEhFQjBNMW1lNHowT3ZFSHl5SVFWeERWaUI5dVhLWQpHV0llUkJmaTNweTE2STh0WlRBZUs3eXZHVXpXYXVkd0JKc3NPSnVqUE5UaDRmTklpUm82QjZmOXRDaExSSll4Ckpzdm5lUHFuZEdCamNmdE1kUGtHSHVDZks0enZrd1I5S2d0TXN3b094amc0d2hFaW0yZStKRFU2dlZiNHZLVWUKeVRnWURqVk1aMzlvSFlySmlHUG5oZWgvT0JyUGpXNWYvM0ppb1MzcVpOdVdLazBkUEQrOHVTOHJRMkpldVFxYwpvZUJHSExua0ZTeWxrN0s2eGFMbnIyS2U5R2RrVFNTdUNFWGZ2YktEc0JST200eDNoMmV2RVZxY3ljdEg0d3FNCnhpb2IyVUZZQXk3aHZ3Yy9ka0t5ZXExWWVOQmxVRXg3MlJHbUNwSkZ3Qi90VGgzRktyMFZTUUlEQVFBQm80R1QKTUlHUU1Bc0dBMVVkRHdRRUF3SUZvREFVQmdOVkhSRUVEVEFMZ2dsc2IyTmhiR2h2YzNRd0hRWURWUjBsQkJZdwpGQVlJS3dZQkJRVUhBd0VHQ0NzR0FRVUZCd01DTUF3R0ExVWRFd0VCL3dRQ01BQXdIUVlEVlIwT0JCWUVGTlZWCnJLdWVWS0JGRUdQSWZESDVVdjA0RG1vVk1COEdBMVVkSXdRWU1CYUFGQlpPcWkzWWxRNExhZ1ZlU0NyaVFLY3cKOWV2UE1BMEdDU3FHU0liM0RRRUJDd1VBQTRJQkFRQmw5VzQ4UlRhSXdsaS9sS08xbVl0NlJnTEU4ZllFR2hTWQo5YU02RVZrRTlSem1uamZ4UmkwNkZTMXZMdUlPUEQycU1kZFZVQjVPdGp4bTBKQ042RTRwbS92M3VRM3NWbStOCnQ0ekxRNFZzV0RHcGpFL2REd1pTZjUwVzJOYUpQcUs4ZVFuVFR1aGc3N1c2Ykl3NnJRYWZFWVRzdWdTT2x5VGMKM1U2WjYwWSt3ZjBjdTVhd3pWME5HSHpWUVluVjNmQU1jcm1iNXlVUWFhYTFKb2FkUzRHbUMzaU93ZGRJRTRPeApsVXdta25ta0p3L2d2UWZYYlVXWEhZSFZZcFZBdjR4dTNvZHVJWGs1NnhvS1praU5ZbG9zRUVDNzd5RjhQY3lVClgzWlhSeWY5WFZkNmNLVll2VGp3WlRPMnE0SkNidWpQelc0MFNNMlZFWjlPWWEzY1p2TSsKLS0tLS1FTkQgQ0VSVElGSUNBVEUtLS0tLQo=", "encoding": "base64", "item": [ "ca", { "ca": "self-sign", "dns": [ "localhost" ], "name": "quadlet_demo" } ], "source": "/etc/pki/tls/certs/quadlet_demo.crt" } TASK [fedora.linux_system_roles.certificate : Create return data] ************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:160 Sunday 04 February 2024 19:38:43 +0000 (0:00:00.762) 0:00:15.706 ******* ok: [sut] => { "ansible_facts": { "certificate_test_certs": { "quadlet_demo": { "ca": "/etc/pki/tls/certs/quadlet_demo.crt", "ca_content": "-----BEGIN CERTIFICATE-----\nMIIDgzCCAmugAwIBAgIRAJrf8hV5C0hFpoAAVX1vdtIwDQYJKoZIhvcNAQELBQAw\nUDEgMB4GA1UEAwwXTG9jYWwgU2lnbmluZyBBdXRob3JpdHkxLDAqBgNVBAMMIzlh\nZGZmMjE1LTc5MGI0ODQ1LWE2ODAwMDU1LTdkNmY3NmQxMB4XDTI0MDIwNDE5Mzg0\nMloXDTI1MDIwNDE5Mzg0MlowFDESMBAGA1UEAxMJbG9jYWxob3N0MIIBIjANBgkq\nhkiG9w0BAQEFAAOCAQ8AMIIBCgKCAQEAvPHEB0M1me4z0OvEHyyIQVxDViB9uXKY\nGWIeRBfi3py16I8tZTAeK7yvGUzWaudwBJssOJujPNTh4fNIiRo6B6f9tChLRJYx\nJsvnePqndGBjcftMdPkGHuCfK4zvkwR9KgtMswoOxjg4whEim2e+JDU6vVb4vKUe\nyTgYDjVMZ39oHYrJiGPnheh/OBrPjW5f/3JioS3qZNuWKk0dPD+8uS8rQ2JeuQqc\noeBGHLnkFSylk7K6xaLnr2Ke9GdkTSSuCEXfvbKDsBROm4x3h2evEVqcyctH4wqM\nxiob2UFYAy7hvwc/dkKyeq1YeNBlUEx72RGmCpJFwB/tTh3FKr0VSQIDAQABo4GT\nMIGQMAsGA1UdDwQEAwIFoDAUBgNVHREEDTALgglsb2NhbGhvc3QwHQYDVR0lBBYw\nFAYIKwYBBQUHAwEGCCsGAQUFBwMCMAwGA1UdEwEB/wQCMAAwHQYDVR0OBBYEFNVV\nrKueVKBFEGPIfDH5Uv04DmoVMB8GA1UdIwQYMBaAFBZOqi3YlQ4LagVeSCriQKcw\n9evPMA0GCSqGSIb3DQEBCwUAA4IBAQBl9W48RTaIwli/lKO1mYt6RgLE8fYEGhSY\n9aM6EVkE9RzmnjfxRi06FS1vLuIOPD2qMddVUB5Otjxm0JCN6E4pm/v3uQ3sVm+N\nt4zLQ4VsWDGpjE/dDwZSf50W2NaJPqK8eQnTTuhg77W6bIw6rQafEYTsugSOlyTc\n3U6Z60Y+wf0cu5awzV0NGHzVQYnV3fAMcrmb5yUQaaa1JoadS4GmC3iOwddIE4Ox\nlUwmknmkJw/gvQfXbUWXHYHVYpVAv4xu3oduIXk56xoKZkiNYlosEEC77yF8PcyU\nX3ZXRyf9XVd6cKVYvTjwZTO2q4JCbujPzW40SM2VEZ9OYa3cZvM+\n-----END CERTIFICATE-----\n", "cert": "/etc/pki/tls/certs/quadlet_demo.crt", "cert_content": "-----BEGIN CERTIFICATE-----\nMIIDgzCCAmugAwIBAgIRAJrf8hV5C0hFpoAAVX1vdtIwDQYJKoZIhvcNAQELBQAw\nUDEgMB4GA1UEAwwXTG9jYWwgU2lnbmluZyBBdXRob3JpdHkxLDAqBgNVBAMMIzlh\nZGZmMjE1LTc5MGI0ODQ1LWE2ODAwMDU1LTdkNmY3NmQxMB4XDTI0MDIwNDE5Mzg0\nMloXDTI1MDIwNDE5Mzg0MlowFDESMBAGA1UEAxMJbG9jYWxob3N0MIIBIjANBgkq\nhkiG9w0BAQEFAAOCAQ8AMIIBCgKCAQEAvPHEB0M1me4z0OvEHyyIQVxDViB9uXKY\nGWIeRBfi3py16I8tZTAeK7yvGUzWaudwBJssOJujPNTh4fNIiRo6B6f9tChLRJYx\nJsvnePqndGBjcftMdPkGHuCfK4zvkwR9KgtMswoOxjg4whEim2e+JDU6vVb4vKUe\nyTgYDjVMZ39oHYrJiGPnheh/OBrPjW5f/3JioS3qZNuWKk0dPD+8uS8rQ2JeuQqc\noeBGHLnkFSylk7K6xaLnr2Ke9GdkTSSuCEXfvbKDsBROm4x3h2evEVqcyctH4wqM\nxiob2UFYAy7hvwc/dkKyeq1YeNBlUEx72RGmCpJFwB/tTh3FKr0VSQIDAQABo4GT\nMIGQMAsGA1UdDwQEAwIFoDAUBgNVHREEDTALgglsb2NhbGhvc3QwHQYDVR0lBBYw\nFAYIKwYBBQUHAwEGCCsGAQUFBwMCMAwGA1UdEwEB/wQCMAAwHQYDVR0OBBYEFNVV\nrKueVKBFEGPIfDH5Uv04DmoVMB8GA1UdIwQYMBaAFBZOqi3YlQ4LagVeSCriQKcw\n9evPMA0GCSqGSIb3DQEBCwUAA4IBAQBl9W48RTaIwli/lKO1mYt6RgLE8fYEGhSY\n9aM6EVkE9RzmnjfxRi06FS1vLuIOPD2qMddVUB5Otjxm0JCN6E4pm/v3uQ3sVm+N\nt4zLQ4VsWDGpjE/dDwZSf50W2NaJPqK8eQnTTuhg77W6bIw6rQafEYTsugSOlyTc\n3U6Z60Y+wf0cu5awzV0NGHzVQYnV3fAMcrmb5yUQaaa1JoadS4GmC3iOwddIE4Ox\nlUwmknmkJw/gvQfXbUWXHYHVYpVAv4xu3oduIXk56xoKZkiNYlosEEC77yF8PcyU\nX3ZXRyf9XVd6cKVYvTjwZTO2q4JCbujPzW40SM2VEZ9OYa3cZvM+\n-----END CERTIFICATE-----\n", "key": "/etc/pki/tls/private/quadlet_demo.key", "key_content": "-----BEGIN PRIVATE KEY-----\nMIIEvQIBADANBgkqhkiG9w0BAQEFAASCBKcwggSjAgEAAoIBAQC88cQHQzWZ7jPQ\n68QfLIhBXENWIH25cpgZYh5EF+LenLXojy1lMB4rvK8ZTNZq53AEmyw4m6M81OHh\n80iJGjoHp/20KEtEljEmy+d4+qd0YGNx+0x0+QYe4J8rjO+TBH0qC0yzCg7GODjC\nESKbZ74kNTq9Vvi8pR7JOBgONUxnf2gdismIY+eF6H84Gs+Nbl//cmKhLepk25Yq\nTR08P7y5LytDYl65Cpyh4EYcueQVLKWTsrrFouevYp70Z2RNJK4IRd+9soOwFE6b\njHeHZ68RWpzJy0fjCozGKhvZQVgDLuG/Bz92QrJ6rVh40GVQTHvZEaYKkkXAH+1O\nHcUqvRVJAgMBAAECggEAINv06QRtvnr4JPXTsk421hMbHM/ivY0p2YRAWjxDNs7P\nLgY0cVwYayfR1dcQlR/x0cNAwgmDLpGhRP/f1gQrImi7TQkS6fcqziH4/AevdqRK\ntyhqiA9E2rraw6KriYhm0hiL+rkdj6h4anu7ZoZ4wWhGcmY9NOWMgJcSS+pGAtuP\nYqt0+8uYeC60elH37iDkFwMidmeoDL8nJwCu2X3D++uRlMiHsdFSX2r/g23NkxGD\n8spDv0nCJ2Srw7nmt+jDt+5vzuS+g1ATsEhT33Qh2LtESznigR8+LmJdMpYOl6d/\na7/2J7FNZPXFQ5wSFHPp92taQ2z7cySlkpwbjzdhWQKBgQDnKnXXsnEEEeE+7tPp\nTKm/r8BdX1u7En3oxD0fzvyG5XMQI9jPvznLcBhelS05MJ5AM/bipNaq0+yxNOA+\n9bFGUk9RFTv1T2yLgchnWJsGgbLzDc+Q7IPlTfu3kGXhdfNaT3wBQ2zlyB4or7Zb\nGZXHqut2CXtZUNalALXUTmuwLQKBgQDRPiC1+Kkb8qMu99tWx+AaOI4yGYjRGYjt\nquMJlec7lYzWK8mP0lFKSbRupUrERtUl4YJPNEs74HZRd2P2O9i2f/O15r8YpYe4\nq68PAi4wWkMu0PgX4eDzu6/bAPFEzpHbQT454fFNHN0k3GZEvEovXEnLoXefCtO0\ng5bbtMyPDQKBgBxYVILSWpFJL/lrh4h/4gCjKE9KthDGP7Ie/e+8jbJsyCMegiNX\nK1sKr7JJUHZGw+DTfl5kH/5qZgeEmEqgDdC1nwwuMl/uET05+LzTXI77sFmlQ+Jd\nXf7sBQWJKzTuJZzZSjdNcTjrxObwRMf+5yLIjn62wu08wFUTNTiHNSUxAoGAHbfR\nNYcnnYyMmLMjN+QTdCXsLDlsMSaLSBYpQMdFnDTVV+XILuEUIpTtR3m6cL4Or96j\nf1RSCyMKKu/mBqbhVKo/luo64elAqX3Y0yzgGoX3XMLZOasfI/7YLyanHbqpBJbz\nm4z3lJgw4HddHw4PC0L0SDOapcAH6W93EaVQGFECgYEA3X6no2WpmlNfCiLCyyWq\nacfYDqAoUo8o6NciUNlbXwEzp2YO5LqEBH/mkFG61iIVkSh+vV1cTh0gg1KQ8WgK\nTr8Uj5zmDQVBkZ7oaU+aI+mcezMZU39Ra00sr/DifNj82o6l/yd8oIUlIt10S7lT\npYbCENhkc/AjAs65LSzJvBc=\n-----END PRIVATE KEY-----\n" } } }, "changed": false } TASK [fedora.linux_system_roles.certificate : Stop tracking certificates] ****** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:176 Sunday 04 February 2024 19:38:43 +0000 (0:00:00.026) 0:00:15.732 ******* ok: [sut] => (item={'cert': '/etc/pki/tls/certs/quadlet_demo.crt', 'cert_content': '-----BEGIN CERTIFICATE-----\nMIIDgzCCAmugAwIBAgIRAJrf8hV5C0hFpoAAVX1vdtIwDQYJKoZIhvcNAQELBQAw\nUDEgMB4GA1UEAwwXTG9jYWwgU2lnbmluZyBBdXRob3JpdHkxLDAqBgNVBAMMIzlh\nZGZmMjE1LTc5MGI0ODQ1LWE2ODAwMDU1LTdkNmY3NmQxMB4XDTI0MDIwNDE5Mzg0\nMloXDTI1MDIwNDE5Mzg0MlowFDESMBAGA1UEAxMJbG9jYWxob3N0MIIBIjANBgkq\nhkiG9w0BAQEFAAOCAQ8AMIIBCgKCAQEAvPHEB0M1me4z0OvEHyyIQVxDViB9uXKY\nGWIeRBfi3py16I8tZTAeK7yvGUzWaudwBJssOJujPNTh4fNIiRo6B6f9tChLRJYx\nJsvnePqndGBjcftMdPkGHuCfK4zvkwR9KgtMswoOxjg4whEim2e+JDU6vVb4vKUe\nyTgYDjVMZ39oHYrJiGPnheh/OBrPjW5f/3JioS3qZNuWKk0dPD+8uS8rQ2JeuQqc\noeBGHLnkFSylk7K6xaLnr2Ke9GdkTSSuCEXfvbKDsBROm4x3h2evEVqcyctH4wqM\nxiob2UFYAy7hvwc/dkKyeq1YeNBlUEx72RGmCpJFwB/tTh3FKr0VSQIDAQABo4GT\nMIGQMAsGA1UdDwQEAwIFoDAUBgNVHREEDTALgglsb2NhbGhvc3QwHQYDVR0lBBYw\nFAYIKwYBBQUHAwEGCCsGAQUFBwMCMAwGA1UdEwEB/wQCMAAwHQYDVR0OBBYEFNVV\nrKueVKBFEGPIfDH5Uv04DmoVMB8GA1UdIwQYMBaAFBZOqi3YlQ4LagVeSCriQKcw\n9evPMA0GCSqGSIb3DQEBCwUAA4IBAQBl9W48RTaIwli/lKO1mYt6RgLE8fYEGhSY\n9aM6EVkE9RzmnjfxRi06FS1vLuIOPD2qMddVUB5Otjxm0JCN6E4pm/v3uQ3sVm+N\nt4zLQ4VsWDGpjE/dDwZSf50W2NaJPqK8eQnTTuhg77W6bIw6rQafEYTsugSOlyTc\n3U6Z60Y+wf0cu5awzV0NGHzVQYnV3fAMcrmb5yUQaaa1JoadS4GmC3iOwddIE4Ox\nlUwmknmkJw/gvQfXbUWXHYHVYpVAv4xu3oduIXk56xoKZkiNYlosEEC77yF8PcyU\nX3ZXRyf9XVd6cKVYvTjwZTO2q4JCbujPzW40SM2VEZ9OYa3cZvM+\n-----END CERTIFICATE-----\n', 'key': '/etc/pki/tls/private/quadlet_demo.key', 'key_content': '-----BEGIN PRIVATE KEY-----\nMIIEvQIBADANBgkqhkiG9w0BAQEFAASCBKcwggSjAgEAAoIBAQC88cQHQzWZ7jPQ\n68QfLIhBXENWIH25cpgZYh5EF+LenLXojy1lMB4rvK8ZTNZq53AEmyw4m6M81OHh\n80iJGjoHp/20KEtEljEmy+d4+qd0YGNx+0x0+QYe4J8rjO+TBH0qC0yzCg7GODjC\nESKbZ74kNTq9Vvi8pR7JOBgONUxnf2gdismIY+eF6H84Gs+Nbl//cmKhLepk25Yq\nTR08P7y5LytDYl65Cpyh4EYcueQVLKWTsrrFouevYp70Z2RNJK4IRd+9soOwFE6b\njHeHZ68RWpzJy0fjCozGKhvZQVgDLuG/Bz92QrJ6rVh40GVQTHvZEaYKkkXAH+1O\nHcUqvRVJAgMBAAECggEAINv06QRtvnr4JPXTsk421hMbHM/ivY0p2YRAWjxDNs7P\nLgY0cVwYayfR1dcQlR/x0cNAwgmDLpGhRP/f1gQrImi7TQkS6fcqziH4/AevdqRK\ntyhqiA9E2rraw6KriYhm0hiL+rkdj6h4anu7ZoZ4wWhGcmY9NOWMgJcSS+pGAtuP\nYqt0+8uYeC60elH37iDkFwMidmeoDL8nJwCu2X3D++uRlMiHsdFSX2r/g23NkxGD\n8spDv0nCJ2Srw7nmt+jDt+5vzuS+g1ATsEhT33Qh2LtESznigR8+LmJdMpYOl6d/\na7/2J7FNZPXFQ5wSFHPp92taQ2z7cySlkpwbjzdhWQKBgQDnKnXXsnEEEeE+7tPp\nTKm/r8BdX1u7En3oxD0fzvyG5XMQI9jPvznLcBhelS05MJ5AM/bipNaq0+yxNOA+\n9bFGUk9RFTv1T2yLgchnWJsGgbLzDc+Q7IPlTfu3kGXhdfNaT3wBQ2zlyB4or7Zb\nGZXHqut2CXtZUNalALXUTmuwLQKBgQDRPiC1+Kkb8qMu99tWx+AaOI4yGYjRGYjt\nquMJlec7lYzWK8mP0lFKSbRupUrERtUl4YJPNEs74HZRd2P2O9i2f/O15r8YpYe4\nq68PAi4wWkMu0PgX4eDzu6/bAPFEzpHbQT454fFNHN0k3GZEvEovXEnLoXefCtO0\ng5bbtMyPDQKBgBxYVILSWpFJL/lrh4h/4gCjKE9KthDGP7Ie/e+8jbJsyCMegiNX\nK1sKr7JJUHZGw+DTfl5kH/5qZgeEmEqgDdC1nwwuMl/uET05+LzTXI77sFmlQ+Jd\nXf7sBQWJKzTuJZzZSjdNcTjrxObwRMf+5yLIjn62wu08wFUTNTiHNSUxAoGAHbfR\nNYcnnYyMmLMjN+QTdCXsLDlsMSaLSBYpQMdFnDTVV+XILuEUIpTtR3m6cL4Or96j\nf1RSCyMKKu/mBqbhVKo/luo64elAqX3Y0yzgGoX3XMLZOasfI/7YLyanHbqpBJbz\nm4z3lJgw4HddHw4PC0L0SDOapcAH6W93EaVQGFECgYEA3X6no2WpmlNfCiLCyyWq\nacfYDqAoUo8o6NciUNlbXwEzp2YO5LqEBH/mkFG61iIVkSh+vV1cTh0gg1KQ8WgK\nTr8Uj5zmDQVBkZ7oaU+aI+mcezMZU39Ra00sr/DifNj82o6l/yd8oIUlIt10S7lT\npYbCENhkc/AjAs65LSzJvBc=\n-----END PRIVATE KEY-----\n', 'ca': '/etc/pki/tls/certs/quadlet_demo.crt', 'ca_content': '-----BEGIN CERTIFICATE-----\nMIIDgzCCAmugAwIBAgIRAJrf8hV5C0hFpoAAVX1vdtIwDQYJKoZIhvcNAQELBQAw\nUDEgMB4GA1UEAwwXTG9jYWwgU2lnbmluZyBBdXRob3JpdHkxLDAqBgNVBAMMIzlh\nZGZmMjE1LTc5MGI0ODQ1LWE2ODAwMDU1LTdkNmY3NmQxMB4XDTI0MDIwNDE5Mzg0\nMloXDTI1MDIwNDE5Mzg0MlowFDESMBAGA1UEAxMJbG9jYWxob3N0MIIBIjANBgkq\nhkiG9w0BAQEFAAOCAQ8AMIIBCgKCAQEAvPHEB0M1me4z0OvEHyyIQVxDViB9uXKY\nGWIeRBfi3py16I8tZTAeK7yvGUzWaudwBJssOJujPNTh4fNIiRo6B6f9tChLRJYx\nJsvnePqndGBjcftMdPkGHuCfK4zvkwR9KgtMswoOxjg4whEim2e+JDU6vVb4vKUe\nyTgYDjVMZ39oHYrJiGPnheh/OBrPjW5f/3JioS3qZNuWKk0dPD+8uS8rQ2JeuQqc\noeBGHLnkFSylk7K6xaLnr2Ke9GdkTSSuCEXfvbKDsBROm4x3h2evEVqcyctH4wqM\nxiob2UFYAy7hvwc/dkKyeq1YeNBlUEx72RGmCpJFwB/tTh3FKr0VSQIDAQABo4GT\nMIGQMAsGA1UdDwQEAwIFoDAUBgNVHREEDTALgglsb2NhbGhvc3QwHQYDVR0lBBYw\nFAYIKwYBBQUHAwEGCCsGAQUFBwMCMAwGA1UdEwEB/wQCMAAwHQYDVR0OBBYEFNVV\nrKueVKBFEGPIfDH5Uv04DmoVMB8GA1UdIwQYMBaAFBZOqi3YlQ4LagVeSCriQKcw\n9evPMA0GCSqGSIb3DQEBCwUAA4IBAQBl9W48RTaIwli/lKO1mYt6RgLE8fYEGhSY\n9aM6EVkE9RzmnjfxRi06FS1vLuIOPD2qMddVUB5Otjxm0JCN6E4pm/v3uQ3sVm+N\nt4zLQ4VsWDGpjE/dDwZSf50W2NaJPqK8eQnTTuhg77W6bIw6rQafEYTsugSOlyTc\n3U6Z60Y+wf0cu5awzV0NGHzVQYnV3fAMcrmb5yUQaaa1JoadS4GmC3iOwddIE4Ox\nlUwmknmkJw/gvQfXbUWXHYHVYpVAv4xu3oduIXk56xoKZkiNYlosEEC77yF8PcyU\nX3ZXRyf9XVd6cKVYvTjwZTO2q4JCbujPzW40SM2VEZ9OYa3cZvM+\n-----END CERTIFICATE-----\n'}) => { "ansible_loop_var": "item", "changed": false, "cmd": [ "getcert", "stop-tracking", "-f", "/etc/pki/tls/certs/quadlet_demo.crt" ], "delta": "0:00:00.033498", "end": "2024-02-04 19:38:44.213262", "item": { "ca": "/etc/pki/tls/certs/quadlet_demo.crt", "ca_content": "-----BEGIN CERTIFICATE-----\nMIIDgzCCAmugAwIBAgIRAJrf8hV5C0hFpoAAVX1vdtIwDQYJKoZIhvcNAQELBQAw\nUDEgMB4GA1UEAwwXTG9jYWwgU2lnbmluZyBBdXRob3JpdHkxLDAqBgNVBAMMIzlh\nZGZmMjE1LTc5MGI0ODQ1LWE2ODAwMDU1LTdkNmY3NmQxMB4XDTI0MDIwNDE5Mzg0\nMloXDTI1MDIwNDE5Mzg0MlowFDESMBAGA1UEAxMJbG9jYWxob3N0MIIBIjANBgkq\nhkiG9w0BAQEFAAOCAQ8AMIIBCgKCAQEAvPHEB0M1me4z0OvEHyyIQVxDViB9uXKY\nGWIeRBfi3py16I8tZTAeK7yvGUzWaudwBJssOJujPNTh4fNIiRo6B6f9tChLRJYx\nJsvnePqndGBjcftMdPkGHuCfK4zvkwR9KgtMswoOxjg4whEim2e+JDU6vVb4vKUe\nyTgYDjVMZ39oHYrJiGPnheh/OBrPjW5f/3JioS3qZNuWKk0dPD+8uS8rQ2JeuQqc\noeBGHLnkFSylk7K6xaLnr2Ke9GdkTSSuCEXfvbKDsBROm4x3h2evEVqcyctH4wqM\nxiob2UFYAy7hvwc/dkKyeq1YeNBlUEx72RGmCpJFwB/tTh3FKr0VSQIDAQABo4GT\nMIGQMAsGA1UdDwQEAwIFoDAUBgNVHREEDTALgglsb2NhbGhvc3QwHQYDVR0lBBYw\nFAYIKwYBBQUHAwEGCCsGAQUFBwMCMAwGA1UdEwEB/wQCMAAwHQYDVR0OBBYEFNVV\nrKueVKBFEGPIfDH5Uv04DmoVMB8GA1UdIwQYMBaAFBZOqi3YlQ4LagVeSCriQKcw\n9evPMA0GCSqGSIb3DQEBCwUAA4IBAQBl9W48RTaIwli/lKO1mYt6RgLE8fYEGhSY\n9aM6EVkE9RzmnjfxRi06FS1vLuIOPD2qMddVUB5Otjxm0JCN6E4pm/v3uQ3sVm+N\nt4zLQ4VsWDGpjE/dDwZSf50W2NaJPqK8eQnTTuhg77W6bIw6rQafEYTsugSOlyTc\n3U6Z60Y+wf0cu5awzV0NGHzVQYnV3fAMcrmb5yUQaaa1JoadS4GmC3iOwddIE4Ox\nlUwmknmkJw/gvQfXbUWXHYHVYpVAv4xu3oduIXk56xoKZkiNYlosEEC77yF8PcyU\nX3ZXRyf9XVd6cKVYvTjwZTO2q4JCbujPzW40SM2VEZ9OYa3cZvM+\n-----END CERTIFICATE-----\n", "cert": "/etc/pki/tls/certs/quadlet_demo.crt", "cert_content": "-----BEGIN CERTIFICATE-----\nMIIDgzCCAmugAwIBAgIRAJrf8hV5C0hFpoAAVX1vdtIwDQYJKoZIhvcNAQELBQAw\nUDEgMB4GA1UEAwwXTG9jYWwgU2lnbmluZyBBdXRob3JpdHkxLDAqBgNVBAMMIzlh\nZGZmMjE1LTc5MGI0ODQ1LWE2ODAwMDU1LTdkNmY3NmQxMB4XDTI0MDIwNDE5Mzg0\nMloXDTI1MDIwNDE5Mzg0MlowFDESMBAGA1UEAxMJbG9jYWxob3N0MIIBIjANBgkq\nhkiG9w0BAQEFAAOCAQ8AMIIBCgKCAQEAvPHEB0M1me4z0OvEHyyIQVxDViB9uXKY\nGWIeRBfi3py16I8tZTAeK7yvGUzWaudwBJssOJujPNTh4fNIiRo6B6f9tChLRJYx\nJsvnePqndGBjcftMdPkGHuCfK4zvkwR9KgtMswoOxjg4whEim2e+JDU6vVb4vKUe\nyTgYDjVMZ39oHYrJiGPnheh/OBrPjW5f/3JioS3qZNuWKk0dPD+8uS8rQ2JeuQqc\noeBGHLnkFSylk7K6xaLnr2Ke9GdkTSSuCEXfvbKDsBROm4x3h2evEVqcyctH4wqM\nxiob2UFYAy7hvwc/dkKyeq1YeNBlUEx72RGmCpJFwB/tTh3FKr0VSQIDAQABo4GT\nMIGQMAsGA1UdDwQEAwIFoDAUBgNVHREEDTALgglsb2NhbGhvc3QwHQYDVR0lBBYw\nFAYIKwYBBQUHAwEGCCsGAQUFBwMCMAwGA1UdEwEB/wQCMAAwHQYDVR0OBBYEFNVV\nrKueVKBFEGPIfDH5Uv04DmoVMB8GA1UdIwQYMBaAFBZOqi3YlQ4LagVeSCriQKcw\n9evPMA0GCSqGSIb3DQEBCwUAA4IBAQBl9W48RTaIwli/lKO1mYt6RgLE8fYEGhSY\n9aM6EVkE9RzmnjfxRi06FS1vLuIOPD2qMddVUB5Otjxm0JCN6E4pm/v3uQ3sVm+N\nt4zLQ4VsWDGpjE/dDwZSf50W2NaJPqK8eQnTTuhg77W6bIw6rQafEYTsugSOlyTc\n3U6Z60Y+wf0cu5awzV0NGHzVQYnV3fAMcrmb5yUQaaa1JoadS4GmC3iOwddIE4Ox\nlUwmknmkJw/gvQfXbUWXHYHVYpVAv4xu3oduIXk56xoKZkiNYlosEEC77yF8PcyU\nX3ZXRyf9XVd6cKVYvTjwZTO2q4JCbujPzW40SM2VEZ9OYa3cZvM+\n-----END CERTIFICATE-----\n", "key": "/etc/pki/tls/private/quadlet_demo.key", "key_content": "-----BEGIN PRIVATE KEY-----\nMIIEvQIBADANBgkqhkiG9w0BAQEFAASCBKcwggSjAgEAAoIBAQC88cQHQzWZ7jPQ\n68QfLIhBXENWIH25cpgZYh5EF+LenLXojy1lMB4rvK8ZTNZq53AEmyw4m6M81OHh\n80iJGjoHp/20KEtEljEmy+d4+qd0YGNx+0x0+QYe4J8rjO+TBH0qC0yzCg7GODjC\nESKbZ74kNTq9Vvi8pR7JOBgONUxnf2gdismIY+eF6H84Gs+Nbl//cmKhLepk25Yq\nTR08P7y5LytDYl65Cpyh4EYcueQVLKWTsrrFouevYp70Z2RNJK4IRd+9soOwFE6b\njHeHZ68RWpzJy0fjCozGKhvZQVgDLuG/Bz92QrJ6rVh40GVQTHvZEaYKkkXAH+1O\nHcUqvRVJAgMBAAECggEAINv06QRtvnr4JPXTsk421hMbHM/ivY0p2YRAWjxDNs7P\nLgY0cVwYayfR1dcQlR/x0cNAwgmDLpGhRP/f1gQrImi7TQkS6fcqziH4/AevdqRK\ntyhqiA9E2rraw6KriYhm0hiL+rkdj6h4anu7ZoZ4wWhGcmY9NOWMgJcSS+pGAtuP\nYqt0+8uYeC60elH37iDkFwMidmeoDL8nJwCu2X3D++uRlMiHsdFSX2r/g23NkxGD\n8spDv0nCJ2Srw7nmt+jDt+5vzuS+g1ATsEhT33Qh2LtESznigR8+LmJdMpYOl6d/\na7/2J7FNZPXFQ5wSFHPp92taQ2z7cySlkpwbjzdhWQKBgQDnKnXXsnEEEeE+7tPp\nTKm/r8BdX1u7En3oxD0fzvyG5XMQI9jPvznLcBhelS05MJ5AM/bipNaq0+yxNOA+\n9bFGUk9RFTv1T2yLgchnWJsGgbLzDc+Q7IPlTfu3kGXhdfNaT3wBQ2zlyB4or7Zb\nGZXHqut2CXtZUNalALXUTmuwLQKBgQDRPiC1+Kkb8qMu99tWx+AaOI4yGYjRGYjt\nquMJlec7lYzWK8mP0lFKSbRupUrERtUl4YJPNEs74HZRd2P2O9i2f/O15r8YpYe4\nq68PAi4wWkMu0PgX4eDzu6/bAPFEzpHbQT454fFNHN0k3GZEvEovXEnLoXefCtO0\ng5bbtMyPDQKBgBxYVILSWpFJL/lrh4h/4gCjKE9KthDGP7Ie/e+8jbJsyCMegiNX\nK1sKr7JJUHZGw+DTfl5kH/5qZgeEmEqgDdC1nwwuMl/uET05+LzTXI77sFmlQ+Jd\nXf7sBQWJKzTuJZzZSjdNcTjrxObwRMf+5yLIjn62wu08wFUTNTiHNSUxAoGAHbfR\nNYcnnYyMmLMjN+QTdCXsLDlsMSaLSBYpQMdFnDTVV+XILuEUIpTtR3m6cL4Or96j\nf1RSCyMKKu/mBqbhVKo/luo64elAqX3Y0yzgGoX3XMLZOasfI/7YLyanHbqpBJbz\nm4z3lJgw4HddHw4PC0L0SDOapcAH6W93EaVQGFECgYEA3X6no2WpmlNfCiLCyyWq\nacfYDqAoUo8o6NciUNlbXwEzp2YO5LqEBH/mkFG61iIVkSh+vV1cTh0gg1KQ8WgK\nTr8Uj5zmDQVBkZ7oaU+aI+mcezMZU39Ra00sr/DifNj82o6l/yd8oIUlIt10S7lT\npYbCENhkc/AjAs65LSzJvBc=\n-----END PRIVATE KEY-----\n" }, "rc": 0, "start": "2024-02-04 19:38:44.179764" } STDOUT: Request "20240204193842" removed. TASK [fedora.linux_system_roles.certificate : Remove files] ******************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:181 Sunday 04 February 2024 19:38:44 +0000 (0:00:00.346) 0:00:16.078 ******* changed: [sut] => (item=/etc/pki/tls/certs/quadlet_demo.crt) => { "ansible_loop_var": "item", "changed": true, "item": "/etc/pki/tls/certs/quadlet_demo.crt", "path": "/etc/pki/tls/certs/quadlet_demo.crt", "state": "absent" } changed: [sut] => (item=/etc/pki/tls/private/quadlet_demo.key) => { "ansible_loop_var": "item", "changed": true, "item": "/etc/pki/tls/private/quadlet_demo.key", "path": "/etc/pki/tls/private/quadlet_demo.key", "state": "absent" } ok: [sut] => (item=/etc/pki/tls/certs/quadlet_demo.crt) => { "ansible_loop_var": "item", "changed": false, "item": "/etc/pki/tls/certs/quadlet_demo.crt", "path": "/etc/pki/tls/certs/quadlet_demo.crt", "state": "absent" } TASK [Run the role] ************************************************************ task path: /WORKDIR/git-fix-linger-handlingciq7eztj/tests/tests_quadlet_demo.yml:50 Sunday 04 February 2024 19:38:44 +0000 (0:00:00.720) 0:00:16.799 ******* TASK [fedora.linux_system_roles.podman : Set platform/version specific variables] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:3 Sunday 04 February 2024 19:38:44 +0000 (0:00:00.038) 0:00:16.837 ******* included: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml for sut TASK [fedora.linux_system_roles.podman : Ensure ansible_facts used by role] **** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:3 Sunday 04 February 2024 19:38:45 +0000 (0:00:00.018) 0:00:16.856 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_required_facts | difference(ansible_facts.keys() | list) | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if system is ostree] ************ task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:11 Sunday 04 February 2024 19:38:45 +0000 (0:00:00.021) 0:00:16.877 ******* ok: [sut] => { "changed": false, "stat": { "exists": false } } TASK [fedora.linux_system_roles.podman : Set flag to indicate system is ostree] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:16 Sunday 04 February 2024 19:38:45 +0000 (0:00:00.232) 0:00:17.109 ******* ok: [sut] => { "ansible_facts": { "__podman_is_ostree": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Set platform/version specific variables] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:20 Sunday 04 February 2024 19:38:45 +0000 (0:00:00.017) 0:00:17.127 ******* ok: [sut] => (item=RedHat.yml) => { "ansible_facts": { "__podman_packages": [ "podman", "shadow-utils-subid" ] }, "ansible_included_var_files": [ "/WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/vars/RedHat.yml" ], "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml" } skipping: [sut] => (item=Fedora.yml) => { "ansible_loop_var": "item", "changed": false, "false_condition": "__vars_file is file", "item": "Fedora.yml", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=Fedora_39.yml) => { "ansible_loop_var": "item", "changed": false, "false_condition": "__vars_file is file", "item": "Fedora_39.yml", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=Fedora_39.yml) => { "ansible_loop_var": "item", "changed": false, "false_condition": "__vars_file is file", "item": "Fedora_39.yml", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Gather the package facts] ************* task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:6 Sunday 04 February 2024 19:38:45 +0000 (0:00:00.027) 0:00:17.154 ******* ok: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Enable copr if requested] ************* task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:10 Sunday 04 February 2024 19:38:46 +0000 (0:00:00.995) 0:00:18.149 ******* skipping: [sut] => { "changed": false, "false_condition": "podman_use_copr | d(false)", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Ensure required packages are installed] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:14 Sunday 04 February 2024 19:38:46 +0000 (0:00:00.031) 0:00:18.181 ******* changed: [sut] => { "changed": true, "rc": 0, "results": [ "Installed: libmd-1.1.0-2.fc39.x86_64", "Installed: slirp4netns-1.2.2-1.fc39.x86_64", "Installed: conmon-2:2.1.8-2.fc39.x86_64", "Installed: libbsd-0.11.7-5.fc39.x86_64", "Installed: container-selinux-2:2.229.0-1.fc39.noarch", "Installed: aardvark-dns-1.9.0-1.fc39.x86_64", "Installed: podman-5:4.9.0-1.fc39.x86_64", "Installed: catatonit-0.1.7-18.fc39.x86_64", "Installed: libslirp-4.7.0-4.fc39.x86_64", "Installed: fuse-common-3.16.1-1.fc39.x86_64", "Installed: gvisor-tap-vsock-6:0.7.2-1.fc39.x86_64", "Installed: gvisor-tap-vsock-gvforwarder-6:0.7.2-1.fc39.x86_64", "Installed: fuse-overlayfs-1.12-2.fc39.x86_64", "Installed: netavark-1.10.1-5.fc39.x86_64", "Installed: criu-3.19-2.fc39.x86_64", "Installed: fuse3-3.16.1-1.fc39.x86_64", "Installed: criu-libs-3.19-2.fc39.x86_64", "Installed: fuse3-libs-3.16.1-1.fc39.x86_64", "Installed: crun-1.14-1.fc39.x86_64", "Installed: shadow-utils-subid-2:4.14.0-2.fc39.x86_64", "Installed: qemu-user-static-2:8.1.3-2.fc39.x86_64", "Installed: qemu-user-static-aarch64-2:8.1.3-2.fc39.x86_64", "Installed: qemu-user-static-alpha-2:8.1.3-2.fc39.x86_64", "Installed: qemu-user-static-arm-2:8.1.3-2.fc39.x86_64", "Installed: qemu-user-static-cris-2:8.1.3-2.fc39.x86_64", "Installed: qemu-user-static-hexagon-2:8.1.3-2.fc39.x86_64", "Installed: qemu-user-static-hppa-2:8.1.3-2.fc39.x86_64", "Installed: yajl-2.1.0-22.fc39.x86_64", "Installed: qemu-user-static-loongarch64-2:8.1.3-2.fc39.x86_64", "Installed: qemu-user-static-m68k-2:8.1.3-2.fc39.x86_64", "Installed: qemu-user-static-microblaze-2:8.1.3-2.fc39.x86_64", "Installed: qemu-user-static-mips-2:8.1.3-2.fc39.x86_64", "Installed: qemu-user-static-nios2-2:8.1.3-2.fc39.x86_64", "Installed: passt-0^20231230.gf091893-1.fc39.x86_64", "Installed: libnet-1.3-1.fc39.x86_64", "Installed: passt-selinux-0^20231230.gf091893-1.fc39.noarch", "Installed: qemu-user-static-or1k-2:8.1.3-2.fc39.x86_64", "Installed: qemu-user-static-ppc-2:8.1.3-2.fc39.x86_64", "Installed: qemu-user-static-riscv-2:8.1.3-2.fc39.x86_64", "Installed: qemu-user-static-s390x-2:8.1.3-2.fc39.x86_64", "Installed: qemu-user-static-sh4-2:8.1.3-2.fc39.x86_64", "Installed: qemu-user-static-sparc-2:8.1.3-2.fc39.x86_64", "Installed: qemu-user-static-x86-2:8.1.3-2.fc39.x86_64", "Installed: qemu-user-static-xtensa-2:8.1.3-2.fc39.x86_64", "Installed: containers-common-4:1-95.fc39.noarch", "Installed: containers-common-extra-4:1-95.fc39.noarch" ] } TASK [fedora.linux_system_roles.podman : Get podman version] ******************* task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:22 Sunday 04 February 2024 19:40:06 +0000 (0:01:20.585) 0:01:38.767 ******* ok: [sut] => { "changed": false, "cmd": [ "podman", "--version" ], "delta": "0:00:00.035069", "end": "2024-02-04 19:40:07.182105", "rc": 0, "start": "2024-02-04 19:40:07.147036" } STDOUT: podman version 4.9.0 TASK [fedora.linux_system_roles.podman : Set podman version] ******************* task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:27 Sunday 04 February 2024 19:40:07 +0000 (0:00:00.288) 0:01:39.055 ******* ok: [sut] => { "ansible_facts": { "podman_version": "4.9.0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Podman package version must be 4.2 or later] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:31 Sunday 04 February 2024 19:40:07 +0000 (0:00:00.023) 0:01:39.079 ******* skipping: [sut] => { "changed": false, "false_condition": "podman_version is version(\"4.2\", \"<\")", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Podman package version must be 4.4 or later for quadlet, secrets] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:38 Sunday 04 February 2024 19:40:07 +0000 (0:00:00.023) 0:01:39.102 ******* skipping: [sut] => { "changed": false, "false_condition": "podman_version is version(\"4.4\", \"<\")", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Podman package version must be 4.4 or later for quadlet, secrets] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:48 Sunday 04 February 2024 19:40:07 +0000 (0:00:00.052) 0:01:39.155 ******* META: end_host conditional evaluated to False, continuing execution for sut skipping: [sut] => { "skip_reason": "end_host conditional evaluated to False, continuing execution for sut" } MSG: end_host conditional evaluated to false, continuing execution for sut TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:55 Sunday 04 February 2024 19:40:07 +0000 (0:00:00.054) 0:01:39.209 ******* included: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for sut TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Sunday 04 February 2024 19:40:07 +0000 (0:00:00.060) 0:01:39.270 ******* ok: [sut] => { "ansible_facts": { "getent_passwd": { "root": [ "x", "0", "0", "Super User", "/root", "/bin/bash" ] } }, "changed": false } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Sunday 04 February 2024 19:40:07 +0000 (0:00:00.365) 0:01:39.636 ******* skipping: [sut] => { "changed": false, "false_condition": "not ansible_facts[\"getent_passwd\"][__podman_user]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Sunday 04 February 2024 19:40:07 +0000 (0:00:00.024) 0:01:39.660 ******* ok: [sut] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get group information] **************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:27 Sunday 04 February 2024 19:40:07 +0000 (0:00:00.032) 0:01:39.692 ******* ok: [sut] => { "ansible_facts": { "getent_group": { "root": [ "x", "0", "" ] } }, "changed": false } TASK [fedora.linux_system_roles.podman : Set group name] *********************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:34 Sunday 04 February 2024 19:40:08 +0000 (0:00:00.402) 0:01:40.095 ******* ok: [sut] => { "ansible_facts": { "__podman_group_name": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:38 Sunday 04 February 2024 19:40:08 +0000 (0:00:00.031) 0:01:40.127 ******* ok: [sut] => { "changed": false, "stat": { "atime": 1696291200.0, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "72ba28a3be19c467e89ab90f5d21f99b605159a7", "ctime": 1707075555.3834877, "dev": 51714, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 203277, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-pie-executable", "mode": "0755", "mtime": 1696291200.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 16032, "uid": 0, "version": "1460241045", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************ task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:49 Sunday 04 February 2024 19:40:08 +0000 (0:00:00.338) 0:01:40.465 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check group with getsubids] *********** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:53 Sunday 04 February 2024 19:40:08 +0000 (0:00:00.024) 0:01:40.489 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if user is in subuid file] ****** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:63 Sunday 04 February 2024 19:40:08 +0000 (0:00:00.027) 0:01:40.517 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:71 Sunday 04 February 2024 19:40:08 +0000 (0:00:00.029) 0:01:40.546 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if group is in subgid file] ***** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:78 Sunday 04 February 2024 19:40:08 +0000 (0:00:00.023) 0:01:40.570 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] ***** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:86 Sunday 04 February 2024 19:40:08 +0000 (0:00:00.019) 0:01:40.589 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set config file paths] **************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:61 Sunday 04 February 2024 19:40:08 +0000 (0:00:00.018) 0:01:40.608 ******* ok: [sut] => { "ansible_facts": { "__podman_container_conf_file": "/etc/containers/containers.conf.d/50-systemroles.conf", "__podman_policy_json_file": "/etc/containers/policy.json", "__podman_registries_conf_file": "/etc/containers/registries.conf.d/50-systemroles.conf", "__podman_storage_conf_file": "/etc/containers/storage.conf" }, "changed": false } TASK [fedora.linux_system_roles.podman : Handle container.conf.d] ************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:70 Sunday 04 February 2024 19:40:08 +0000 (0:00:00.041) 0:01:40.649 ******* included: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_container_conf_d.yml for sut TASK [fedora.linux_system_roles.podman : Ensure containers.d exists] *********** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_container_conf_d.yml:5 Sunday 04 February 2024 19:40:08 +0000 (0:00:00.038) 0:01:40.688 ******* skipping: [sut] => { "changed": false, "false_condition": "podman_containers_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Update container config file] ********* task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_container_conf_d.yml:13 Sunday 04 February 2024 19:40:08 +0000 (0:00:00.019) 0:01:40.708 ******* skipping: [sut] => { "changed": false, "false_condition": "podman_containers_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Handle registries.conf.d] ************* task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:73 Sunday 04 February 2024 19:40:08 +0000 (0:00:00.019) 0:01:40.728 ******* included: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_registries_conf_d.yml for sut TASK [fedora.linux_system_roles.podman : Ensure registries.d exists] *********** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_registries_conf_d.yml:5 Sunday 04 February 2024 19:40:08 +0000 (0:00:00.062) 0:01:40.791 ******* skipping: [sut] => { "changed": false, "false_condition": "podman_registries_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Update registries config file] ******** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_registries_conf_d.yml:13 Sunday 04 February 2024 19:40:08 +0000 (0:00:00.024) 0:01:40.816 ******* skipping: [sut] => { "changed": false, "false_condition": "podman_registries_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Handle storage.conf] ****************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:76 Sunday 04 February 2024 19:40:08 +0000 (0:00:00.021) 0:01:40.837 ******* included: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_storage_conf.yml for sut TASK [fedora.linux_system_roles.podman : Ensure storage.conf parent dir exists] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_storage_conf.yml:5 Sunday 04 February 2024 19:40:09 +0000 (0:00:00.043) 0:01:40.881 ******* skipping: [sut] => { "changed": false, "false_condition": "podman_storage_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Update storage config file] *********** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_storage_conf.yml:13 Sunday 04 February 2024 19:40:09 +0000 (0:00:00.022) 0:01:40.904 ******* skipping: [sut] => { "changed": false, "false_condition": "podman_storage_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Handle policy.json] ******************* task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:79 Sunday 04 February 2024 19:40:09 +0000 (0:00:00.023) 0:01:40.927 ******* included: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml for sut TASK [fedora.linux_system_roles.podman : Ensure policy.json parent dir exists] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:6 Sunday 04 February 2024 19:40:09 +0000 (0:00:00.044) 0:01:40.972 ******* skipping: [sut] => { "changed": false, "false_condition": "podman_policy_json | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Stat the policy.json file] ************ task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:14 Sunday 04 February 2024 19:40:09 +0000 (0:00:00.025) 0:01:40.997 ******* skipping: [sut] => { "changed": false, "false_condition": "podman_policy_json | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get the existing policy.json] ********* task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:19 Sunday 04 February 2024 19:40:09 +0000 (0:00:00.021) 0:01:41.019 ******* skipping: [sut] => { "changed": false, "false_condition": "podman_policy_json | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Write new policy.json file] *********** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:25 Sunday 04 February 2024 19:40:09 +0000 (0:00:00.022) 0:01:41.041 ******* skipping: [sut] => { "changed": false, "false_condition": "podman_policy_json | length > 0", "skip_reason": "Conditional result was False" } TASK [Manage firewall for specified ports] ************************************* task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:85 Sunday 04 February 2024 19:40:09 +0000 (0:00:00.022) 0:01:41.063 ******* TASK [fedora.linux_system_roles.firewall : Setup firewalld] ******************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:2 Sunday 04 February 2024 19:40:09 +0000 (0:00:00.093) 0:01:41.157 ******* included: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml for sut TASK [fedora.linux_system_roles.firewall : Ensure ansible_facts used by role] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:2 Sunday 04 February 2024 19:40:09 +0000 (0:00:00.037) 0:01:41.194 ******* skipping: [sut] => { "changed": false, "false_condition": "__firewall_required_facts | difference(ansible_facts.keys() | list) | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Check if system is ostree] ********** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:10 Sunday 04 February 2024 19:40:09 +0000 (0:00:00.055) 0:01:41.250 ******* ok: [sut] => { "changed": false, "stat": { "exists": false } } TASK [fedora.linux_system_roles.firewall : Set flag to indicate system is ostree] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:15 Sunday 04 February 2024 19:40:09 +0000 (0:00:00.247) 0:01:41.498 ******* ok: [sut] => { "ansible_facts": { "__firewall_is_ostree": false }, "changed": false } TASK [fedora.linux_system_roles.firewall : Install firewalld] ****************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:19 Sunday 04 February 2024 19:40:09 +0000 (0:00:00.026) 0:01:41.525 ******* ok: [sut] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [fedora.linux_system_roles.firewall : Collect service facts] ************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:5 Sunday 04 February 2024 19:40:11 +0000 (0:00:02.020) 0:01:43.545 ******* skipping: [sut] => { "changed": false, "false_condition": "firewall_disable_conflicting_services | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Attempt to stop and disable conflicting services] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:9 Sunday 04 February 2024 19:40:11 +0000 (0:00:00.020) 0:01:43.565 ******* skipping: [sut] => (item=nftables) => { "ansible_loop_var": "item", "changed": false, "false_condition": "firewall_disable_conflicting_services | bool", "item": "nftables", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=iptables) => { "ansible_loop_var": "item", "changed": false, "false_condition": "firewall_disable_conflicting_services | bool", "item": "iptables", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=ufw) => { "ansible_loop_var": "item", "changed": false, "false_condition": "firewall_disable_conflicting_services | bool", "item": "ufw", "skip_reason": "Conditional result was False" } skipping: [sut] => { "changed": false } MSG: All items skipped TASK [fedora.linux_system_roles.firewall : Unmask firewalld service] *********** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:22 Sunday 04 February 2024 19:40:11 +0000 (0:00:00.026) 0:01:43.591 ******* ok: [sut] => { "changed": false, "name": "firewalld", "status": { "AccessSELinuxContext": "system_u:object_r:firewalld_unit_file_t:s0", "ActiveEnterTimestampMonotonic": "0", "ActiveExitTimestampMonotonic": "0", "ActiveState": "inactive", "After": "basic.target dbus.socket system.slice polkit.service dbus-broker.service sysinit.target", "AllowIsolate": "no", "AssertResult": "no", "AssertTimestampMonotonic": "0", "Before": "network-pre.target shutdown.target", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "BusName": "org.fedoraproject.FirewallD1", "CPUAccounting": "yes", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "[not set]", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "yes", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore", "CleanResult": "success", "CollectMode": "inactive", "ConditionResult": "no", "ConditionTimestampMonotonic": "0", "ConfigurationDirectoryMode": "0755", "Conflicts": "iptables.service ebtables.service shutdown.target nftables.service ip6tables.service ipset.service", "ControlGroupId": "0", "ControlPID": "0", "CoredumpFilter": "0x33", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "DefaultStartupMemoryLow": "0", "Delegate": "no", "Description": "firewalld - dynamic firewall daemon", "DevicePolicy": "auto", "Documentation": "\"man:firewalld(1)\"", "DropInPaths": "/usr/lib/systemd/system/service.d/10-timeout-abort.conf", "DynamicUser": "no", "EnvironmentFiles": "/etc/sysconfig/firewalld (ignore_errors=yes)", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "0", "ExecMainStartTimestampMonotonic": "0", "ExecMainStatus": "0", "ExecReload": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecReloadEx": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStart": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStartEx": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExitType": "main", "ExtensionImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FileDescriptorStorePreserve": "restart", "FinalKillSignal": "9", "FragmentPath": "/usr/lib/systemd/system/firewalld.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOReadBytes": "18446744073709551615", "IOReadOperations": "18446744073709551615", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "IOWeight": "[not set]", "IOWriteBytes": "18446744073709551615", "IOWriteOperations": "18446744073709551615", "IPAccounting": "no", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "Id": "firewalld.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestampMonotonic": "0", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "mixed", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "infinity", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "8388608", "LimitMEMLOCKSoft": "8388608", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "524288", "LimitNOFILESoft": "1024", "LimitNPROC": "14778", "LimitNPROCSoft": "14778", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "14778", "LimitSIGPENDINGSoft": "14778", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "0", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "ManagedOOMSwap": "auto", "MemoryAccounting": "yes", "MemoryAvailable": "infinity", "MemoryCurrent": "[not set]", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryKSM": "no", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemoryPressureThresholdUSec": "200ms", "MemoryPressureWatch": "auto", "MemorySwapMax": "infinity", "MemoryZSwapMax": "infinity", "MountAPIVFS": "no", "MountImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAPolicy": "n/a", "Names": "firewalld.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMPolicy": "stop", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "OnSuccessJobMode": "fail", "Perpetual": "no", "PrivateDevices": "no", "PrivateIPC": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProcSubset": "all", "ProtectClock": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectHostname": "no", "ProtectKernelLogs": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectProc": "default", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "ReloadResult": "success", "ReloadSignal": "1", "RemainAfterExit": "no", "RemoveIPC": "no", "Requires": "dbus.socket system.slice sysinit.target", "Restart": "no", "RestartKillSignal": "15", "RestartMaxDelayUSec": "infinity", "RestartMode": "normal", "RestartSteps": "0", "RestartUSec": "100ms", "RestartUSecNext": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RootEphemeral": "no", "RootImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "Slice": "system.slice", "StandardError": "null", "StandardInput": "null", "StandardOutput": "null", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StartupMemoryHigh": "18446744073709551615", "StartupMemoryLow": "0", "StartupMemoryMax": "18446744073709551615", "StartupMemorySwapMax": "18446744073709551615", "StartupMemoryZSwapMax": "18446744073709551615", "StateChangeTimestampMonotonic": "0", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "dead", "SuccessAction": "none", "SyslogFacility": "3", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "2147483646", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "[not set]", "TasksMax": "4433", "TimeoutAbortUSec": "45s", "TimeoutCleanUSec": "infinity", "TimeoutStartFailureMode": "terminate", "TimeoutStartUSec": "45s", "TimeoutStopFailureMode": "abort", "TimeoutStopUSec": "45s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "dbus", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "enabled", "UnitFileState": "disabled", "UtmpMode": "init", "Wants": "network-pre.target", "WatchdogSignal": "6", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "infinity" } } TASK [fedora.linux_system_roles.firewall : Enable and start firewalld service] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:28 Sunday 04 February 2024 19:40:12 +0000 (0:00:00.446) 0:01:44.038 ******* changed: [sut] => { "changed": true, "enabled": true, "name": "firewalld", "state": "started", "status": { "AccessSELinuxContext": "system_u:object_r:firewalld_unit_file_t:s0", "ActiveEnterTimestampMonotonic": "0", "ActiveExitTimestampMonotonic": "0", "ActiveState": "inactive", "After": "basic.target sysinit.target dbus.socket polkit.service system.slice dbus-broker.service", "AllowIsolate": "no", "AssertResult": "no", "AssertTimestampMonotonic": "0", "Before": "network-pre.target shutdown.target", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "BusName": "org.fedoraproject.FirewallD1", "CPUAccounting": "yes", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "[not set]", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "yes", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore", "CleanResult": "success", "CollectMode": "inactive", "ConditionResult": "no", "ConditionTimestampMonotonic": "0", "ConfigurationDirectoryMode": "0755", "Conflicts": "ip6tables.service iptables.service ipset.service ebtables.service nftables.service shutdown.target", "ControlGroupId": "0", "ControlPID": "0", "CoredumpFilter": "0x33", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "DefaultStartupMemoryLow": "0", "Delegate": "no", "Description": "firewalld - dynamic firewall daemon", "DevicePolicy": "auto", "Documentation": "\"man:firewalld(1)\"", "DropInPaths": "/usr/lib/systemd/system/service.d/10-timeout-abort.conf", "DynamicUser": "no", "EnvironmentFiles": "/etc/sysconfig/firewalld (ignore_errors=yes)", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "0", "ExecMainStartTimestampMonotonic": "0", "ExecMainStatus": "0", "ExecReload": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecReloadEx": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStart": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStartEx": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExitType": "main", "ExtensionImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FileDescriptorStorePreserve": "restart", "FinalKillSignal": "9", "FragmentPath": "/usr/lib/systemd/system/firewalld.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOReadBytes": "18446744073709551615", "IOReadOperations": "18446744073709551615", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "IOWeight": "[not set]", "IOWriteBytes": "18446744073709551615", "IOWriteOperations": "18446744073709551615", "IPAccounting": "no", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "Id": "firewalld.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestampMonotonic": "0", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "mixed", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "infinity", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "8388608", "LimitMEMLOCKSoft": "8388608", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "524288", "LimitNOFILESoft": "1024", "LimitNPROC": "14778", "LimitNPROCSoft": "14778", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "14778", "LimitSIGPENDINGSoft": "14778", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "0", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "ManagedOOMSwap": "auto", "MemoryAccounting": "yes", "MemoryAvailable": "infinity", "MemoryCurrent": "[not set]", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryKSM": "no", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemoryPressureThresholdUSec": "200ms", "MemoryPressureWatch": "auto", "MemorySwapMax": "infinity", "MemoryZSwapMax": "infinity", "MountAPIVFS": "no", "MountImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAPolicy": "n/a", "Names": "firewalld.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMPolicy": "stop", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "OnSuccessJobMode": "fail", "Perpetual": "no", "PrivateDevices": "no", "PrivateIPC": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProcSubset": "all", "ProtectClock": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectHostname": "no", "ProtectKernelLogs": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectProc": "default", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "ReloadResult": "success", "ReloadSignal": "1", "RemainAfterExit": "no", "RemoveIPC": "no", "Requires": "system.slice sysinit.target dbus.socket", "Restart": "no", "RestartKillSignal": "15", "RestartMaxDelayUSec": "infinity", "RestartMode": "normal", "RestartSteps": "0", "RestartUSec": "100ms", "RestartUSecNext": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RootEphemeral": "no", "RootImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "Slice": "system.slice", "StandardError": "null", "StandardInput": "null", "StandardOutput": "null", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StartupMemoryHigh": "18446744073709551615", "StartupMemoryLow": "0", "StartupMemoryMax": "18446744073709551615", "StartupMemorySwapMax": "18446744073709551615", "StartupMemoryZSwapMax": "18446744073709551615", "StateChangeTimestampMonotonic": "0", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "dead", "SuccessAction": "none", "SyslogFacility": "3", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "2147483646", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "[not set]", "TasksMax": "4433", "TimeoutAbortUSec": "45s", "TimeoutCleanUSec": "infinity", "TimeoutStartFailureMode": "terminate", "TimeoutStartUSec": "45s", "TimeoutStopFailureMode": "abort", "TimeoutStopUSec": "45s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "dbus", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "enabled", "UnitFileState": "disabled", "UtmpMode": "init", "Wants": "network-pre.target", "WatchdogSignal": "6", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "infinity" } } TASK [fedora.linux_system_roles.firewall : Check if previous replaced is defined] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:34 Sunday 04 February 2024 19:40:14 +0000 (0:00:02.083) 0:01:46.121 ******* ok: [sut] => { "ansible_facts": { "__firewall_previous_replaced": false, "__firewall_python_cmd": "/usr/bin/python3", "__firewall_report_changed": true }, "changed": false } TASK [fedora.linux_system_roles.firewall : Get config files, checksums before and remove] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:43 Sunday 04 February 2024 19:40:14 +0000 (0:00:00.033) 0:01:46.155 ******* skipping: [sut] => { "changed": false, "false_condition": "__firewall_previous_replaced | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Tell firewall module it is able to report changed] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:55 Sunday 04 February 2024 19:40:14 +0000 (0:00:00.018) 0:01:46.174 ******* skipping: [sut] => { "changed": false, "false_condition": "__firewall_previous_replaced | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Configure firewall] ***************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:71 Sunday 04 February 2024 19:40:14 +0000 (0:00:00.023) 0:01:46.197 ******* changed: [sut] => (item={'port': '8000/tcp', 'state': 'enabled'}) => { "__firewall_changed": true, "ansible_loop_var": "item", "changed": true, "item": { "port": "8000/tcp", "state": "enabled" } } changed: [sut] => (item={'port': '9000/tcp', 'state': 'enabled'}) => { "__firewall_changed": true, "ansible_loop_var": "item", "changed": true, "item": { "port": "9000/tcp", "state": "enabled" } } TASK [fedora.linux_system_roles.firewall : Gather firewall config information] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:120 Sunday 04 February 2024 19:40:16 +0000 (0:00:01.657) 0:01:47.854 ******* skipping: [sut] => (item={'port': '8000/tcp', 'state': 'enabled'}) => { "ansible_loop_var": "item", "changed": false, "false_condition": "firewall | length == 1", "item": { "port": "8000/tcp", "state": "enabled" }, "skip_reason": "Conditional result was False" } skipping: [sut] => (item={'port': '9000/tcp', 'state': 'enabled'}) => { "ansible_loop_var": "item", "changed": false, "false_condition": "firewall | length == 1", "item": { "port": "9000/tcp", "state": "enabled" }, "skip_reason": "Conditional result was False" } skipping: [sut] => { "changed": false } MSG: All items skipped TASK [fedora.linux_system_roles.firewall : Update firewalld_config fact] ******* task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:130 Sunday 04 February 2024 19:40:16 +0000 (0:00:00.043) 0:01:47.898 ******* skipping: [sut] => { "changed": false, "false_condition": "firewall | length == 1", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Gather firewall config if no arguments] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:139 Sunday 04 February 2024 19:40:16 +0000 (0:00:00.030) 0:01:47.928 ******* skipping: [sut] => { "changed": false, "false_condition": "firewall == None or firewall | length == 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Update firewalld_config fact] ******* task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:144 Sunday 04 February 2024 19:40:16 +0000 (0:00:00.030) 0:01:47.959 ******* skipping: [sut] => { "changed": false, "false_condition": "firewall == None or firewall | length == 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Get config files, checksums after] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:153 Sunday 04 February 2024 19:40:16 +0000 (0:00:00.028) 0:01:47.987 ******* skipping: [sut] => { "changed": false, "false_condition": "__firewall_previous_replaced | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Calculate what has changed] ********* task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:163 Sunday 04 February 2024 19:40:16 +0000 (0:00:00.022) 0:01:48.009 ******* skipping: [sut] => { "changed": false, "false_condition": "__firewall_previous_replaced | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Show diffs] ************************* task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:169 Sunday 04 February 2024 19:40:16 +0000 (0:00:00.019) 0:01:48.028 ******* skipping: [sut] => { "false_condition": "__firewall_previous_replaced | bool" } TASK [Manage selinux for specified ports] ************************************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:92 Sunday 04 February 2024 19:40:16 +0000 (0:00:00.052) 0:01:48.081 ******* skipping: [sut] => { "changed": false, "false_condition": "podman_selinux_ports | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Keep track of users that need to cancel linger] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:99 Sunday 04 February 2024 19:40:16 +0000 (0:00:00.021) 0:01:48.102 ******* ok: [sut] => { "ansible_facts": { "__podman_cancel_user_linger": [] }, "changed": false } TASK [fedora.linux_system_roles.podman : Handle secrets] *********************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:103 Sunday 04 February 2024 19:40:16 +0000 (0:00:00.021) 0:01:48.124 ******* included: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml for sut => (item=(censored due to no_log)) included: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml for sut => (item=(censored due to no_log)) included: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml for sut => (item=(censored due to no_log)) TASK [fedora.linux_system_roles.podman : Set variables part 0] ***************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:3 Sunday 04 February 2024 19:40:16 +0000 (0:00:00.092) 0:01:48.217 ******* ok: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Set variables part 1] ***************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:10 Sunday 04 February 2024 19:40:16 +0000 (0:00:00.026) 0:01:48.244 ******* ok: [sut] => { "ansible_facts": { "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set variables part 2] ***************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:14 Sunday 04 February 2024 19:40:16 +0000 (0:00:00.023) 0:01:48.267 ******* ok: [sut] => { "ansible_facts": { "__podman_rootless": false, "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:20 Sunday 04 February 2024 19:40:16 +0000 (0:00:00.027) 0:01:48.295 ******* included: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for sut TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Sunday 04 February 2024 19:40:16 +0000 (0:00:00.032) 0:01:48.328 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Sunday 04 February 2024 19:40:16 +0000 (0:00:00.020) 0:01:48.348 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Sunday 04 February 2024 19:40:16 +0000 (0:00:00.018) 0:01:48.367 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Manage each secret] ******************* task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:25 Sunday 04 February 2024 19:40:16 +0000 (0:00:00.019) 0:01:48.386 ******* ok: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Set variables part 0] ***************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:3 Sunday 04 February 2024 19:40:16 +0000 (0:00:00.405) 0:01:48.792 ******* ok: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Set variables part 1] ***************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:10 Sunday 04 February 2024 19:40:16 +0000 (0:00:00.028) 0:01:48.821 ******* ok: [sut] => { "ansible_facts": { "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set variables part 2] ***************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:14 Sunday 04 February 2024 19:40:17 +0000 (0:00:00.024) 0:01:48.846 ******* ok: [sut] => { "ansible_facts": { "__podman_rootless": false, "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:20 Sunday 04 February 2024 19:40:17 +0000 (0:00:00.028) 0:01:48.874 ******* included: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for sut TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Sunday 04 February 2024 19:40:17 +0000 (0:00:00.032) 0:01:48.906 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Sunday 04 February 2024 19:40:17 +0000 (0:00:00.043) 0:01:48.949 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Sunday 04 February 2024 19:40:17 +0000 (0:00:00.020) 0:01:48.970 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Manage each secret] ******************* task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:25 Sunday 04 February 2024 19:40:17 +0000 (0:00:00.019) 0:01:48.989 ******* ok: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Set variables part 0] ***************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:3 Sunday 04 February 2024 19:40:17 +0000 (0:00:00.297) 0:01:49.286 ******* ok: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Set variables part 1] ***************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:10 Sunday 04 February 2024 19:40:17 +0000 (0:00:00.027) 0:01:49.314 ******* ok: [sut] => { "ansible_facts": { "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set variables part 2] ***************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:14 Sunday 04 February 2024 19:40:17 +0000 (0:00:00.025) 0:01:49.339 ******* ok: [sut] => { "ansible_facts": { "__podman_rootless": false, "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:20 Sunday 04 February 2024 19:40:17 +0000 (0:00:00.028) 0:01:49.368 ******* included: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for sut TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Sunday 04 February 2024 19:40:17 +0000 (0:00:00.035) 0:01:49.403 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Sunday 04 February 2024 19:40:17 +0000 (0:00:00.019) 0:01:49.423 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Sunday 04 February 2024 19:40:17 +0000 (0:00:00.019) 0:01:49.443 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Manage each secret] ******************* task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:25 Sunday 04 February 2024 19:40:17 +0000 (0:00:00.018) 0:01:49.462 ******* ok: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Handle Kubernetes specifications] ***** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:110 Sunday 04 February 2024 19:40:17 +0000 (0:00:00.297) 0:01:49.759 ******* skipping: [sut] => { "changed": false, "skipped_reason": "No items in the list" } TASK [fedora.linux_system_roles.podman : Handle Quadlet specifications] ******** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:116 Sunday 04 February 2024 19:40:17 +0000 (0:00:00.018) 0:01:49.778 ******* included: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for sut => (item={'file_src': 'quadlet-demo.network'}) included: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for sut => (item={'file_src': 'quadlet-demo-mysql.volume'}) included: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for sut => (item={'template_src': 'quadlet-demo-mysql.container.j2'}) included: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for sut => (item={'file_src': 'envoy-proxy-configmap.yml'}) included: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for sut => (item={'template_src': 'quadlet-demo.yml.j2'}) included: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for sut => (item={'file_src': 'quadlet-demo.kube'}) TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14 Sunday 04 February 2024 19:40:18 +0000 (0:00:00.104) 0:01:49.883 ******* ok: [sut] => { "ansible_facts": { "__podman_quadlet_file_src": "quadlet-demo.network", "__podman_quadlet_spec": {}, "__podman_quadlet_str": "[Network]\nSubnet=192.168.30.0/24\nGateway=192.168.30.1\nLabel=app=wordpress", "__podman_quadlet_template_src": "" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25 Sunday 04 February 2024 19:40:18 +0000 (0:00:00.033) 0:01:49.916 ******* ok: [sut] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_pull_image": true, "__podman_state": "created", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] ***** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35 Sunday 04 February 2024 19:40:18 +0000 (0:00:00.051) 0:01:49.967 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_quadlet_file_src", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48 Sunday 04 February 2024 19:40:18 +0000 (0:00:00.022) 0:01:49.990 ******* ok: [sut] => { "ansible_facts": { "__podman_quadlet_name": "quadlet-demo", "__podman_quadlet_type": "network", "__podman_rootless": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57 Sunday 04 February 2024 19:40:18 +0000 (0:00:00.036) 0:01:50.026 ******* included: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for sut TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Sunday 04 February 2024 19:40:18 +0000 (0:00:00.041) 0:01:50.067 ******* skipping: [sut] => { "changed": false, "false_condition": "'getent_passwd' not in ansible_facts or __podman_user not in ansible_facts['getent_passwd']", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Sunday 04 February 2024 19:40:18 +0000 (0:00:00.024) 0:01:50.091 ******* skipping: [sut] => { "changed": false, "false_condition": "not ansible_facts[\"getent_passwd\"][__podman_user]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Sunday 04 February 2024 19:40:18 +0000 (0:00:00.023) 0:01:50.114 ******* ok: [sut] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get group information] **************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:27 Sunday 04 February 2024 19:40:18 +0000 (0:00:00.029) 0:01:50.144 ******* ok: [sut] => { "ansible_facts": { "getent_group": { "root": [ "x", "0", "" ] } }, "changed": false } TASK [fedora.linux_system_roles.podman : Set group name] *********************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:34 Sunday 04 February 2024 19:40:18 +0000 (0:00:00.252) 0:01:50.397 ******* ok: [sut] => { "ansible_facts": { "__podman_group_name": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:38 Sunday 04 February 2024 19:40:18 +0000 (0:00:00.029) 0:01:50.426 ******* ok: [sut] => { "changed": false, "stat": { "atime": 1707075608.5761826, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "72ba28a3be19c467e89ab90f5d21f99b605159a7", "ctime": 1707075555.3834877, "dev": 51714, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 203277, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-pie-executable", "mode": "0755", "mtime": 1696291200.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 16032, "uid": 0, "version": "1460241045", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************ task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:49 Sunday 04 February 2024 19:40:18 +0000 (0:00:00.248) 0:01:50.675 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check group with getsubids] *********** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:53 Sunday 04 February 2024 19:40:18 +0000 (0:00:00.020) 0:01:50.695 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if user is in subuid file] ****** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:63 Sunday 04 February 2024 19:40:18 +0000 (0:00:00.019) 0:01:50.714 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:71 Sunday 04 February 2024 19:40:18 +0000 (0:00:00.020) 0:01:50.735 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if group is in subgid file] ***** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:78 Sunday 04 February 2024 19:40:18 +0000 (0:00:00.019) 0:01:50.755 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] ***** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:86 Sunday 04 February 2024 19:40:18 +0000 (0:00:00.019) 0:01:50.774 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62 Sunday 04 February 2024 19:40:18 +0000 (0:00:00.019) 0:01:50.794 ******* ok: [sut] => { "ansible_facts": { "__podman_activate_systemd_unit": true, "__podman_images_found": [], "__podman_kube_yamls_raw": "", "__podman_service_name": "", "__podman_systemd_scope": "system", "__podman_user_home_dir": "/root", "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73 Sunday 04 February 2024 19:40:18 +0000 (0:00:00.038) 0:01:50.832 ******* ok: [sut] => { "ansible_facts": { "__podman_quadlet_path": "/etc/containers/systemd" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get kube yaml contents] *************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77 Sunday 04 February 2024 19:40:19 +0000 (0:00:00.023) 0:01:50.855 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_kube_yamls_raw | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87 Sunday 04 February 2024 19:40:19 +0000 (0:00:00.021) 0:01:50.877 ******* ok: [sut] => { "ansible_facts": { "__podman_images": [], "__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo.network", "__podman_volumes": [] }, "changed": false } TASK [fedora.linux_system_roles.podman : Cleanup quadlets] ********************* task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:103 Sunday 04 February 2024 19:40:19 +0000 (0:00:00.080) 0:01:50.957 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_state == \"absent\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create and update quadlets] *********** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:107 Sunday 04 February 2024 19:40:19 +0000 (0:00:00.020) 0:01:50.978 ******* included: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml for sut TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:2 Sunday 04 February 2024 19:40:19 +0000 (0:00:00.047) 0:01:51.026 ******* included: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for sut TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Sunday 04 February 2024 19:40:19 +0000 (0:00:00.035) 0:01:51.061 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Sunday 04 February 2024 19:40:19 +0000 (0:00:00.019) 0:01:51.080 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Sunday 04 February 2024 19:40:19 +0000 (0:00:00.018) 0:01:51.099 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create host directories] ************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:7 Sunday 04 February 2024 19:40:19 +0000 (0:00:00.019) 0:01:51.118 ******* skipping: [sut] => { "changed": false, "skipped_reason": "No items in the list" } TASK [fedora.linux_system_roles.podman : Ensure container images are present] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:20 Sunday 04 February 2024 19:40:19 +0000 (0:00:00.017) 0:01:51.136 ******* skipping: [sut] => { "changed": false, "skipped_reason": "No items in the list" } TASK [fedora.linux_system_roles.podman : Ensure the quadlet directory is present] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:37 Sunday 04 February 2024 19:40:19 +0000 (0:00:00.020) 0:01:51.157 ******* ok: [sut] => { "changed": false, "gid": 0, "group": "root", "mode": "0755", "owner": "root", "path": "/etc/containers/systemd", "secontext": "system_u:object_r:etc_t:s0", "size": 4096, "state": "directory", "uid": 0 } TASK [fedora.linux_system_roles.podman : Ensure quadlet file is copied] ******** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:46 Sunday 04 February 2024 19:40:19 +0000 (0:00:00.261) 0:01:51.418 ******* changed: [sut] => { "changed": true, "checksum": "e57c08d49aff4bae8daab138d913aeddaa8682a0", "dest": "/etc/containers/systemd/quadlet-demo.network", "gid": 0, "group": "root", "md5sum": "061f3cf318cbd8ab5794bb1173831fb8", "mode": "0644", "owner": "root", "secontext": "system_u:object_r:etc_t:s0", "size": 74, "src": "/root/.ansible/tmp/ansible-tmp-1707075619.6203864-7913-42245081858217/source", "state": "file", "uid": 0 } TASK [fedora.linux_system_roles.podman : Ensure quadlet file content is present] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:56 Sunday 04 February 2024 19:40:20 +0000 (0:00:00.726) 0:01:52.144 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_quadlet_file_src", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Ensure quadlet file is present] ******* task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:68 Sunday 04 February 2024 19:40:20 +0000 (0:00:00.022) 0:01:52.167 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_copy_file is skipped", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Reload systemctl] ********************* task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:80 Sunday 04 February 2024 19:40:20 +0000 (0:00:00.020) 0:01:52.187 ******* ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.podman : Start service] ************************ task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:108 Sunday 04 February 2024 19:40:21 +0000 (0:00:00.881) 0:01:53.068 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_service_name | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Restart service] ********************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:123 Sunday 04 February 2024 19:40:21 +0000 (0:00:00.023) 0:01:53.092 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_service_name | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14 Sunday 04 February 2024 19:40:21 +0000 (0:00:00.023) 0:01:53.116 ******* ok: [sut] => { "ansible_facts": { "__podman_quadlet_file_src": "quadlet-demo-mysql.volume", "__podman_quadlet_spec": {}, "__podman_quadlet_str": "[Volume]", "__podman_quadlet_template_src": "" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25 Sunday 04 February 2024 19:40:21 +0000 (0:00:00.031) 0:01:53.148 ******* ok: [sut] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_pull_image": true, "__podman_state": "created", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] ***** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35 Sunday 04 February 2024 19:40:21 +0000 (0:00:00.052) 0:01:53.201 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_quadlet_file_src", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48 Sunday 04 February 2024 19:40:21 +0000 (0:00:00.022) 0:01:53.223 ******* ok: [sut] => { "ansible_facts": { "__podman_quadlet_name": "quadlet-demo-mysql", "__podman_quadlet_type": "volume", "__podman_rootless": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57 Sunday 04 February 2024 19:40:21 +0000 (0:00:00.034) 0:01:53.258 ******* included: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for sut TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Sunday 04 February 2024 19:40:21 +0000 (0:00:00.040) 0:01:53.298 ******* skipping: [sut] => { "changed": false, "false_condition": "'getent_passwd' not in ansible_facts or __podman_user not in ansible_facts['getent_passwd']", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Sunday 04 February 2024 19:40:21 +0000 (0:00:00.024) 0:01:53.323 ******* skipping: [sut] => { "changed": false, "false_condition": "not ansible_facts[\"getent_passwd\"][__podman_user]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Sunday 04 February 2024 19:40:21 +0000 (0:00:00.024) 0:01:53.347 ******* ok: [sut] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get group information] **************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:27 Sunday 04 February 2024 19:40:21 +0000 (0:00:00.028) 0:01:53.376 ******* ok: [sut] => { "ansible_facts": { "getent_group": { "root": [ "x", "0", "" ] } }, "changed": false } TASK [fedora.linux_system_roles.podman : Set group name] *********************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:34 Sunday 04 February 2024 19:40:21 +0000 (0:00:00.256) 0:01:53.633 ******* ok: [sut] => { "ansible_facts": { "__podman_group_name": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:38 Sunday 04 February 2024 19:40:21 +0000 (0:00:00.029) 0:01:53.662 ******* ok: [sut] => { "changed": false, "stat": { "atime": 1707075608.5761826, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "72ba28a3be19c467e89ab90f5d21f99b605159a7", "ctime": 1707075555.3834877, "dev": 51714, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 203277, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-pie-executable", "mode": "0755", "mtime": 1696291200.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 16032, "uid": 0, "version": "1460241045", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************ task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:49 Sunday 04 February 2024 19:40:22 +0000 (0:00:00.261) 0:01:53.924 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check group with getsubids] *********** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:53 Sunday 04 February 2024 19:40:22 +0000 (0:00:00.024) 0:01:53.948 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if user is in subuid file] ****** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:63 Sunday 04 February 2024 19:40:22 +0000 (0:00:00.019) 0:01:53.968 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:71 Sunday 04 February 2024 19:40:22 +0000 (0:00:00.020) 0:01:53.988 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if group is in subgid file] ***** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:78 Sunday 04 February 2024 19:40:22 +0000 (0:00:00.018) 0:01:54.007 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] ***** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:86 Sunday 04 February 2024 19:40:22 +0000 (0:00:00.019) 0:01:54.027 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62 Sunday 04 February 2024 19:40:22 +0000 (0:00:00.019) 0:01:54.046 ******* ok: [sut] => { "ansible_facts": { "__podman_activate_systemd_unit": true, "__podman_images_found": [], "__podman_kube_yamls_raw": "", "__podman_service_name": "quadlet-demo-mysql-volume.service", "__podman_systemd_scope": "system", "__podman_user_home_dir": "/root", "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73 Sunday 04 February 2024 19:40:22 +0000 (0:00:00.039) 0:01:54.086 ******* ok: [sut] => { "ansible_facts": { "__podman_quadlet_path": "/etc/containers/systemd" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get kube yaml contents] *************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77 Sunday 04 February 2024 19:40:22 +0000 (0:00:00.023) 0:01:54.110 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_kube_yamls_raw | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87 Sunday 04 February 2024 19:40:22 +0000 (0:00:00.047) 0:01:54.157 ******* ok: [sut] => { "ansible_facts": { "__podman_images": [], "__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo-mysql.volume", "__podman_volumes": [] }, "changed": false } TASK [fedora.linux_system_roles.podman : Cleanup quadlets] ********************* task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:103 Sunday 04 February 2024 19:40:22 +0000 (0:00:00.057) 0:01:54.214 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_state == \"absent\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create and update quadlets] *********** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:107 Sunday 04 February 2024 19:40:22 +0000 (0:00:00.018) 0:01:54.233 ******* included: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml for sut TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:2 Sunday 04 February 2024 19:40:22 +0000 (0:00:00.045) 0:01:54.279 ******* included: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for sut TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Sunday 04 February 2024 19:40:22 +0000 (0:00:00.035) 0:01:54.314 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Sunday 04 February 2024 19:40:22 +0000 (0:00:00.019) 0:01:54.334 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Sunday 04 February 2024 19:40:22 +0000 (0:00:00.020) 0:01:54.354 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create host directories] ************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:7 Sunday 04 February 2024 19:40:22 +0000 (0:00:00.018) 0:01:54.373 ******* skipping: [sut] => { "changed": false, "skipped_reason": "No items in the list" } TASK [fedora.linux_system_roles.podman : Ensure container images are present] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:20 Sunday 04 February 2024 19:40:22 +0000 (0:00:00.018) 0:01:54.391 ******* skipping: [sut] => { "changed": false, "skipped_reason": "No items in the list" } TASK [fedora.linux_system_roles.podman : Ensure the quadlet directory is present] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:37 Sunday 04 February 2024 19:40:22 +0000 (0:00:00.020) 0:01:54.411 ******* ok: [sut] => { "changed": false, "gid": 0, "group": "root", "mode": "0755", "owner": "root", "path": "/etc/containers/systemd", "secontext": "system_u:object_r:etc_t:s0", "size": 4096, "state": "directory", "uid": 0 } TASK [fedora.linux_system_roles.podman : Ensure quadlet file is copied] ******** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:46 Sunday 04 February 2024 19:40:22 +0000 (0:00:00.259) 0:01:54.671 ******* changed: [sut] => { "changed": true, "checksum": "585f8cbdf0ec73000f9227dcffbef71e9552ea4a", "dest": "/etc/containers/systemd/quadlet-demo-mysql.volume", "gid": 0, "group": "root", "md5sum": "5ddd03a022aeb4502d9bc8ce436b4233", "mode": "0644", "owner": "root", "secontext": "system_u:object_r:etc_t:s0", "size": 9, "src": "/root/.ansible/tmp/ansible-tmp-1707075622.8734992-7961-161658592115508/source", "state": "file", "uid": 0 } TASK [fedora.linux_system_roles.podman : Ensure quadlet file content is present] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:56 Sunday 04 February 2024 19:40:23 +0000 (0:00:00.615) 0:01:55.287 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_quadlet_file_src", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Ensure quadlet file is present] ******* task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:68 Sunday 04 February 2024 19:40:23 +0000 (0:00:00.022) 0:01:55.310 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_copy_file is skipped", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Reload systemctl] ********************* task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:80 Sunday 04 February 2024 19:40:23 +0000 (0:00:00.020) 0:01:55.330 ******* ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.podman : Start service] ************************ task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:108 Sunday 04 February 2024 19:40:24 +0000 (0:00:00.877) 0:01:56.207 ******* changed: [sut] => { "changed": true, "name": "quadlet-demo-mysql-volume.service", "state": "started", "status": { "AccessSELinuxContext": "system_u:object_r:systemd_unit_file_t:s0", "ActiveEnterTimestampMonotonic": "0", "ActiveExitTimestampMonotonic": "0", "ActiveState": "inactive", "After": "systemd-journald.socket system.slice -.mount sysinit.target basic.target", "AllowIsolate": "no", "AssertResult": "no", "AssertTimestampMonotonic": "0", "Before": "shutdown.target", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "CPUAccounting": "yes", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "[not set]", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "no", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore", "CleanResult": "success", "CollectMode": "inactive", "ConditionResult": "no", "ConditionTimestampMonotonic": "0", "ConfigurationDirectoryMode": "0755", "Conflicts": "shutdown.target", "ControlGroupId": "0", "ControlPID": "0", "CoredumpFilter": "0x33", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "DefaultStartupMemoryLow": "0", "Delegate": "no", "Description": "quadlet-demo-mysql-volume.service", "DevicePolicy": "auto", "DropInPaths": "/usr/lib/systemd/system/service.d/10-timeout-abort.conf", "DynamicUser": "no", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "0", "ExecMainStartTimestampMonotonic": "0", "ExecMainStatus": "0", "ExecStart": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman volume create --ignore systemd-quadlet-demo-mysql ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStartEx": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman volume create --ignore systemd-quadlet-demo-mysql ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExitType": "main", "ExtensionImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FileDescriptorStorePreserve": "restart", "FinalKillSignal": "9", "FragmentPath": "/run/systemd/generator/quadlet-demo-mysql-volume.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOReadBytes": "18446744073709551615", "IOReadOperations": "18446744073709551615", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "IOWeight": "[not set]", "IOWriteBytes": "18446744073709551615", "IOWriteOperations": "18446744073709551615", "IPAccounting": "no", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "Id": "quadlet-demo-mysql-volume.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestampMonotonic": "0", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "control-group", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "infinity", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "8388608", "LimitMEMLOCKSoft": "8388608", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "524288", "LimitNOFILESoft": "1024", "LimitNPROC": "14778", "LimitNPROCSoft": "14778", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "14778", "LimitSIGPENDINGSoft": "14778", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "0", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "ManagedOOMSwap": "auto", "MemoryAccounting": "yes", "MemoryAvailable": "infinity", "MemoryCurrent": "[not set]", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryKSM": "no", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemoryPressureThresholdUSec": "200ms", "MemoryPressureWatch": "auto", "MemorySwapMax": "infinity", "MemoryZSwapMax": "infinity", "MountAPIVFS": "no", "MountImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAPolicy": "n/a", "Names": "quadlet-demo-mysql-volume.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMPolicy": "stop", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "OnSuccessJobMode": "fail", "Perpetual": "no", "PrivateDevices": "no", "PrivateIPC": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProcSubset": "all", "ProtectClock": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectHostname": "no", "ProtectKernelLogs": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectProc": "default", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "ReloadResult": "success", "ReloadSignal": "1", "RemainAfterExit": "yes", "RemoveIPC": "no", "Requires": "system.slice -.mount sysinit.target", "RequiresMountsFor": "/run/containers", "Restart": "no", "RestartKillSignal": "15", "RestartMaxDelayUSec": "infinity", "RestartMode": "normal", "RestartSteps": "0", "RestartUSec": "100ms", "RestartUSecNext": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RootEphemeral": "no", "RootImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "Slice": "system.slice", "StandardError": "inherit", "StandardInput": "null", "StandardOutput": "journal", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StartupMemoryHigh": "18446744073709551615", "StartupMemoryLow": "0", "StartupMemoryMax": "18446744073709551615", "StartupMemorySwapMax": "18446744073709551615", "StartupMemoryZSwapMax": "18446744073709551615", "StateChangeTimestampMonotonic": "0", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "dead", "SuccessAction": "none", "SyslogFacility": "3", "SyslogIdentifier": "quadlet-demo-mysql-volume", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "2147483646", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "[not set]", "TasksMax": "4433", "TimeoutAbortUSec": "45s", "TimeoutCleanUSec": "infinity", "TimeoutStartFailureMode": "terminate", "TimeoutStartUSec": "infinity", "TimeoutStopFailureMode": "abort", "TimeoutStopUSec": "45s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "oneshot", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "disabled", "UnitFileState": "generated", "UtmpMode": "init", "WatchdogSignal": "6", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "infinity" } } TASK [fedora.linux_system_roles.podman : Restart service] ********************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:123 Sunday 04 February 2024 19:40:25 +0000 (0:00:00.679) 0:01:56.887 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_service_started is changed", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14 Sunday 04 February 2024 19:40:25 +0000 (0:00:00.050) 0:01:56.938 ******* ok: [sut] => { "ansible_facts": { "__podman_quadlet_file_src": "", "__podman_quadlet_spec": {}, "__podman_quadlet_str": "[Install]\nWantedBy=default.target\n\n[Container]\nImage=quay.io/linux-system-roles/mysql:5.6\nContainerName=quadlet-demo-mysql\nVolume=quadlet-demo-mysql.volume:/var/lib/mysql\nVolume=/tmp/quadlet_demo:/var/lib/quadlet_demo:Z\nNetwork=quadlet-demo.network\nSecret=mysql-root-password-container,type=env,target=MYSQL_ROOT_PASSWORD\nHealthCmd=/usr/bin/true\nHealthOnFailure=kill\n", "__podman_quadlet_template_src": "quadlet-demo-mysql.container.j2" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25 Sunday 04 February 2024 19:40:25 +0000 (0:00:00.068) 0:01:57.006 ******* ok: [sut] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_pull_image": true, "__podman_state": "created", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] ***** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35 Sunday 04 February 2024 19:40:25 +0000 (0:00:00.028) 0:01:57.035 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_quadlet_str", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48 Sunday 04 February 2024 19:40:25 +0000 (0:00:00.022) 0:01:57.058 ******* ok: [sut] => { "ansible_facts": { "__podman_quadlet_name": "quadlet-demo-mysql", "__podman_quadlet_type": "container", "__podman_rootless": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57 Sunday 04 February 2024 19:40:25 +0000 (0:00:00.034) 0:01:57.092 ******* included: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for sut TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Sunday 04 February 2024 19:40:25 +0000 (0:00:00.041) 0:01:57.134 ******* skipping: [sut] => { "changed": false, "false_condition": "'getent_passwd' not in ansible_facts or __podman_user not in ansible_facts['getent_passwd']", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Sunday 04 February 2024 19:40:25 +0000 (0:00:00.024) 0:01:57.159 ******* skipping: [sut] => { "changed": false, "false_condition": "not ansible_facts[\"getent_passwd\"][__podman_user]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Sunday 04 February 2024 19:40:25 +0000 (0:00:00.023) 0:01:57.182 ******* ok: [sut] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get group information] **************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:27 Sunday 04 February 2024 19:40:25 +0000 (0:00:00.029) 0:01:57.212 ******* ok: [sut] => { "ansible_facts": { "getent_group": { "root": [ "x", "0", "" ] } }, "changed": false } TASK [fedora.linux_system_roles.podman : Set group name] *********************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:34 Sunday 04 February 2024 19:40:25 +0000 (0:00:00.254) 0:01:57.467 ******* ok: [sut] => { "ansible_facts": { "__podman_group_name": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:38 Sunday 04 February 2024 19:40:25 +0000 (0:00:00.029) 0:01:57.496 ******* ok: [sut] => { "changed": false, "stat": { "atime": 1707075608.5761826, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "72ba28a3be19c467e89ab90f5d21f99b605159a7", "ctime": 1707075555.3834877, "dev": 51714, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 203277, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-pie-executable", "mode": "0755", "mtime": 1696291200.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 16032, "uid": 0, "version": "1460241045", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************ task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:49 Sunday 04 February 2024 19:40:25 +0000 (0:00:00.247) 0:01:57.744 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check group with getsubids] *********** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:53 Sunday 04 February 2024 19:40:25 +0000 (0:00:00.020) 0:01:57.765 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if user is in subuid file] ****** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:63 Sunday 04 February 2024 19:40:25 +0000 (0:00:00.019) 0:01:57.784 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:71 Sunday 04 February 2024 19:40:25 +0000 (0:00:00.018) 0:01:57.803 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if group is in subgid file] ***** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:78 Sunday 04 February 2024 19:40:25 +0000 (0:00:00.020) 0:01:57.824 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] ***** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:86 Sunday 04 February 2024 19:40:26 +0000 (0:00:00.019) 0:01:57.844 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62 Sunday 04 February 2024 19:40:26 +0000 (0:00:00.020) 0:01:57.864 ******* ok: [sut] => { "ansible_facts": { "__podman_activate_systemd_unit": true, "__podman_images_found": [ "quay.io/linux-system-roles/mysql:5.6" ], "__podman_kube_yamls_raw": "", "__podman_service_name": "quadlet-demo-mysql.service", "__podman_systemd_scope": "system", "__podman_user_home_dir": "/root", "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73 Sunday 04 February 2024 19:40:26 +0000 (0:00:00.038) 0:01:57.902 ******* ok: [sut] => { "ansible_facts": { "__podman_quadlet_path": "/etc/containers/systemd" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get kube yaml contents] *************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77 Sunday 04 February 2024 19:40:26 +0000 (0:00:00.050) 0:01:57.953 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_kube_yamls_raw | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87 Sunday 04 February 2024 19:40:26 +0000 (0:00:00.022) 0:01:57.975 ******* ok: [sut] => { "ansible_facts": { "__podman_images": [ "quay.io/linux-system-roles/mysql:5.6" ], "__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo-mysql.container", "__podman_volumes": [ "/tmp/quadlet_demo" ] }, "changed": false } TASK [fedora.linux_system_roles.podman : Cleanup quadlets] ********************* task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:103 Sunday 04 February 2024 19:40:26 +0000 (0:00:00.058) 0:01:58.033 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_state == \"absent\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create and update quadlets] *********** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:107 Sunday 04 February 2024 19:40:26 +0000 (0:00:00.027) 0:01:58.061 ******* included: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml for sut TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:2 Sunday 04 February 2024 19:40:26 +0000 (0:00:00.049) 0:01:58.110 ******* included: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for sut TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Sunday 04 February 2024 19:40:26 +0000 (0:00:00.036) 0:01:58.147 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Sunday 04 February 2024 19:40:26 +0000 (0:00:00.019) 0:01:58.167 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Sunday 04 February 2024 19:40:26 +0000 (0:00:00.019) 0:01:58.187 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create host directories] ************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:7 Sunday 04 February 2024 19:40:26 +0000 (0:00:00.019) 0:01:58.206 ******* changed: [sut] => (item=/tmp/quadlet_demo) => { "ansible_loop_var": "item", "changed": true, "gid": 0, "group": "root", "item": "/tmp/quadlet_demo", "mode": "0777", "owner": "root", "path": "/tmp/quadlet_demo", "secontext": "unconfined_u:object_r:user_tmp_t:s0", "size": 40, "state": "directory", "uid": 0 } TASK [fedora.linux_system_roles.podman : Ensure container images are present] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:20 Sunday 04 February 2024 19:40:26 +0000 (0:00:00.269) 0:01:58.476 ******* changed: [sut] => (item=quay.io/linux-system-roles/mysql:5.6) => { "actions": [ "Pulled image quay.io/linux-system-roles/mysql:5.6" ], "ansible_loop_var": "item", "changed": true, "failed_when_result": false, "image": [ { "Annotations": {}, "Architecture": "amd64", "Author": "", "Comment": "", "Config": { "Cmd": [ "mysqld" ], "Entrypoint": [ "docker-entrypoint.sh" ], "Env": [ "PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin", "GOSU_VERSION=1.12", "MYSQL_MAJOR=5.6", "MYSQL_VERSION=5.6.51-1debian9" ], "ExposedPorts": { "3306/tcp": {} }, "Volumes": { "/var/lib/mysql": {} } }, "Created": "2021-12-21T02:57:59.587337844Z", "Digest": "sha256:2d0419de75076a55e0d0c0680b4f86ecdbb8a88de92054698a0028bcb218a3a8", "GraphDriver": { "Data": { "LowerDir": "/var/lib/containers/storage/overlay/63c8cde32bc583ca7971591c23194c1fd072197949f28fcad7f10780ad1af022/diff:/var/lib/containers/storage/overlay/4904a0ac4db0b6060c944fc7c882cba99b79b50f90151785350c0e4ba7f94390/diff:/var/lib/containers/storage/overlay/1822e40336d4b2b31d888f669f443965d4476dae35fd0a67e58eefd9c5ecb9e5/diff:/var/lib/containers/storage/overlay/988a3979591d296974121b0a893ec621ea5e9939ab3cc5e7c194dbe27f7b5677/diff:/var/lib/containers/storage/overlay/ba0315059f870fcb66ebd6517f455f714fb4eeeb4fdee9cefb3d7f59a7b0cdb7/diff:/var/lib/containers/storage/overlay/0dbdd9024cc38ecc226b2c4772718d32ff0abc288b41669d58d506274c16c67a/diff:/var/lib/containers/storage/overlay/f8ddb99f39386eac409b19f8b347a7a10ee91771cfb838c88793ad8947a233f5/diff:/var/lib/containers/storage/overlay/2ecb8e9889676e4b9b129bab56314644612905ce6e6ce1c92d5a86ac8f08f03b/diff:/var/lib/containers/storage/overlay/07e715cdf99b679f1e067f4819d1b7e1ba9ea24229ede4e6b45b9240fe5865f8/diff:/var/lib/containers/storage/overlay/2b83e5699838047f936a3875bcce7fe1b169983bf86785ae7519c5bc488558ae/diff", "UpperDir": "/var/lib/containers/storage/overlay/bb762056c5cde2e3213fe6ccf595176708249829a607b0afc5799075d4a852ec/diff", "WorkDir": "/var/lib/containers/storage/overlay/bb762056c5cde2e3213fe6ccf595176708249829a607b0afc5799075d4a852ec/work" }, "Name": "overlay" }, "History": [ { "created": "2021-12-21T01:24:41.626621944Z", "created_by": "/bin/sh -c #(nop) ADD file:b1650c19c1dd1f4cff71553f2bb0bc949944d0bc24b54c318b2880c14538648a in / " }, { "created": "2021-12-21T01:24:42.056427889Z", "created_by": "/bin/sh -c #(nop) CMD [\"bash\"]", "empty_layer": true }, { "created": "2021-12-21T02:56:59.289452902Z", "created_by": "/bin/sh -c groupadd -r mysql && useradd -r -g mysql mysql" }, { "created": "2021-12-21T02:57:05.692613282Z", "created_by": "/bin/sh -c apt-get update && apt-get install -y --no-install-recommends gnupg dirmngr && rm -rf /var/lib/apt/lists/*" }, { "created": "2021-12-21T02:57:05.921738618Z", "created_by": "/bin/sh -c #(nop) ENV GOSU_VERSION=1.12", "empty_layer": true }, { "created": "2021-12-21T02:57:18.981216995Z", "created_by": "/bin/sh -c set -eux; \tsavedAptMark=\"$(apt-mark showmanual)\"; \tapt-get update; \tapt-get install -y --no-install-recommends ca-certificates wget; \trm -rf /var/lib/apt/lists/*; \tdpkgArch=\"$(dpkg --print-architecture | awk -F- '{ print $NF }')\"; \twget -O /usr/local/bin/gosu \"https://github.com/tianon/gosu/releases/download/$GOSU_VERSION/gosu-$dpkgArch\"; \twget -O /usr/local/bin/gosu.asc \"https://github.com/tianon/gosu/releases/download/$GOSU_VERSION/gosu-$dpkgArch.asc\"; \texport GNUPGHOME=\"$(mktemp -d)\"; \tgpg --batch --keyserver hkps://keys.openpgp.org --recv-keys B42F6819007F00F88E364FD4036A9C25BF357DD4; \tgpg --batch --verify /usr/local/bin/gosu.asc /usr/local/bin/gosu; \tgpgconf --kill all; \trm -rf \"$GNUPGHOME\" /usr/local/bin/gosu.asc; \tapt-mark auto '.*' > /dev/null; \t[ -z \"$savedAptMark\" ] || apt-mark manual $savedAptMark > /dev/null; \tapt-get purge -y --auto-remove -o APT::AutoRemove::RecommendsImportant=false; \tchmod +x /usr/local/bin/gosu; \tgosu --version; \tgosu nobody true" }, { "created": "2021-12-21T02:57:19.781116502Z", "created_by": "/bin/sh -c mkdir /docker-entrypoint-initdb.d" }, { "created": "2021-12-21T02:57:25.555452451Z", "created_by": "/bin/sh -c apt-get update && apt-get install -y --no-install-recommends \t\tpwgen \t\tperl \t\txz-utils \t&& rm -rf /var/lib/apt/lists/*" }, { "created": "2021-12-21T02:57:37.684730941Z", "created_by": "/bin/sh -c set -ex; \tkey='A4A9406876FCBD3C456770C88C718D3B5072E1F5'; \texport GNUPGHOME=\"$(mktemp -d)\"; \tgpg --batch --keyserver keyserver.ubuntu.com --recv-keys \"$key\"; \tgpg --batch --export \"$key\" > /etc/apt/trusted.gpg.d/mysql.gpg; \tgpgconf --kill all; \trm -rf \"$GNUPGHOME\"; \tapt-key list > /dev/null" }, { "created": "2021-12-21T02:57:37.903115972Z", "created_by": "/bin/sh -c #(nop) ENV MYSQL_MAJOR=5.6", "empty_layer": true }, { "created": "2021-12-21T02:57:38.155528708Z", "created_by": "/bin/sh -c #(nop) ENV MYSQL_VERSION=5.6.51-1debian9", "empty_layer": true }, { "created": "2021-12-21T02:57:39.106901365Z", "created_by": "/bin/sh -c echo 'deb http://repo.mysql.com/apt/debian/ stretch mysql-5.6' > /etc/apt/sources.list.d/mysql.list" }, { "created": "2021-12-21T02:57:56.693791892Z", "created_by": "/bin/sh -c { \t\techo mysql-community-server mysql-community-server/data-dir select ''; \t\techo mysql-community-server mysql-community-server/root-pass password ''; \t\techo mysql-community-server mysql-community-server/re-root-pass password ''; \t\techo mysql-community-server mysql-community-server/remove-test-db select false; \t} | debconf-set-selections \t&& apt-get update \t&& apt-get install -y \t\tmysql-server=\"${MYSQL_VERSION}\" \t&& find /etc/mysql/ -name '*.cnf' -print0 \t\t| xargs -0 grep -lZE '^(bind-address|log)' \t\t| xargs -rt -0 sed -Ei 's/^(bind-address|log)/#&/' \t&& echo '[mysqld]\\nskip-host-cache\\nskip-name-resolve' > /etc/mysql/conf.d/docker.cnf \t&& rm -rf /var/lib/apt/lists/* \t&& rm -rf /var/lib/mysql && mkdir -p /var/lib/mysql /var/run/mysqld \t&& chown -R mysql:mysql /var/lib/mysql /var/run/mysqld \t&& chmod 1777 /var/run/mysqld /var/lib/mysql" }, { "created": "2021-12-21T02:57:57.29721744Z", "created_by": "/bin/sh -c #(nop) VOLUME [/var/lib/mysql]", "empty_layer": true }, { "created": "2021-12-21T02:57:57.591077869Z", "created_by": "/bin/sh -c #(nop) COPY file:345a22fe55d3e6783a17075612415413487e7dba27fbf1000a67c7870364b739 in /usr/local/bin/ " }, { "created": "2021-12-21T02:57:58.761392735Z", "created_by": "/bin/sh -c ln -s usr/local/bin/docker-entrypoint.sh /entrypoint.sh # backwards compat" }, { "created": "2021-12-21T02:57:59.04809683Z", "created_by": "/bin/sh -c #(nop) ENTRYPOINT [\"docker-entrypoint.sh\"]", "empty_layer": true }, { "created": "2021-12-21T02:57:59.31392792Z", "created_by": "/bin/sh -c #(nop) EXPOSE 3306", "empty_layer": true }, { "created": "2021-12-21T02:57:59.587337844Z", "created_by": "/bin/sh -c #(nop) CMD [\"mysqld\"]", "empty_layer": true } ], "Id": "dd3b2a5dcb48ff61113592ed5ddd762581be4387c7bc552375a2159422aa6bf5", "Labels": null, "ManifestType": "application/vnd.docker.distribution.manifest.v2+json", "NamesHistory": [ "quay.io/linux-system-roles/mysql:5.6" ], "Os": "linux", "Parent": "", "RepoDigests": [ "quay.io/linux-system-roles/mysql@sha256:2d0419de75076a55e0d0c0680b4f86ecdbb8a88de92054698a0028bcb218a3a8", "quay.io/linux-system-roles/mysql@sha256:4faa0a30c0055742c72daff1482976814f340d4bfb647f0fa2c3562fd1a7abda" ], "RepoTags": [ "quay.io/linux-system-roles/mysql:5.6" ], "RootFS": { "Layers": [ "sha256:2b83e5699838047f936a3875bcce7fe1b169983bf86785ae7519c5bc488558ae", "sha256:2e10295573911964a8acc37c738aed958d0bc385bc51df6f184b196fb592c975", "sha256:d414fdead0b9c934781e5611c016d7c237062c9470db810ef28b321a492dbb9c", "sha256:4085e588967d10dafb76ad52f66f661b41b9008cc16568ce02853de7cbbf7433", "sha256:7ea96a4e341b4d139f5e076e71d19229537292226c3b24ec2d588567667a1191", "sha256:e3dce1c82d4e4142e8e1bf2fb9c6fd2555196d40088167270ff95c9af37b88af", "sha256:2612088e90f60ecb4a13e68adbfb2b1366f82fc8fbf1f0970f8769cf3cd753ee", "sha256:eba393347f896c4028340bf1a40cc5d6b8f21eeb7abf3e79799d842a6babb93f", "sha256:7c5a5c1986b16b0d8ad4eaa8066627910ec7794a657467e8f18d223f1c0034fd", "sha256:49a1ca1cd2b80f84d6616df2b5123458f4bcb065bca28c43d9c3fd5a0480b907", "sha256:7137327a7221ddb123883146beb68003ee31d652ec46ab8bea3ba12c2df8df8f" ], "Type": "layers" }, "Size": 307607194, "User": "", "Version": "20.10.7", "VirtualSize": 307607194 } ], "item": "quay.io/linux-system-roles/mysql:5.6", "podman_actions": [ "/usr/bin/podman image ls quay.io/linux-system-roles/mysql:5.6 --format json", "/usr/bin/podman image exists quay.io/linux-system-roles/mysql:5.6", "/usr/bin/podman pull quay.io/linux-system-roles/mysql:5.6 -q", "/usr/bin/podman inspect dd3b2a5dcb48ff61113592ed5ddd762581be4387c7bc552375a2159422aa6bf5 --format json" ] } TASK [fedora.linux_system_roles.podman : Ensure the quadlet directory is present] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:37 Sunday 04 February 2024 19:40:36 +0000 (0:00:09.880) 0:02:08.356 ******* ok: [sut] => { "changed": false, "gid": 0, "group": "root", "mode": "0755", "owner": "root", "path": "/etc/containers/systemd", "secontext": "system_u:object_r:etc_t:s0", "size": 4096, "state": "directory", "uid": 0 } TASK [fedora.linux_system_roles.podman : Ensure quadlet file is copied] ******** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:46 Sunday 04 February 2024 19:40:36 +0000 (0:00:00.264) 0:02:08.621 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_quadlet_file_src | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Ensure quadlet file content is present] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:56 Sunday 04 February 2024 19:40:36 +0000 (0:00:00.021) 0:02:08.642 ******* changed: [sut] => { "changed": true, "checksum": "7bf9312d57274d4dabbd222b01717e2cc0cfa0ed", "dest": "/etc/containers/systemd/quadlet-demo-mysql.container", "gid": 0, "group": "root", "md5sum": "f1cd3875deb45b806389825bb4b5ad4a", "mode": "0644", "owner": "root", "secontext": "system_u:object_r:etc_t:s0", "size": 367, "src": "/root/.ansible/tmp/ansible-tmp-1707075636.8441758-8013-248821270230451/source", "state": "file", "uid": 0 } TASK [fedora.linux_system_roles.podman : Ensure quadlet file is present] ******* task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:68 Sunday 04 February 2024 19:40:37 +0000 (0:00:00.614) 0:02:09.256 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_copy_content is skipped", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Reload systemctl] ********************* task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:80 Sunday 04 February 2024 19:40:37 +0000 (0:00:00.020) 0:02:09.277 ******* ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.podman : Start service] ************************ task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:108 Sunday 04 February 2024 19:40:38 +0000 (0:00:00.890) 0:02:10.167 ******* fatal: [sut]: FAILED! => { "changed": false } MSG: Unable to start service quadlet-demo-mysql.service: Job for quadlet-demo-mysql.service failed because the control process exited with error code. See "systemctl status quadlet-demo-mysql.service" and "journalctl -xeu quadlet-demo-mysql.service" for details. TASK [Dump journal] ************************************************************ task path: /WORKDIR/git-fix-linger-handlingciq7eztj/tests/tests_quadlet_demo.yml:131 Sunday 04 February 2024 19:40:38 +0000 (0:00:00.611) 0:02:10.778 ******* ok: [sut] => { "changed": false, "cmd": [ "journalctl", "-ex" ], "delta": "0:00:00.043578", "end": "2024-02-04 19:40:39.194478", "rc": 0, "start": "2024-02-04 19:40:39.150900" } STDOUT: Feb 04 19:37:03 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_START pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:37:03 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: CRYPTO_KEY_USER pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:4d:0e:a0:55:e6:42:83:ce:eb:49:54:e8:e7:6b:94:70:b8:87:2b:24:6c:0c:40:0b:23:42:6b:40:30:1b:d1:22 direction=? spid=4417 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=? res=success' Feb 04 19:37:03 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm python3[4433]: ansible-ansible.legacy.command Invoked with _raw_params=curl -skL -o /etc/pki/ca-trust/source/anchors/Current-IT-Root-CAs.pem https://certs.corp.redhat.com/certs/Current-IT-Root-CAs.pem _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 04 19:37:03 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_END pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:37:03 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGOUT pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:37:03 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGIN pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:37:03 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_START pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:37:03 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: CRYPTO_KEY_USER pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:4d:0e:a0:55:e6:42:83:ce:eb:49:54:e8:e7:6b:94:70:b8:87:2b:24:6c:0c:40:0b:23:42:6b:40:30:1b:d1:22 direction=? spid=4436 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=? res=success' Feb 04 19:37:03 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm python3[4452]: ansible-ansible.legacy.command Invoked with _raw_params=bash -c 'update-ca-trust' _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 04 19:37:04 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_END pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:37:04 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGOUT pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:37:04 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGIN pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:37:04 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_START pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:37:04 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: CRYPTO_KEY_USER pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:4d:0e:a0:55:e6:42:83:ce:eb:49:54:e8:e7:6b:94:70:b8:87:2b:24:6c:0c:40:0b:23:42:6b:40:30:1b:d1:22 direction=? spid=4463 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=? res=success' Feb 04 19:37:05 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm python3[4479]: ansible-ansible.legacy.command Invoked with _raw_params=bash -c 'update-ca-trust extract' _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 04 19:37:06 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_END pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:37:06 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGOUT pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:37:06 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGIN pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:37:06 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_START pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:37:06 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: CRYPTO_KEY_USER pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:4d:0e:a0:55:e6:42:83:ce:eb:49:54:e8:e7:6b:94:70:b8:87:2b:24:6c:0c:40:0b:23:42:6b:40:30:1b:d1:22 direction=? spid=4490 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=? res=success' Feb 04 19:37:06 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm python3[4506]: ansible-ansible.legacy.command Invoked with _raw_params=curl -v https://gitlab.cee.redhat.com _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 04 19:37:06 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_END pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:37:06 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGOUT pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:37:06 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGIN pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:37:06 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_START pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:37:06 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: CRYPTO_KEY_USER pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:4d:0e:a0:55:e6:42:83:ce:eb:49:54:e8:e7:6b:94:70:b8:87:2b:24:6c:0c:40:0b:23:42:6b:40:30:1b:d1:22 direction=? spid=4509 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=? res=success' Feb 04 19:37:06 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm python3[4525]: ansible-ansible.legacy.command Invoked with _raw_params=curl -v https://beaker.engineering.redhat.com/ _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 04 19:37:07 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_END pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:37:07 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGOUT pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:37:07 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGIN pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:37:07 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_START pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:37:07 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: CRYPTO_KEY_USER pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:4d:0e:a0:55:e6:42:83:ce:eb:49:54:e8:e7:6b:94:70:b8:87:2b:24:6c:0c:40:0b:23:42:6b:40:30:1b:d1:22 direction=? spid=4528 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=? res=success' Feb 04 19:37:07 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm python3[4544]: ansible-service_facts Invoked Feb 04 19:37:09 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_END pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:37:09 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGOUT pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:37:09 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGIN pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:37:09 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_START pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:37:09 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: CRYPTO_KEY_USER pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:4d:0e:a0:55:e6:42:83:ce:eb:49:54:e8:e7:6b:94:70:b8:87:2b:24:6c:0c:40:0b:23:42:6b:40:30:1b:d1:22 direction=? spid=4652 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=? res=success' Feb 04 19:37:09 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm python3[4668]: ansible-ansible.legacy.command Invoked with _raw_params=rpm -qa _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 04 19:37:10 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_END pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:37:10 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGOUT pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:37:10 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGIN pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:37:10 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_START pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:37:10 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: CRYPTO_KEY_USER pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:4d:0e:a0:55:e6:42:83:ce:eb:49:54:e8:e7:6b:94:70:b8:87:2b:24:6c:0c:40:0b:23:42:6b:40:30:1b:d1:22 direction=? spid=4670 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=? res=success' Feb 04 19:37:10 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm python3[4686]: ansible-service_facts Invoked Feb 04 19:37:12 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_END pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:37:12 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGOUT pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:37:12 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGIN pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:37:12 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_START pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:37:12 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: CRYPTO_KEY_USER pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:4d:0e:a0:55:e6:42:83:ce:eb:49:54:e8:e7:6b:94:70:b8:87:2b:24:6c:0c:40:0b:23:42:6b:40:30:1b:d1:22 direction=? spid=4794 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=? res=success' Feb 04 19:37:13 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm python3[4810]: ansible-ansible.legacy.command Invoked with _raw_params=rpm -qa _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 04 19:37:13 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_END pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:37:13 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGOUT pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:37:13 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGIN pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:37:13 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_START pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:37:13 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: CRYPTO_KEY_USER pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:4d:0e:a0:55:e6:42:83:ce:eb:49:54:e8:e7:6b:94:70:b8:87:2b:24:6c:0c:40:0b:23:42:6b:40:30:1b:d1:22 direction=? spid=4812 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=? res=success' Feb 04 19:37:13 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm python3[4828]: ansible-ansible.legacy.command Invoked with _raw_params=for repofile in $(ls -1 /etc/yum.repos.d/); do echo "# vvvvv $repofile vvvvv vvvvv"; cat "/etc/yum.repos.d/$repofile"; echo "# ^^^^^ ^^^^^ ^^^^^ ^^^^^"; echo ""; done _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 04 19:37:13 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_END pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:37:13 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGOUT pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:37:13 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGIN pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:37:13 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_START pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:37:13 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: CRYPTO_KEY_USER pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:4d:0e:a0:55:e6:42:83:ce:eb:49:54:e8:e7:6b:94:70:b8:87:2b:24:6c:0c:40:0b:23:42:6b:40:30:1b:d1:22 direction=? spid=4839 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=? res=success' Feb 04 19:37:14 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm python3[4855]: ansible-ansible.legacy.command Invoked with _raw_params=yum -q repolist _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 04 19:37:14 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_END pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:37:14 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGOUT pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:37:14 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGIN pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:37:14 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_START pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:37:14 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: CRYPTO_KEY_USER pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:4d:0e:a0:55:e6:42:83:ce:eb:49:54:e8:e7:6b:94:70:b8:87:2b:24:6c:0c:40:0b:23:42:6b:40:30:1b:d1:22 direction=? spid=4857 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=? res=success' Feb 04 19:37:14 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm python3[4873]: ansible-lineinfile Invoked with path=/etc/dnf/dnf.conf regexp=^best= line=best=False state=present backrefs=False create=False backup=False firstmatch=False unsafe_writes=False search_string=None insertafter=None insertbefore=None validate=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Feb 04 19:37:14 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_END pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:37:14 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGOUT pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:37:15 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGIN pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:37:15 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_START pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:37:15 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: CRYPTO_KEY_USER pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:4d:0e:a0:55:e6:42:83:ce:eb:49:54:e8:e7:6b:94:70:b8:87:2b:24:6c:0c:40:0b:23:42:6b:40:30:1b:d1:22 direction=? spid=4874 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=? res=success' Feb 04 19:37:15 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm python3[4890]: ansible-ansible.legacy.command Invoked with _raw_params=dnf clean all --enablerepo=* _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 04 19:37:15 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_END pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:37:15 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGOUT pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:37:15 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGIN pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:37:15 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_START pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:37:15 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: CRYPTO_KEY_USER pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:4d:0e:a0:55:e6:42:83:ce:eb:49:54:e8:e7:6b:94:70:b8:87:2b:24:6c:0c:40:0b:23:42:6b:40:30:1b:d1:22 direction=? spid=4892 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=? res=success' Feb 04 19:37:15 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm python3[4908]: ansible-ansible.legacy.command Invoked with _raw_params=dnf makecache --enablerepo=* _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 04 19:38:26 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_END pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:38:26 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGOUT pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:38:27 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGIN pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=/dev/pts/0 res=success' Feb 04 19:38:27 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_START pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=/dev/pts/0 res=success' Feb 04 19:38:27 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: CRYPTO_KEY_USER pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:4d:0e:a0:55:e6:42:83:ce:eb:49:54:e8:e7:6b:94:70:b8:87:2b:24:6c:0c:40:0b:23:42:6b:40:30:1b:d1:22 direction=? spid=4974 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=? res=success' Feb 04 19:38:27 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_END pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=? terminal=/dev/pts/0 res=success' Feb 04 19:38:27 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGOUT pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=? terminal=/dev/pts/0 res=success' Feb 04 19:38:28 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGIN pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:38:28 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_START pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:38:28 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: CRYPTO_KEY_USER pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:4d:0e:a0:55:e6:42:83:ce:eb:49:54:e8:e7:6b:94:70:b8:87:2b:24:6c:0c:40:0b:23:42:6b:40:30:1b:d1:22 direction=? spid=4994 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=? res=success' Feb 04 19:38:28 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm python3[5010]: ansible-ansible.legacy.setup Invoked with gather_subset=['all'] gather_timeout=10 filter=[] fact_path=/etc/ansible/facts.d Feb 04 19:38:29 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_END pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:38:29 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGOUT pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:38:29 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGIN pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:38:29 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_START pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:38:29 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: CRYPTO_KEY_USER pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:4d:0e:a0:55:e6:42:83:ce:eb:49:54:e8:e7:6b:94:70:b8:87:2b:24:6c:0c:40:0b:23:42:6b:40:30:1b:d1:22 direction=? spid=5035 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=? res=success' Feb 04 19:38:29 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm python3[5051]: ansible-stat Invoked with path=/run/ostree-booted follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Feb 04 19:38:29 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_END pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:38:29 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGOUT pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:38:29 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGIN pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:38:29 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_START pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:38:29 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: CRYPTO_KEY_USER pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:4d:0e:a0:55:e6:42:83:ce:eb:49:54:e8:e7:6b:94:70:b8:87:2b:24:6c:0c:40:0b:23:42:6b:40:30:1b:d1:22 direction=? spid=5052 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=? res=success' Feb 04 19:38:29 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm python3[5068]: ansible-ansible.legacy.dnf Invoked with name=['python3-pyasn1', 'python3-cryptography', 'python3-dbus'] state=present allow_downgrade=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 allowerasing=False nobest=False use_backend=auto conf_file=None disable_excludes=None download_dir=None list=None releasever=None Feb 04 19:38:35 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_END pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:38:35 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGOUT pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:38:35 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGIN pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:38:35 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_START pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:38:35 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: CRYPTO_KEY_USER pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:4d:0e:a0:55:e6:42:83:ce:eb:49:54:e8:e7:6b:94:70:b8:87:2b:24:6c:0c:40:0b:23:42:6b:40:30:1b:d1:22 direction=? spid=5082 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=? res=success' Feb 04 19:38:36 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm python3[5098]: ansible-ansible.legacy.dnf Invoked with name=['certmonger', 'python3-packaging'] state=present allow_downgrade=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 allowerasing=False nobest=False use_backend=auto conf_file=None disable_excludes=None download_dir=None list=None releasever=None Feb 04 19:38:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm dbus-broker-launch[585]: Noticed file-system modification, trigger reload. ░░ Subject: A configuration directory was written to ░░ Defined-By: dbus-broker ░░ Support: https://groups.google.com/forum/#!forum/bus1-devel ░░ ░░ A write was detected to one of the directories containing D-Bus configuration ░░ files, triggering a configuration reload. ░░ ░░ This functionality exists for backwards compatibility to pick up changes to ░░ D-Bus configuration without an explicit reolad request. Typically when ░░ installing or removing third-party software causes D-Bus configuration files ░░ to be added or removed. ░░ ░░ It is worth noting that this may cause partial configuration to be loaded in ░░ case dispatching this notification races with the writing of the configuration ░░ files. However, a future notification will then cause the configuration to be ░░ reladed again. Feb 04 19:38:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm dbus-broker-launch[585]: Noticed file-system modification, trigger reload. ░░ Subject: A configuration directory was written to ░░ Defined-By: dbus-broker ░░ Support: https://groups.google.com/forum/#!forum/bus1-devel ░░ ░░ A write was detected to one of the directories containing D-Bus configuration ░░ files, triggering a configuration reload. ░░ ░░ This functionality exists for backwards compatibility to pick up changes to ░░ D-Bus configuration without an explicit reolad request. Typically when ░░ installing or removing third-party software causes D-Bus configuration files ░░ to be added or removed. ░░ ░░ It is worth noting that this may cause partial configuration to be loaded in ░░ case dispatching this notification races with the writing of the configuration ░░ files. However, a future notification will then cause the configuration to be ░░ reladed again. Feb 04 19:38:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm dbus-broker-launch[585]: Noticed file-system modification, trigger reload. ░░ Subject: A configuration directory was written to ░░ Defined-By: dbus-broker ░░ Support: https://groups.google.com/forum/#!forum/bus1-devel ░░ ░░ A write was detected to one of the directories containing D-Bus configuration ░░ files, triggering a configuration reload. ░░ ░░ This functionality exists for backwards compatibility to pick up changes to ░░ D-Bus configuration without an explicit reolad request. Typically when ░░ installing or removing third-party software causes D-Bus configuration files ░░ to be added or removed. ░░ ░░ It is worth noting that this may cause partial configuration to be loaded in ░░ case dispatching this notification races with the writing of the configuration ░░ files. However, a future notification will then cause the configuration to be ░░ reladed again. Feb 04 19:38:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm systemd[1]: Reloading requested from client PID 5110 ('systemctl') (unit session-6.scope)... Feb 04 19:38:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm systemd[1]: Reloading... Feb 04 19:38:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm systemd[1]: Reloading finished in 410 ms. Feb 04 19:38:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=93 op=LOAD Feb 04 19:38:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=69 op=UNLOAD Feb 04 19:38:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=94 op=LOAD Feb 04 19:38:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=95 op=LOAD Feb 04 19:38:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=70 op=UNLOAD Feb 04 19:38:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=71 op=UNLOAD Feb 04 19:38:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=96 op=LOAD Feb 04 19:38:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=97 op=LOAD Feb 04 19:38:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=72 op=UNLOAD Feb 04 19:38:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=73 op=UNLOAD Feb 04 19:38:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=98 op=LOAD Feb 04 19:38:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=74 op=UNLOAD Feb 04 19:38:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=99 op=LOAD Feb 04 19:38:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=75 op=UNLOAD Feb 04 19:38:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=100 op=LOAD Feb 04 19:38:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=101 op=LOAD Feb 04 19:38:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=76 op=UNLOAD Feb 04 19:38:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=77 op=UNLOAD Feb 04 19:38:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=102 op=LOAD Feb 04 19:38:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=78 op=UNLOAD Feb 04 19:38:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=103 op=LOAD Feb 04 19:38:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=104 op=LOAD Feb 04 19:38:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=79 op=UNLOAD Feb 04 19:38:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=80 op=UNLOAD Feb 04 19:38:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=105 op=LOAD Feb 04 19:38:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=81 op=UNLOAD Feb 04 19:38:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=106 op=LOAD Feb 04 19:38:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=82 op=UNLOAD Feb 04 19:38:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=107 op=LOAD Feb 04 19:38:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=83 op=UNLOAD Feb 04 19:38:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=108 op=LOAD Feb 04 19:38:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=84 op=UNLOAD Feb 04 19:38:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=109 op=LOAD Feb 04 19:38:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=110 op=LOAD Feb 04 19:38:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=85 op=UNLOAD Feb 04 19:38:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=86 op=UNLOAD Feb 04 19:38:39 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm systemd[1]: Started run-r924c57ac7cdc4bc5959577c9c93bdd23.service - /usr/bin/systemctl start man-db-cache-update. ░░ Subject: A start job for unit run-r924c57ac7cdc4bc5959577c9c93bdd23.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ A start job for unit run-r924c57ac7cdc4bc5959577c9c93bdd23.service has finished successfully. ░░ ░░ The job identifier is 2005. Feb 04 19:38:39 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:init_t:s0 msg='unit=run-r924c57ac7cdc4bc5959577c9c93bdd23 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 04 19:38:39 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm systemd[1]: Starting man-db-cache-update.service... ░░ Subject: A start job for unit man-db-cache-update.service has begun execution ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ A start job for unit man-db-cache-update.service has begun execution. ░░ ░░ The job identifier is 2090. Feb 04 19:38:39 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm systemd[1]: Reloading requested from client PID 5163 ('systemctl') (unit session-6.scope)... Feb 04 19:38:39 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm systemd[1]: Reloading... Feb 04 19:38:39 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm systemd[1]: Reloading finished in 597 ms. Feb 04 19:38:40 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=111 op=LOAD Feb 04 19:38:40 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=93 op=UNLOAD Feb 04 19:38:40 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=112 op=LOAD Feb 04 19:38:40 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=113 op=LOAD Feb 04 19:38:40 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=94 op=UNLOAD Feb 04 19:38:40 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=95 op=UNLOAD Feb 04 19:38:40 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=114 op=LOAD Feb 04 19:38:40 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=115 op=LOAD Feb 04 19:38:40 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=96 op=UNLOAD Feb 04 19:38:40 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=97 op=UNLOAD Feb 04 19:38:40 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=116 op=LOAD Feb 04 19:38:40 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=98 op=UNLOAD Feb 04 19:38:40 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=117 op=LOAD Feb 04 19:38:40 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=99 op=UNLOAD Feb 04 19:38:40 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=118 op=LOAD Feb 04 19:38:40 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=119 op=LOAD Feb 04 19:38:40 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=100 op=UNLOAD Feb 04 19:38:40 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=101 op=UNLOAD Feb 04 19:38:40 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=120 op=LOAD Feb 04 19:38:40 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=102 op=UNLOAD Feb 04 19:38:40 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=121 op=LOAD Feb 04 19:38:40 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=122 op=LOAD Feb 04 19:38:40 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=103 op=UNLOAD Feb 04 19:38:40 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=104 op=UNLOAD Feb 04 19:38:40 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=123 op=LOAD Feb 04 19:38:40 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=105 op=UNLOAD Feb 04 19:38:40 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=124 op=LOAD Feb 04 19:38:40 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=106 op=UNLOAD Feb 04 19:38:40 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=125 op=LOAD Feb 04 19:38:40 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=107 op=UNLOAD Feb 04 19:38:40 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=126 op=LOAD Feb 04 19:38:40 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=108 op=UNLOAD Feb 04 19:38:40 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=127 op=LOAD Feb 04 19:38:40 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=128 op=LOAD Feb 04 19:38:40 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=109 op=UNLOAD Feb 04 19:38:40 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=110 op=UNLOAD Feb 04 19:38:40 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm systemd[1]: Queuing reload/restart jobs for marked units… Feb 04 19:38:40 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_END pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:38:40 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGOUT pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:38:40 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGIN pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:38:40 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_START pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:38:40 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: CRYPTO_KEY_USER pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:4d:0e:a0:55:e6:42:83:ce:eb:49:54:e8:e7:6b:94:70:b8:87:2b:24:6c:0c:40:0b:23:42:6b:40:30:1b:d1:22 direction=? spid=5215 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=? res=success' Feb 04 19:38:40 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm systemd[1]: man-db-cache-update.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ The unit man-db-cache-update.service has successfully entered the 'dead' state. Feb 04 19:38:40 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm systemd[1]: Finished man-db-cache-update.service. ░░ Subject: A start job for unit man-db-cache-update.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ A start job for unit man-db-cache-update.service has finished successfully. ░░ ░░ The job identifier is 2090. Feb 04 19:38:40 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:init_t:s0 msg='unit=man-db-cache-update comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 04 19:38:40 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:init_t:s0 msg='unit=man-db-cache-update comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 04 19:38:40 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm python3[5233]: ansible-file Invoked with name=/etc/certmonger//pre-scripts owner=root group=root mode=0700 state=directory path=/etc/certmonger//pre-scripts recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Feb 04 19:38:40 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm systemd[1]: run-r924c57ac7cdc4bc5959577c9c93bdd23.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ The unit run-r924c57ac7cdc4bc5959577c9c93bdd23.service has successfully entered the 'dead' state. Feb 04 19:38:40 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:init_t:s0 msg='unit=run-r924c57ac7cdc4bc5959577c9c93bdd23 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 04 19:38:40 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_END pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:38:40 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGOUT pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:38:40 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGIN pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:38:40 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_START pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:38:40 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: CRYPTO_KEY_USER pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:4d:0e:a0:55:e6:42:83:ce:eb:49:54:e8:e7:6b:94:70:b8:87:2b:24:6c:0c:40:0b:23:42:6b:40:30:1b:d1:22 direction=? spid=5237 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=? res=success' Feb 04 19:38:41 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm python3[5253]: ansible-file Invoked with name=/etc/certmonger//post-scripts owner=root group=root mode=0700 state=directory path=/etc/certmonger//post-scripts recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Feb 04 19:38:41 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_END pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:38:41 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGOUT pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:38:41 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGIN pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:38:41 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_START pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:38:41 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: CRYPTO_KEY_USER pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:4d:0e:a0:55:e6:42:83:ce:eb:49:54:e8:e7:6b:94:70:b8:87:2b:24:6c:0c:40:0b:23:42:6b:40:30:1b:d1:22 direction=? spid=5254 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=? res=success' Feb 04 19:38:41 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm python3[5270]: ansible-ansible.legacy.systemd Invoked with name=certmonger state=started enabled=True daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Feb 04 19:38:41 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm systemd[1]: Reloading requested from client PID 5273 ('systemctl') (unit session-6.scope)... Feb 04 19:38:41 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm systemd[1]: Reloading... Feb 04 19:38:42 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm systemd[1]: Reloading finished in 406 ms. Feb 04 19:38:42 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=129 op=LOAD Feb 04 19:38:42 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=111 op=UNLOAD Feb 04 19:38:42 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=130 op=LOAD Feb 04 19:38:42 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=131 op=LOAD Feb 04 19:38:42 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=112 op=UNLOAD Feb 04 19:38:42 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=113 op=UNLOAD Feb 04 19:38:42 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=132 op=LOAD Feb 04 19:38:42 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=133 op=LOAD Feb 04 19:38:42 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=114 op=UNLOAD Feb 04 19:38:42 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=115 op=UNLOAD Feb 04 19:38:42 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=134 op=LOAD Feb 04 19:38:42 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=116 op=UNLOAD Feb 04 19:38:42 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=135 op=LOAD Feb 04 19:38:42 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=117 op=UNLOAD Feb 04 19:38:42 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=136 op=LOAD Feb 04 19:38:42 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=137 op=LOAD Feb 04 19:38:42 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=118 op=UNLOAD Feb 04 19:38:42 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=119 op=UNLOAD Feb 04 19:38:42 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=138 op=LOAD Feb 04 19:38:42 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=120 op=UNLOAD Feb 04 19:38:42 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=139 op=LOAD Feb 04 19:38:42 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=140 op=LOAD Feb 04 19:38:42 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=121 op=UNLOAD Feb 04 19:38:42 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=122 op=UNLOAD Feb 04 19:38:42 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=141 op=LOAD Feb 04 19:38:42 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=123 op=UNLOAD Feb 04 19:38:42 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=142 op=LOAD Feb 04 19:38:42 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=124 op=UNLOAD Feb 04 19:38:42 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=143 op=LOAD Feb 04 19:38:42 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=125 op=UNLOAD Feb 04 19:38:42 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=144 op=LOAD Feb 04 19:38:42 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=126 op=UNLOAD Feb 04 19:38:42 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=145 op=LOAD Feb 04 19:38:42 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=146 op=LOAD Feb 04 19:38:42 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=127 op=UNLOAD Feb 04 19:38:42 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=128 op=UNLOAD Feb 04 19:38:42 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm systemd[1]: Starting certmonger.service - Certificate monitoring and PKI enrollment... ░░ Subject: A start job for unit certmonger.service has begun execution ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ A start job for unit certmonger.service has begun execution. ░░ ░░ The job identifier is 2175. Feb 04 19:38:42 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm (rtmonger)[5317]: certmonger.service: Referenced but unset environment variable evaluates to an empty string: OPTS Feb 04 19:38:42 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm systemd[1]: Started certmonger.service - Certificate monitoring and PKI enrollment. ░░ Subject: A start job for unit certmonger.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ A start job for unit certmonger.service has finished successfully. ░░ ░░ The job identifier is 2175. Feb 04 19:38:42 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:init_t:s0 msg='unit=certmonger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 04 19:38:42 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_END pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:38:42 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGOUT pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:38:42 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGIN pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:38:42 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_START pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:38:42 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: CRYPTO_KEY_USER pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:4d:0e:a0:55:e6:42:83:ce:eb:49:54:e8:e7:6b:94:70:b8:87:2b:24:6c:0c:40:0b:23:42:6b:40:30:1b:d1:22 direction=? spid=5342 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=? res=success' Feb 04 19:38:42 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm python3[5358]: ansible-fedora.linux_system_roles.certificate_request Invoked with name=quadlet_demo dns=['localhost'] directory=/etc/pki/tls wait=True ca=self-sign __header=# # Ansible managed # # system_role:certificate provider_config_directory=/etc/certmonger provider=certmonger key_usage=['digitalSignature', 'keyEncipherment'] extended_key_usage=['id-kp-serverAuth', 'id-kp-clientAuth'] auto_renew=True ip=None email=None common_name=None country=None state=None locality=None organization=None organizational_unit=None contact_email=None key_size=None owner=None group=None mode=None principal=None run_before=None run_after=None Feb 04 19:38:42 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm certmonger[5317]: 2024-02-04 19:38:42 [5317] Wrote to /var/lib/certmonger/requests/20240204193842 Feb 04 19:38:42 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm certmonger[5317]: 2024-02-04 19:38:42 [5317] Wrote to /var/lib/certmonger/requests/20240204193842 Feb 04 19:38:42 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm certmonger[5317]: 2024-02-04 19:38:42 [5317] Wrote to /var/lib/certmonger/requests/20240204193842 Feb 04 19:38:42 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm certmonger[5317]: 2024-02-04 19:38:42 [5317] Wrote to /var/lib/certmonger/requests/20240204193842 Feb 04 19:38:42 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm certmonger[5317]: 2024-02-04 19:38:42 [5317] Wrote to /var/lib/certmonger/requests/20240204193842 Feb 04 19:38:42 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm certmonger[5317]: 2024-02-04 19:38:42 [5317] Wrote to /var/lib/certmonger/requests/20240204193842 Feb 04 19:38:42 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm certmonger[5317]: 2024-02-04 19:38:42 [5317] Wrote to /var/lib/certmonger/requests/20240204193842 Feb 04 19:38:42 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm certmonger[5317]: 2024-02-04 19:38:42 [5317] Wrote to /var/lib/certmonger/requests/20240204193842 Feb 04 19:38:42 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm certmonger[5317]: 2024-02-04 19:38:42 [5317] Wrote to /var/lib/certmonger/requests/20240204193842 Feb 04 19:38:42 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm certmonger[5317]: 2024-02-04 19:38:42 [5317] Wrote to /var/lib/certmonger/requests/20240204193842 Feb 04 19:38:42 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm certmonger[5317]: 2024-02-04 19:38:42 [5317] Wrote to /var/lib/certmonger/requests/20240204193842 Feb 04 19:38:42 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm certmonger[5317]: 2024-02-04 19:38:42 [5317] Wrote to /var/lib/certmonger/requests/20240204193842 Feb 04 19:38:42 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm certmonger[5317]: 2024-02-04 19:38:42 [5317] Wrote to /var/lib/certmonger/requests/20240204193842 Feb 04 19:38:42 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm certmonger[5317]: 2024-02-04 19:38:42 [5317] Wrote to /var/lib/certmonger/requests/20240204193842 Feb 04 19:38:42 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm certmonger[5317]: 2024-02-04 19:38:42 [5317] Wrote to /var/lib/certmonger/requests/20240204193842 Feb 04 19:38:42 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm certmonger[5317]: 2024-02-04 19:38:42 [5317] Wrote to /var/lib/certmonger/requests/20240204193842 Feb 04 19:38:42 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm certmonger[5317]: 2024-02-04 19:38:42 [5317] Wrote to /var/lib/certmonger/requests/20240204193842 Feb 04 19:38:42 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm certmonger[5317]: 2024-02-04 19:38:42 [5317] Wrote to /var/lib/certmonger/requests/20240204193842 Feb 04 19:38:42 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm certmonger[5317]: 2024-02-04 19:38:42 [5317] Wrote to /var/lib/certmonger/requests/20240204193842 Feb 04 19:38:42 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm certmonger[5317]: 2024-02-04 19:38:42 [5317] Wrote to /var/lib/certmonger/requests/20240204193842 Feb 04 19:38:42 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm certmonger[5317]: 2024-02-04 19:38:42 [5317] Wrote to /var/lib/certmonger/requests/20240204193842 Feb 04 19:38:42 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm certmonger[5317]: 2024-02-04 19:38:42 [5317] Wrote to /var/lib/certmonger/requests/20240204193842 Feb 04 19:38:42 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm certmonger[5317]: 2024-02-04 19:38:42 [5317] Wrote to /var/lib/certmonger/requests/20240204193842 Feb 04 19:38:42 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm certmonger[5317]: 2024-02-04 19:38:42 [5317] Wrote to /var/lib/certmonger/requests/20240204193842 Feb 04 19:38:42 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm certmonger[5317]: 2024-02-04 19:38:42 [5317] Wrote to /var/lib/certmonger/requests/20240204193842 Feb 04 19:38:42 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm certmonger[5317]: 2024-02-04 19:38:42 [5317] Wrote to /var/lib/certmonger/requests/20240204193842 Feb 04 19:38:42 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm certmonger[5317]: 2024-02-04 19:38:42 [5317] Wrote to /var/lib/certmonger/requests/20240204193842 Feb 04 19:38:43 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm certmonger[5317]: 2024-02-04 19:38:43 [5317] Wrote to /var/lib/certmonger/requests/20240204193842 Feb 04 19:38:43 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm certmonger[5317]: 2024-02-04 19:38:43 [5317] Wrote to /var/lib/certmonger/requests/20240204193842 Feb 04 19:38:43 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm certmonger[5376]: Certificate in file "/etc/pki/tls/certs/quadlet_demo.crt" issued by CA and saved. Feb 04 19:38:43 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm certmonger[5317]: 2024-02-04 19:38:43 [5317] Wrote to /var/lib/certmonger/requests/20240204193842 Feb 04 19:38:43 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_END pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:38:43 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGOUT pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:38:43 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGIN pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:38:43 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_START pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:38:43 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: CRYPTO_KEY_USER pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:4d:0e:a0:55:e6:42:83:ce:eb:49:54:e8:e7:6b:94:70:b8:87:2b:24:6c:0c:40:0b:23:42:6b:40:30:1b:d1:22 direction=? spid=5377 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=? res=success' Feb 04 19:38:43 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm python3[5393]: ansible-slurp Invoked with path=/etc/pki/tls/certs/quadlet_demo.crt src=/etc/pki/tls/certs/quadlet_demo.crt Feb 04 19:38:43 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_END pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:38:43 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGOUT pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:38:43 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGIN pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:38:43 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_START pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:38:43 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: CRYPTO_KEY_USER pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:4d:0e:a0:55:e6:42:83:ce:eb:49:54:e8:e7:6b:94:70:b8:87:2b:24:6c:0c:40:0b:23:42:6b:40:30:1b:d1:22 direction=? spid=5394 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=? res=success' Feb 04 19:38:43 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm python3[5410]: ansible-slurp Invoked with path=/etc/pki/tls/private/quadlet_demo.key src=/etc/pki/tls/private/quadlet_demo.key Feb 04 19:38:43 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_END pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:38:43 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGOUT pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:38:43 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGIN pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:38:43 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_START pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:38:43 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: CRYPTO_KEY_USER pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:4d:0e:a0:55:e6:42:83:ce:eb:49:54:e8:e7:6b:94:70:b8:87:2b:24:6c:0c:40:0b:23:42:6b:40:30:1b:d1:22 direction=? spid=5411 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=? res=success' Feb 04 19:38:43 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm python3[5427]: ansible-slurp Invoked with path=/etc/pki/tls/certs/quadlet_demo.crt src=/etc/pki/tls/certs/quadlet_demo.crt Feb 04 19:38:43 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_END pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:38:43 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGOUT pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:38:43 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGIN pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:38:43 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_START pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:38:43 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: CRYPTO_KEY_USER pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:4d:0e:a0:55:e6:42:83:ce:eb:49:54:e8:e7:6b:94:70:b8:87:2b:24:6c:0c:40:0b:23:42:6b:40:30:1b:d1:22 direction=? spid=5428 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=? res=success' Feb 04 19:38:44 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm python3[5444]: ansible-ansible.legacy.command Invoked with _raw_params=getcert stop-tracking -f /etc/pki/tls/certs/quadlet_demo.crt _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 04 19:38:44 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm certmonger[5317]: 2024-02-04 19:38:44 [5317] Wrote to /var/lib/certmonger/requests/20240204193842 Feb 04 19:38:44 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_END pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:38:44 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGOUT pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:38:44 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGIN pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:38:44 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_START pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:38:44 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: CRYPTO_KEY_USER pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:4d:0e:a0:55:e6:42:83:ce:eb:49:54:e8:e7:6b:94:70:b8:87:2b:24:6c:0c:40:0b:23:42:6b:40:30:1b:d1:22 direction=? spid=5446 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=? res=success' Feb 04 19:38:44 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm python3[5462]: ansible-file Invoked with path=/etc/pki/tls/certs/quadlet_demo.crt state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Feb 04 19:38:44 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_END pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:38:44 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGOUT pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:38:44 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGIN pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:38:44 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_START pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:38:44 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: CRYPTO_KEY_USER pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:4d:0e:a0:55:e6:42:83:ce:eb:49:54:e8:e7:6b:94:70:b8:87:2b:24:6c:0c:40:0b:23:42:6b:40:30:1b:d1:22 direction=? spid=5463 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=? res=success' Feb 04 19:38:44 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm python3[5479]: ansible-file Invoked with path=/etc/pki/tls/private/quadlet_demo.key state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Feb 04 19:38:44 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_END pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:38:44 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGOUT pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:38:44 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGIN pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:38:44 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_START pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:38:44 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: CRYPTO_KEY_USER pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:4d:0e:a0:55:e6:42:83:ce:eb:49:54:e8:e7:6b:94:70:b8:87:2b:24:6c:0c:40:0b:23:42:6b:40:30:1b:d1:22 direction=? spid=5480 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=? res=success' Feb 04 19:38:44 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm python3[5496]: ansible-file Invoked with path=/etc/pki/tls/certs/quadlet_demo.crt state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Feb 04 19:38:44 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_END pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:38:44 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGOUT pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:38:45 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGIN pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:38:45 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_START pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:38:45 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: CRYPTO_KEY_USER pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:4d:0e:a0:55:e6:42:83:ce:eb:49:54:e8:e7:6b:94:70:b8:87:2b:24:6c:0c:40:0b:23:42:6b:40:30:1b:d1:22 direction=? spid=5497 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=? res=success' Feb 04 19:38:45 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm python3[5513]: ansible-stat Invoked with path=/run/ostree-booted follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Feb 04 19:38:45 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_END pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:38:45 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGOUT pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:38:45 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGIN pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:38:45 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_START pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:38:45 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: CRYPTO_KEY_USER pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:4d:0e:a0:55:e6:42:83:ce:eb:49:54:e8:e7:6b:94:70:b8:87:2b:24:6c:0c:40:0b:23:42:6b:40:30:1b:d1:22 direction=? spid=5514 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=? res=success' Feb 04 19:38:46 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_END pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:38:46 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGOUT pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:38:46 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGIN pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:38:46 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_START pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:38:46 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: CRYPTO_KEY_USER pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:4d:0e:a0:55:e6:42:83:ce:eb:49:54:e8:e7:6b:94:70:b8:87:2b:24:6c:0c:40:0b:23:42:6b:40:30:1b:d1:22 direction=? spid=5531 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=? res=success' Feb 04 19:38:46 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm python3[5547]: ansible-ansible.legacy.dnf Invoked with name=['podman', 'shadow-utils-subid'] state=present allow_downgrade=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 allowerasing=False nobest=False use_backend=auto conf_file=None disable_excludes=None download_dir=None list=None releasever=None Feb 04 19:39:05 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm kernel: SELinux: Converting 398 SID table entries... Feb 04 19:39:05 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm kernel: SELinux: policy capability network_peer_controls=1 Feb 04 19:39:05 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm kernel: SELinux: policy capability open_perms=1 Feb 04 19:39:05 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm kernel: SELinux: policy capability extended_socket_class=1 Feb 04 19:39:05 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm kernel: SELinux: policy capability always_check_network=0 Feb 04 19:39:05 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm kernel: SELinux: policy capability cgroup_seclabel=1 Feb 04 19:39:05 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm kernel: SELinux: policy capability nnp_nosuid_transition=1 Feb 04 19:39:05 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm kernel: SELinux: policy capability genfs_seclabel_symlinks=1 Feb 04 19:39:05 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm kernel: SELinux: policy capability ioctl_skip_cloexec=0 Feb 04 19:39:05 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: MAC_POLICY_LOAD auid=0 ses=6 lsm=selinux res=1 Feb 04 19:39:14 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm kernel: SELinux: Converting 398 SID table entries... Feb 04 19:39:14 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm kernel: SELinux: policy capability network_peer_controls=1 Feb 04 19:39:14 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm kernel: SELinux: policy capability open_perms=1 Feb 04 19:39:14 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm kernel: SELinux: policy capability extended_socket_class=1 Feb 04 19:39:14 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm kernel: SELinux: policy capability always_check_network=0 Feb 04 19:39:14 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm kernel: SELinux: policy capability cgroup_seclabel=1 Feb 04 19:39:14 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm kernel: SELinux: policy capability nnp_nosuid_transition=1 Feb 04 19:39:14 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm kernel: SELinux: policy capability genfs_seclabel_symlinks=1 Feb 04 19:39:14 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm kernel: SELinux: policy capability ioctl_skip_cloexec=0 Feb 04 19:39:14 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: MAC_POLICY_LOAD auid=0 ses=6 lsm=selinux res=1 Feb 04 19:39:15 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[586]: USER_MAC_POLICY_LOAD pid=586 uid=81 auid=4294967295 ses=4294967295 subj=system_u:system_r:system_dbusd_t:s0-s0:c0.c1023 msg='avc: op=load_policy lsm=selinux seqno=4 res=1 exe="/usr/bin/dbus-broker" sauid=81 hostname=? addr=? terminal=?' Feb 04 19:39:17 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: MAC_CONFIG_CHANGE bool=virt_use_nfs val=1 old_val=0 auid=0 ses=6 Feb 04 19:39:20 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm setsebool[5660]: The virt_use_nfs policy boolean was changed to 1 by root Feb 04 19:39:20 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm setsebool[5660]: The virt_sandbox_use_all_caps policy boolean was changed to 1 by root Feb 04 19:39:29 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm kernel: SELinux: Converting 400 SID table entries... Feb 04 19:39:29 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm kernel: SELinux: policy capability network_peer_controls=1 Feb 04 19:39:29 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm kernel: SELinux: policy capability open_perms=1 Feb 04 19:39:29 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm kernel: SELinux: policy capability extended_socket_class=1 Feb 04 19:39:29 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm kernel: SELinux: policy capability always_check_network=0 Feb 04 19:39:29 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm kernel: SELinux: policy capability cgroup_seclabel=1 Feb 04 19:39:29 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm kernel: SELinux: policy capability nnp_nosuid_transition=1 Feb 04 19:39:29 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm kernel: SELinux: policy capability genfs_seclabel_symlinks=1 Feb 04 19:39:29 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm kernel: SELinux: policy capability ioctl_skip_cloexec=0 Feb 04 19:39:29 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: MAC_POLICY_LOAD auid=0 ses=6 lsm=selinux res=1 Feb 04 19:40:04 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm systemd[1]: proc-sys-fs-binfmt_misc.automount: Got automount request for /proc/sys/fs/binfmt_misc, triggered by 6317 (systemd-binfmt) Feb 04 19:40:04 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[586]: USER_MAC_POLICY_LOAD pid=586 uid=81 auid=4294967295 ses=4294967295 subj=system_u:system_r:system_dbusd_t:s0-s0:c0.c1023 msg='avc: op=load_policy lsm=selinux seqno=6 res=1 exe="/usr/bin/dbus-broker" sauid=81 hostname=? addr=? terminal=?' Feb 04 19:40:04 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm systemd[1]: Mounting proc-sys-fs-binfmt_misc.mount - Arbitrary Executable File Formats File System... ░░ Subject: A start job for unit proc-sys-fs-binfmt_misc.mount has begun execution ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ A start job for unit proc-sys-fs-binfmt_misc.mount has begun execution. ░░ ░░ The job identifier is 2280. Feb 04 19:40:04 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm systemd[1]: Mounted proc-sys-fs-binfmt_misc.mount - Arbitrary Executable File Formats File System. ░░ Subject: A start job for unit proc-sys-fs-binfmt_misc.mount has finished successfully ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ A start job for unit proc-sys-fs-binfmt_misc.mount has finished successfully. ░░ ░░ The job identifier is 2280. Feb 04 19:40:04 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:init_t:s0 msg='unit=run-r7ea71933f17140e28477a3a804dbd288 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 04 19:40:04 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm systemd[1]: Started run-r7ea71933f17140e28477a3a804dbd288.service - /usr/bin/systemctl start man-db-cache-update. ░░ Subject: A start job for unit run-r7ea71933f17140e28477a3a804dbd288.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ A start job for unit run-r7ea71933f17140e28477a3a804dbd288.service has finished successfully. ░░ ░░ The job identifier is 2284. Feb 04 19:40:04 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm systemd[1]: Starting man-db-cache-update.service... ░░ Subject: A start job for unit man-db-cache-update.service has begun execution ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ A start job for unit man-db-cache-update.service has begun execution. ░░ ░░ The job identifier is 2369. Feb 04 19:40:04 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm systemd[1]: Reloading requested from client PID 6329 ('systemctl') (unit session-6.scope)... Feb 04 19:40:04 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm systemd[1]: Reloading... Feb 04 19:40:05 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm systemd[1]: Reloading finished in 641 ms. Feb 04 19:40:05 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=147 op=LOAD Feb 04 19:40:05 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=129 op=UNLOAD Feb 04 19:40:05 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=148 op=LOAD Feb 04 19:40:05 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=149 op=LOAD Feb 04 19:40:05 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=130 op=UNLOAD Feb 04 19:40:05 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=131 op=UNLOAD Feb 04 19:40:05 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=150 op=LOAD Feb 04 19:40:05 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=151 op=LOAD Feb 04 19:40:05 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=132 op=UNLOAD Feb 04 19:40:05 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=133 op=UNLOAD Feb 04 19:40:05 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=152 op=LOAD Feb 04 19:40:05 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=134 op=UNLOAD Feb 04 19:40:05 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=153 op=LOAD Feb 04 19:40:05 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=135 op=UNLOAD Feb 04 19:40:05 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=154 op=LOAD Feb 04 19:40:05 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=155 op=LOAD Feb 04 19:40:05 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=136 op=UNLOAD Feb 04 19:40:05 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=137 op=UNLOAD Feb 04 19:40:05 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=156 op=LOAD Feb 04 19:40:05 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=138 op=UNLOAD Feb 04 19:40:05 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=157 op=LOAD Feb 04 19:40:05 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=158 op=LOAD Feb 04 19:40:05 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=139 op=UNLOAD Feb 04 19:40:05 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=140 op=UNLOAD Feb 04 19:40:05 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=159 op=LOAD Feb 04 19:40:05 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=141 op=UNLOAD Feb 04 19:40:05 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=160 op=LOAD Feb 04 19:40:05 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=142 op=UNLOAD Feb 04 19:40:05 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=161 op=LOAD Feb 04 19:40:05 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=143 op=UNLOAD Feb 04 19:40:05 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=162 op=LOAD Feb 04 19:40:05 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=144 op=UNLOAD Feb 04 19:40:05 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=163 op=LOAD Feb 04 19:40:05 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=164 op=LOAD Feb 04 19:40:05 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=145 op=UNLOAD Feb 04 19:40:05 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=146 op=UNLOAD Feb 04 19:40:05 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm systemd[1]: Starting dnf-makecache.service - dnf makecache... ░░ Subject: A start job for unit dnf-makecache.service has begun execution ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ A start job for unit dnf-makecache.service has begun execution. ░░ ░░ The job identifier is 2454. Feb 04 19:40:05 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm systemd[1]: Queuing reload/restart jobs for marked units… Feb 04 19:40:05 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm systemd[886]: Created slice session.slice - User Core Session Slice. ░░ Subject: A start job for unit UNIT has finished successfully ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ A start job for unit UNIT has finished successfully. ░░ ░░ The job identifier is 25. Feb 04 19:40:05 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm systemd[886]: Starting dbus-broker.service - D-Bus User Message Bus... ░░ Subject: A start job for unit UNIT has begun execution ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ A start job for unit UNIT has begun execution. ░░ ░░ The job identifier is 20. Feb 04 19:40:05 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm dbus-broker-launch[6394]: Policy to allow eavesdropping in /usr/share/dbus-1/session.conf +31: Eavesdropping is deprecated and ignored Feb 04 19:40:05 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm dbus-broker-launch[6394]: Policy to allow eavesdropping in /usr/share/dbus-1/session.conf +33: Eavesdropping is deprecated and ignored Feb 04 19:40:05 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm systemd[886]: Started dbus-broker.service - D-Bus User Message Bus. ░░ Subject: A start job for unit UNIT has finished successfully ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ A start job for unit UNIT has finished successfully. ░░ ░░ The job identifier is 20. Feb 04 19:40:05 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm dbus-broker-launch[6394]: Ready Feb 04 19:40:05 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm systemd[886]: Reloading requested from client PID 6393 ('systemctl')... Feb 04 19:40:05 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm systemd[886]: Reloading... Feb 04 19:40:06 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm systemd[886]: Reloading finished in 183 ms. Feb 04 19:40:06 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm systemd[886]: Queuing reload/restart jobs for marked units… Feb 04 19:40:06 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm dnf[6387]: Failed determining last makecache time. Feb 04 19:40:06 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm systemd[1]: man-db-cache-update.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ The unit man-db-cache-update.service has successfully entered the 'dead' state. Feb 04 19:40:06 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:init_t:s0 msg='unit=man-db-cache-update comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 04 19:40:06 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:init_t:s0 msg='unit=man-db-cache-update comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 04 19:40:06 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm systemd[1]: Finished man-db-cache-update.service. ░░ Subject: A start job for unit man-db-cache-update.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ A start job for unit man-db-cache-update.service has finished successfully. ░░ ░░ The job identifier is 2369. Feb 04 19:40:06 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm systemd[1]: man-db-cache-update.service: Consumed 1.123s CPU time. ░░ Subject: Resources consumed by unit runtime ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ The unit man-db-cache-update.service completed and consumed the indicated resources. Feb 04 19:40:06 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm systemd[1]: run-r7ea71933f17140e28477a3a804dbd288.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ The unit run-r7ea71933f17140e28477a3a804dbd288.service has successfully entered the 'dead' state. Feb 04 19:40:06 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_END pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:06 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGOUT pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:06 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:init_t:s0 msg='unit=run-r7ea71933f17140e28477a3a804dbd288 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 04 19:40:06 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGIN pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:06 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_START pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:06 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: CRYPTO_KEY_USER pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:4d:0e:a0:55:e6:42:83:ce:eb:49:54:e8:e7:6b:94:70:b8:87:2b:24:6c:0c:40:0b:23:42:6b:40:30:1b:d1:22 direction=? spid=6415 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=? res=success' Feb 04 19:40:07 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm dnf[6387]: Beaker Client - Fedora8 1.5 kB/s | 244 B 00:00 Feb 04 19:40:07 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm dnf[6387]: Errors during downloading metadata for repository 'beaker-client': Feb 04 19:40:07 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm dnf[6387]: - Status code: 404 for http://download.eng.bos.redhat.com/beakerrepos/client/Fedora8/repodata/repomd.xml (IP: 10.19.165.239) Feb 04 19:40:07 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm dnf[6387]: Error: Failed to download metadata for repo 'beaker-client': Cannot download repomd.xml: Status code: 404 for http://download.eng.bos.redhat.com/beakerrepos/client/Fedora8/repodata/repomd.xml (IP: 10.19.165.239) Feb 04 19:40:07 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm python3[6431]: ansible-ansible.legacy.command Invoked with _raw_params=podman --version _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 04 19:40:07 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_END pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:07 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGOUT pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:07 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm dnf[6387]: Beaker tasks 5.7 kB/s | 1.3 kB 00:00 Feb 04 19:40:07 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGIN pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:07 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_START pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:07 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: CRYPTO_KEY_USER pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:4d:0e:a0:55:e6:42:83:ce:eb:49:54:e8:e7:6b:94:70:b8:87:2b:24:6c:0c:40:0b:23:42:6b:40:30:1b:d1:22 direction=? spid=6440 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=? res=success' Feb 04 19:40:07 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm dnf[6387]: Fedora 39 - x86_64 64 kB/s | 24 kB 00:00 Feb 04 19:40:07 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm python3[6456]: ansible-getent Invoked with database=passwd key=root fail_key=False service=None split=None Feb 04 19:40:07 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_END pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:07 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGOUT pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:07 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGIN pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:07 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_START pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:07 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: CRYPTO_KEY_USER pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:4d:0e:a0:55:e6:42:83:ce:eb:49:54:e8:e7:6b:94:70:b8:87:2b:24:6c:0c:40:0b:23:42:6b:40:30:1b:d1:22 direction=? spid=6458 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=? res=success' Feb 04 19:40:08 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm python3[6474]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None Feb 04 19:40:08 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_END pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:08 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGOUT pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:08 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm dnf[6387]: Fedora 39 openh264 (From Cisco) - x86_64 16 kB/s | 989 B 00:00 Feb 04 19:40:08 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGIN pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:08 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_START pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:08 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: CRYPTO_KEY_USER pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:4d:0e:a0:55:e6:42:83:ce:eb:49:54:e8:e7:6b:94:70:b8:87:2b:24:6c:0c:40:0b:23:42:6b:40:30:1b:d1:22 direction=? spid=6477 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=? res=success' Feb 04 19:40:08 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm dnf[6387]: Fedora 39 - x86_64 - Updates 193 kB/s | 22 kB 00:00 Feb 04 19:40:08 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm python3[6494]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Feb 04 19:40:08 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_END pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:08 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGOUT pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:08 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm dnf[6387]: Copr repo for qa-tools owned by lpol 28 kB/s | 1.8 kB 00:00 Feb 04 19:40:08 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm dnf[6387]: Ignoring repositories: beaker-client Feb 04 19:40:08 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm dnf[6387]: Metadata cache created. Feb 04 19:40:08 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm systemd[1]: dnf-makecache.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ The unit dnf-makecache.service has successfully entered the 'dead' state. Feb 04 19:40:08 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm systemd[1]: Finished dnf-makecache.service - dnf makecache. ░░ Subject: A start job for unit dnf-makecache.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ A start job for unit dnf-makecache.service has finished successfully. ░░ ░░ The job identifier is 2454. Feb 04 19:40:08 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:init_t:s0 msg='unit=dnf-makecache comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 04 19:40:08 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:init_t:s0 msg='unit=dnf-makecache comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 04 19:40:08 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm systemd[1]: dnf-makecache.service: Consumed 2.031s CPU time. ░░ Subject: Resources consumed by unit runtime ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ The unit dnf-makecache.service completed and consumed the indicated resources. Feb 04 19:40:09 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGIN pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:09 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_START pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:09 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: CRYPTO_KEY_USER pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:4d:0e:a0:55:e6:42:83:ce:eb:49:54:e8:e7:6b:94:70:b8:87:2b:24:6c:0c:40:0b:23:42:6b:40:30:1b:d1:22 direction=? spid=6500 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=? res=success' Feb 04 19:40:09 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm python3[6516]: ansible-stat Invoked with path=/run/ostree-booted follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Feb 04 19:40:09 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_END pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:09 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGOUT pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:09 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGIN pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:09 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_START pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:09 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: CRYPTO_KEY_USER pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:4d:0e:a0:55:e6:42:83:ce:eb:49:54:e8:e7:6b:94:70:b8:87:2b:24:6c:0c:40:0b:23:42:6b:40:30:1b:d1:22 direction=? spid=6517 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=? res=success' Feb 04 19:40:10 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm python3[6533]: ansible-ansible.legacy.dnf Invoked with name=['firewalld'] state=present allow_downgrade=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 allowerasing=False nobest=False use_backend=auto conf_file=None disable_excludes=None download_dir=None list=None releasever=None Feb 04 19:40:11 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_END pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:11 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGOUT pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:11 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGIN pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:11 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_START pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:11 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: CRYPTO_KEY_USER pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:4d:0e:a0:55:e6:42:83:ce:eb:49:54:e8:e7:6b:94:70:b8:87:2b:24:6c:0c:40:0b:23:42:6b:40:30:1b:d1:22 direction=? spid=6536 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=? res=success' Feb 04 19:40:12 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm python3[6552]: ansible-systemd Invoked with name=firewalld masked=False daemon_reload=False daemon_reexec=False scope=system no_block=False state=None enabled=None force=None Feb 04 19:40:12 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_END pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:12 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGOUT pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:12 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGIN pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:12 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_START pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:12 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: CRYPTO_KEY_USER pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:4d:0e:a0:55:e6:42:83:ce:eb:49:54:e8:e7:6b:94:70:b8:87:2b:24:6c:0c:40:0b:23:42:6b:40:30:1b:d1:22 direction=? spid=6555 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=? res=success' Feb 04 19:40:12 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm python3[6571]: ansible-ansible.legacy.systemd Invoked with name=firewalld state=started enabled=True daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Feb 04 19:40:12 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm systemd[1]: Reloading requested from client PID 6574 ('systemctl') (unit session-6.scope)... Feb 04 19:40:12 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm systemd[1]: Reloading... Feb 04 19:40:13 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm systemd[1]: Reloading finished in 420 ms. Feb 04 19:40:13 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=165 op=LOAD Feb 04 19:40:13 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=147 op=UNLOAD Feb 04 19:40:13 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=166 op=LOAD Feb 04 19:40:13 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=167 op=LOAD Feb 04 19:40:13 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=148 op=UNLOAD Feb 04 19:40:13 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=149 op=UNLOAD Feb 04 19:40:13 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=168 op=LOAD Feb 04 19:40:13 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=169 op=LOAD Feb 04 19:40:13 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=150 op=UNLOAD Feb 04 19:40:13 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=151 op=UNLOAD Feb 04 19:40:13 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=170 op=LOAD Feb 04 19:40:13 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=152 op=UNLOAD Feb 04 19:40:13 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=171 op=LOAD Feb 04 19:40:13 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=153 op=UNLOAD Feb 04 19:40:13 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=172 op=LOAD Feb 04 19:40:13 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=173 op=LOAD Feb 04 19:40:13 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=154 op=UNLOAD Feb 04 19:40:13 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=155 op=UNLOAD Feb 04 19:40:13 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=174 op=LOAD Feb 04 19:40:13 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=156 op=UNLOAD Feb 04 19:40:13 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=175 op=LOAD Feb 04 19:40:13 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=176 op=LOAD Feb 04 19:40:13 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=157 op=UNLOAD Feb 04 19:40:13 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=158 op=UNLOAD Feb 04 19:40:13 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=177 op=LOAD Feb 04 19:40:13 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=159 op=UNLOAD Feb 04 19:40:13 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=178 op=LOAD Feb 04 19:40:13 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=160 op=UNLOAD Feb 04 19:40:13 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=179 op=LOAD Feb 04 19:40:13 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=161 op=UNLOAD Feb 04 19:40:13 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=180 op=LOAD Feb 04 19:40:13 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=162 op=UNLOAD Feb 04 19:40:13 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=181 op=LOAD Feb 04 19:40:13 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=182 op=LOAD Feb 04 19:40:13 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=163 op=UNLOAD Feb 04 19:40:13 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=164 op=UNLOAD Feb 04 19:40:13 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm systemd[1]: Starting firewalld.service - firewalld - dynamic firewall daemon... ░░ Subject: A start job for unit firewalld.service has begun execution ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ A start job for unit firewalld.service has begun execution. ░░ ░░ The job identifier is 2540. Feb 04 19:40:14 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm systemd[1]: Started firewalld.service - firewalld - dynamic firewall daemon. ░░ Subject: A start job for unit firewalld.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ A start job for unit firewalld.service has finished successfully. ░░ ░░ The job identifier is 2540. Feb 04 19:40:14 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:init_t:s0 msg='unit=firewalld comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 04 19:40:14 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_END pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:14 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGOUT pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:14 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGIN pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:14 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_START pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:14 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: CRYPTO_KEY_USER pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:4d:0e:a0:55:e6:42:83:ce:eb:49:54:e8:e7:6b:94:70:b8:87:2b:24:6c:0c:40:0b:23:42:6b:40:30:1b:d1:22 direction=? spid=6638 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=? res=success' Feb 04 19:40:14 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[6625]: NETFILTER_CFG table=firewalld:2 family=1 entries=1 op=nft_register_table pid=6625 subj=system_u:system_r:firewalld_t:s0 comm="firewalld" Feb 04 19:40:14 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[6625]: NETFILTER_CFG table=firewalld:2 family=1 entries=2 op=nft_unregister_table pid=6625 subj=system_u:system_r:firewalld_t:s0 comm="firewalld" Feb 04 19:40:14 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm python3[6654]: ansible-fedora.linux_system_roles.firewall_lib Invoked with port=['8000/tcp'] permanent=True runtime=True state=enabled __report_changed=True service=[] source_port=[] forward_port=[] rich_rule=[] source=[] interface=[] interface_pci_id=[] icmp_block=[] timeout=0 ipset_entries=[] protocol=[] helper_module=[] destination=[] firewalld_conf=None masquerade=None icmp_block_inversion=None target=None zone=None set_default_zone=None ipset=None ipset_type=None description=None short=None Feb 04 19:40:15 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[6625]: NETFILTER_CFG table=firewalld:3 family=1 entries=227 op=nft_register_chain pid=6625 subj=system_u:system_r:firewalld_t:s0 comm="firewalld" Feb 04 19:40:15 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[6625]: NETFILTER_CFG table=firewalld_policy_drop:4 family=1 entries=2 op=nft_unregister_table pid=6625 subj=system_u:system_r:firewalld_t:s0 comm="firewalld" Feb 04 19:40:15 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=183 op=LOAD Feb 04 19:40:15 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm systemd[1]: Starting polkit.service - Authorization Manager... ░░ Subject: A start job for unit polkit.service has begun execution ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ A start job for unit polkit.service has begun execution. ░░ ░░ The job identifier is 2632. Feb 04 19:40:15 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm polkitd[6685]: Started polkitd version 123 Feb 04 19:40:15 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm polkitd[6685]: Loading rules from directory /etc/polkit-1/rules.d Feb 04 19:40:15 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm polkitd[6685]: Loading rules from directory /usr/share/polkit-1/rules.d Feb 04 19:40:15 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm polkitd[6685]: Finished loading, compiling and executing 4 rules Feb 04 19:40:15 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm systemd[1]: Started polkit.service - Authorization Manager. ░░ Subject: A start job for unit polkit.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ A start job for unit polkit.service has finished successfully. ░░ ░░ The job identifier is 2632. Feb 04 19:40:15 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:init_t:s0 msg='unit=polkit comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 04 19:40:15 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm polkitd[6685]: Acquired the name org.freedesktop.PolicyKit1 on the system bus Feb 04 19:40:15 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[6625]: NETFILTER_CFG table=firewalld:5 family=1 entries=26 op=nft_register_rule pid=6625 subj=system_u:system_r:firewalld_t:s0 comm="firewalld" Feb 04 19:40:15 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[6625]: NETFILTER_CFG table=firewalld:6 family=1 entries=1 op=nft_register_rule pid=6625 subj=system_u:system_r:firewalld_t:s0 comm="firewalld" Feb 04 19:40:15 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_END pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:15 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGOUT pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:15 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGIN pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:15 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_START pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:15 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: CRYPTO_KEY_USER pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:4d:0e:a0:55:e6:42:83:ce:eb:49:54:e8:e7:6b:94:70:b8:87:2b:24:6c:0c:40:0b:23:42:6b:40:30:1b:d1:22 direction=? spid=6696 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=? res=success' Feb 04 19:40:15 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm python3[6712]: ansible-fedora.linux_system_roles.firewall_lib Invoked with port=['9000/tcp'] permanent=True runtime=True state=enabled __report_changed=True service=[] source_port=[] forward_port=[] rich_rule=[] source=[] interface=[] interface_pci_id=[] icmp_block=[] timeout=0 ipset_entries=[] protocol=[] helper_module=[] destination=[] firewalld_conf=None masquerade=None icmp_block_inversion=None target=None zone=None set_default_zone=None ipset=None ipset_type=None description=None short=None Feb 04 19:40:15 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[6625]: NETFILTER_CFG table=firewalld:7 family=1 entries=1 op=nft_register_rule pid=6625 subj=system_u:system_r:firewalld_t:s0 comm="firewalld" Feb 04 19:40:15 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_END pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:15 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGOUT pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:16 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGIN pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:16 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_START pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:16 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: CRYPTO_KEY_USER pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:4d:0e:a0:55:e6:42:83:ce:eb:49:54:e8:e7:6b:94:70:b8:87:2b:24:6c:0c:40:0b:23:42:6b:40:30:1b:d1:22 direction=? spid=6713 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=? res=success' Feb 04 19:40:16 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_END pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:16 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGOUT pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:17 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGIN pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:17 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_START pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:17 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: CRYPTO_KEY_USER pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:4d:0e:a0:55:e6:42:83:ce:eb:49:54:e8:e7:6b:94:70:b8:87:2b:24:6c:0c:40:0b:23:42:6b:40:30:1b:d1:22 direction=? spid=6737 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=? res=success' Feb 04 19:40:17 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_END pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:17 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGOUT pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:17 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGIN pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:17 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_START pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:17 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: CRYPTO_KEY_USER pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:4d:0e:a0:55:e6:42:83:ce:eb:49:54:e8:e7:6b:94:70:b8:87:2b:24:6c:0c:40:0b:23:42:6b:40:30:1b:d1:22 direction=? spid=6760 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=? res=success' Feb 04 19:40:17 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_END pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:17 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGOUT pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:18 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGIN pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:18 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_START pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:18 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: CRYPTO_KEY_USER pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:4d:0e:a0:55:e6:42:83:ce:eb:49:54:e8:e7:6b:94:70:b8:87:2b:24:6c:0c:40:0b:23:42:6b:40:30:1b:d1:22 direction=? spid=6783 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=? res=success' Feb 04 19:40:18 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm python3[6799]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None Feb 04 19:40:18 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_END pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:18 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGOUT pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:18 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGIN pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:18 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_START pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:18 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: CRYPTO_KEY_USER pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:4d:0e:a0:55:e6:42:83:ce:eb:49:54:e8:e7:6b:94:70:b8:87:2b:24:6c:0c:40:0b:23:42:6b:40:30:1b:d1:22 direction=? spid=6801 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=? res=success' Feb 04 19:40:18 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm python3[6817]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Feb 04 19:40:18 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_END pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:18 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGOUT pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:19 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGIN pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:19 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_START pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:19 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: CRYPTO_KEY_USER pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:4d:0e:a0:55:e6:42:83:ce:eb:49:54:e8:e7:6b:94:70:b8:87:2b:24:6c:0c:40:0b:23:42:6b:40:30:1b:d1:22 direction=? spid=6820 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=? res=success' Feb 04 19:40:19 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm python3[6836]: ansible-file Invoked with path=/etc/containers/systemd state=directory owner=root group=0 mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Feb 04 19:40:19 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_END pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:19 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGOUT pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:19 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGIN pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:19 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_START pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:19 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: CRYPTO_KEY_USER pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:4d:0e:a0:55:e6:42:83:ce:eb:49:54:e8:e7:6b:94:70:b8:87:2b:24:6c:0c:40:0b:23:42:6b:40:30:1b:d1:22 direction=? spid=6837 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=? res=success' Feb 04 19:40:19 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_END pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:19 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGOUT pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:19 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGIN pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:19 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_START pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:19 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: CRYPTO_KEY_USER pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:4d:0e:a0:55:e6:42:83:ce:eb:49:54:e8:e7:6b:94:70:b8:87:2b:24:6c:0c:40:0b:23:42:6b:40:30:1b:d1:22 direction=? spid=6853 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=? res=success' Feb 04 19:40:19 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_END pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:19 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGOUT pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:19 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGIN pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:19 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_START pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:19 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: CRYPTO_KEY_USER pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:4d:0e:a0:55:e6:42:83:ce:eb:49:54:e8:e7:6b:94:70:b8:87:2b:24:6c:0c:40:0b:23:42:6b:40:30:1b:d1:22 direction=? spid=6875 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=? res=success' Feb 04 19:40:19 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm python3[6891]: ansible-ansible.legacy.stat Invoked with path=/etc/containers/systemd/quadlet-demo.network follow=False get_checksum=True checksum_algorithm=sha1 get_mime=True get_attributes=True Feb 04 19:40:19 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_END pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:19 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGOUT pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:19 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGIN pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:19 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_START pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:19 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: CRYPTO_KEY_USER pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:4d:0e:a0:55:e6:42:83:ce:eb:49:54:e8:e7:6b:94:70:b8:87:2b:24:6c:0c:40:0b:23:42:6b:40:30:1b:d1:22 direction=? spid=6892 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=? res=success' Feb 04 19:40:19 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_END pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:19 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGOUT pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:19 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGIN pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:19 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_START pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:19 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: CRYPTO_KEY_USER pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:4d:0e:a0:55:e6:42:83:ce:eb:49:54:e8:e7:6b:94:70:b8:87:2b:24:6c:0c:40:0b:23:42:6b:40:30:1b:d1:22 direction=? spid=6908 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=? res=success' Feb 04 19:40:19 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_END pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:19 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGOUT pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:20 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGIN pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:20 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_START pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:20 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: CRYPTO_KEY_USER pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:4d:0e:a0:55:e6:42:83:ce:eb:49:54:e8:e7:6b:94:70:b8:87:2b:24:6c:0c:40:0b:23:42:6b:40:30:1b:d1:22 direction=? spid=6925 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=? res=success' Feb 04 19:40:20 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm python3[6941]: ansible-ansible.legacy.copy Invoked with src=/root/.ansible/tmp/ansible-tmp-1707075619.6203864-7913-42245081858217/source dest=/etc/containers/systemd/quadlet-demo.network owner=root group=0 mode=0644 _original_basename=quadlet-demo.network follow=False checksum=e57c08d49aff4bae8daab138d913aeddaa8682a0 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Feb 04 19:40:20 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_END pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:20 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGOUT pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:20 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGIN pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:20 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_START pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:20 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: CRYPTO_KEY_USER pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:4d:0e:a0:55:e6:42:83:ce:eb:49:54:e8:e7:6b:94:70:b8:87:2b:24:6c:0c:40:0b:23:42:6b:40:30:1b:d1:22 direction=? spid=6942 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=? res=success' Feb 04 19:40:20 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_END pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:20 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGOUT pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:20 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGIN pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:20 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_START pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:20 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: CRYPTO_KEY_USER pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:4d:0e:a0:55:e6:42:83:ce:eb:49:54:e8:e7:6b:94:70:b8:87:2b:24:6c:0c:40:0b:23:42:6b:40:30:1b:d1:22 direction=? spid=6959 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=? res=success' Feb 04 19:40:20 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm python3[6975]: ansible-systemd Invoked with daemon_reload=True scope=system daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None Feb 04 19:40:20 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm systemd[1]: Reloading requested from client PID 6976 ('systemctl') (unit session-6.scope)... Feb 04 19:40:20 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm systemd[1]: Reloading... Feb 04 19:40:21 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm systemd[1]: Reloading finished in 433 ms. Feb 04 19:40:21 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=184 op=LOAD Feb 04 19:40:21 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=165 op=UNLOAD Feb 04 19:40:21 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=185 op=LOAD Feb 04 19:40:21 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=186 op=LOAD Feb 04 19:40:21 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=166 op=UNLOAD Feb 04 19:40:21 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=167 op=UNLOAD Feb 04 19:40:21 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=187 op=LOAD Feb 04 19:40:21 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=188 op=LOAD Feb 04 19:40:21 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=168 op=UNLOAD Feb 04 19:40:21 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=169 op=UNLOAD Feb 04 19:40:21 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=189 op=LOAD Feb 04 19:40:21 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=170 op=UNLOAD Feb 04 19:40:21 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=190 op=LOAD Feb 04 19:40:21 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=171 op=UNLOAD Feb 04 19:40:21 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=191 op=LOAD Feb 04 19:40:21 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=192 op=LOAD Feb 04 19:40:21 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=172 op=UNLOAD Feb 04 19:40:21 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=173 op=UNLOAD Feb 04 19:40:21 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=193 op=LOAD Feb 04 19:40:21 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=174 op=UNLOAD Feb 04 19:40:21 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=194 op=LOAD Feb 04 19:40:21 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=195 op=LOAD Feb 04 19:40:21 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=175 op=UNLOAD Feb 04 19:40:21 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=176 op=UNLOAD Feb 04 19:40:21 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=196 op=LOAD Feb 04 19:40:21 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=177 op=UNLOAD Feb 04 19:40:21 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=197 op=LOAD Feb 04 19:40:21 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=183 op=UNLOAD Feb 04 19:40:21 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=198 op=LOAD Feb 04 19:40:21 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=178 op=UNLOAD Feb 04 19:40:21 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=199 op=LOAD Feb 04 19:40:21 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=179 op=UNLOAD Feb 04 19:40:21 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=200 op=LOAD Feb 04 19:40:21 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=180 op=UNLOAD Feb 04 19:40:21 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=201 op=LOAD Feb 04 19:40:21 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=202 op=LOAD Feb 04 19:40:21 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=181 op=UNLOAD Feb 04 19:40:21 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=182 op=UNLOAD Feb 04 19:40:21 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_END pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:21 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGOUT pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:21 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGIN pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:21 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_START pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:21 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: CRYPTO_KEY_USER pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:4d:0e:a0:55:e6:42:83:ce:eb:49:54:e8:e7:6b:94:70:b8:87:2b:24:6c:0c:40:0b:23:42:6b:40:30:1b:d1:22 direction=? spid=7026 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=? res=success' Feb 04 19:40:21 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm python3[7042]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None Feb 04 19:40:21 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_END pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:21 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGOUT pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:21 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGIN pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:21 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_START pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:21 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: CRYPTO_KEY_USER pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:4d:0e:a0:55:e6:42:83:ce:eb:49:54:e8:e7:6b:94:70:b8:87:2b:24:6c:0c:40:0b:23:42:6b:40:30:1b:d1:22 direction=? spid=7044 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=? res=success' Feb 04 19:40:22 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm python3[7060]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Feb 04 19:40:22 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_END pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:22 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGOUT pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:22 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGIN pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:22 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_START pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:22 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: CRYPTO_KEY_USER pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:4d:0e:a0:55:e6:42:83:ce:eb:49:54:e8:e7:6b:94:70:b8:87:2b:24:6c:0c:40:0b:23:42:6b:40:30:1b:d1:22 direction=? spid=7063 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=? res=success' Feb 04 19:40:22 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm python3[7079]: ansible-file Invoked with path=/etc/containers/systemd state=directory owner=root group=0 mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Feb 04 19:40:22 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_END pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:22 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGOUT pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:22 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGIN pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:22 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_START pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:22 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: CRYPTO_KEY_USER pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:4d:0e:a0:55:e6:42:83:ce:eb:49:54:e8:e7:6b:94:70:b8:87:2b:24:6c:0c:40:0b:23:42:6b:40:30:1b:d1:22 direction=? spid=7080 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=? res=success' Feb 04 19:40:22 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_END pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:22 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGOUT pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:22 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGIN pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:22 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_START pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:22 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: CRYPTO_KEY_USER pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:4d:0e:a0:55:e6:42:83:ce:eb:49:54:e8:e7:6b:94:70:b8:87:2b:24:6c:0c:40:0b:23:42:6b:40:30:1b:d1:22 direction=? spid=7096 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=? res=success' Feb 04 19:40:22 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_END pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:22 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGOUT pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:22 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGIN pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:22 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_START pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:22 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: CRYPTO_KEY_USER pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:4d:0e:a0:55:e6:42:83:ce:eb:49:54:e8:e7:6b:94:70:b8:87:2b:24:6c:0c:40:0b:23:42:6b:40:30:1b:d1:22 direction=? spid=7118 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=? res=success' Feb 04 19:40:23 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm python3[7134]: ansible-ansible.legacy.stat Invoked with path=/etc/containers/systemd/quadlet-demo-mysql.volume follow=False get_checksum=True checksum_algorithm=sha1 get_mime=True get_attributes=True Feb 04 19:40:23 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_END pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:23 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGOUT pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:23 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGIN pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:23 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_START pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:23 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: CRYPTO_KEY_USER pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:4d:0e:a0:55:e6:42:83:ce:eb:49:54:e8:e7:6b:94:70:b8:87:2b:24:6c:0c:40:0b:23:42:6b:40:30:1b:d1:22 direction=? spid=7135 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=? res=success' Feb 04 19:40:23 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_END pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:23 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGOUT pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:23 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGIN pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:23 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_START pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:23 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: CRYPTO_KEY_USER pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:4d:0e:a0:55:e6:42:83:ce:eb:49:54:e8:e7:6b:94:70:b8:87:2b:24:6c:0c:40:0b:23:42:6b:40:30:1b:d1:22 direction=? spid=7151 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=? res=success' Feb 04 19:40:23 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_END pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:23 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGOUT pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:23 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGIN pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:23 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_START pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:23 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: CRYPTO_KEY_USER pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:4d:0e:a0:55:e6:42:83:ce:eb:49:54:e8:e7:6b:94:70:b8:87:2b:24:6c:0c:40:0b:23:42:6b:40:30:1b:d1:22 direction=? spid=7168 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=? res=success' Feb 04 19:40:23 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm python3[7184]: ansible-ansible.legacy.copy Invoked with src=/root/.ansible/tmp/ansible-tmp-1707075622.8734992-7961-161658592115508/source dest=/etc/containers/systemd/quadlet-demo-mysql.volume owner=root group=0 mode=0644 _original_basename=quadlet-demo-mysql.volume follow=False checksum=585f8cbdf0ec73000f9227dcffbef71e9552ea4a backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Feb 04 19:40:23 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_END pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:23 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGOUT pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:23 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGIN pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:23 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_START pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:23 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: CRYPTO_KEY_USER pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:4d:0e:a0:55:e6:42:83:ce:eb:49:54:e8:e7:6b:94:70:b8:87:2b:24:6c:0c:40:0b:23:42:6b:40:30:1b:d1:22 direction=? spid=7185 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=? res=success' Feb 04 19:40:23 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_END pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:23 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGOUT pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:23 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGIN pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:23 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_START pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:23 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: CRYPTO_KEY_USER pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:4d:0e:a0:55:e6:42:83:ce:eb:49:54:e8:e7:6b:94:70:b8:87:2b:24:6c:0c:40:0b:23:42:6b:40:30:1b:d1:22 direction=? spid=7202 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=? res=success' Feb 04 19:40:23 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm python3[7218]: ansible-systemd Invoked with daemon_reload=True scope=system daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None Feb 04 19:40:23 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm systemd[1]: Reloading requested from client PID 7219 ('systemctl') (unit session-6.scope)... Feb 04 19:40:23 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm systemd[1]: Reloading... Feb 04 19:40:24 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm systemd[1]: Reloading finished in 423 ms. Feb 04 19:40:24 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=203 op=LOAD Feb 04 19:40:24 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=184 op=UNLOAD Feb 04 19:40:24 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=204 op=LOAD Feb 04 19:40:24 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=205 op=LOAD Feb 04 19:40:24 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=185 op=UNLOAD Feb 04 19:40:24 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=186 op=UNLOAD Feb 04 19:40:24 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=206 op=LOAD Feb 04 19:40:24 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=207 op=LOAD Feb 04 19:40:24 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=187 op=UNLOAD Feb 04 19:40:24 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=188 op=UNLOAD Feb 04 19:40:24 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=208 op=LOAD Feb 04 19:40:24 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=189 op=UNLOAD Feb 04 19:40:24 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=209 op=LOAD Feb 04 19:40:24 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=190 op=UNLOAD Feb 04 19:40:24 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=210 op=LOAD Feb 04 19:40:24 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=211 op=LOAD Feb 04 19:40:24 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=191 op=UNLOAD Feb 04 19:40:24 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=192 op=UNLOAD Feb 04 19:40:24 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=212 op=LOAD Feb 04 19:40:24 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=193 op=UNLOAD Feb 04 19:40:24 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=213 op=LOAD Feb 04 19:40:24 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=214 op=LOAD Feb 04 19:40:24 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=194 op=UNLOAD Feb 04 19:40:24 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=195 op=UNLOAD Feb 04 19:40:24 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=215 op=LOAD Feb 04 19:40:24 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=196 op=UNLOAD Feb 04 19:40:24 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=216 op=LOAD Feb 04 19:40:24 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=197 op=UNLOAD Feb 04 19:40:24 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=217 op=LOAD Feb 04 19:40:24 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=198 op=UNLOAD Feb 04 19:40:24 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=218 op=LOAD Feb 04 19:40:24 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=199 op=UNLOAD Feb 04 19:40:24 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=219 op=LOAD Feb 04 19:40:24 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=200 op=UNLOAD Feb 04 19:40:24 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=220 op=LOAD Feb 04 19:40:24 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=221 op=LOAD Feb 04 19:40:24 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=201 op=UNLOAD Feb 04 19:40:24 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=202 op=UNLOAD Feb 04 19:40:24 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_END pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:24 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGOUT pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:24 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGIN pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:24 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_START pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:24 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: CRYPTO_KEY_USER pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:4d:0e:a0:55:e6:42:83:ce:eb:49:54:e8:e7:6b:94:70:b8:87:2b:24:6c:0c:40:0b:23:42:6b:40:30:1b:d1:22 direction=? spid=7269 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=? res=success' Feb 04 19:40:24 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm python3[7285]: ansible-systemd Invoked with name=quadlet-demo-mysql-volume.service scope=system state=started daemon_reload=False daemon_reexec=False no_block=False enabled=None force=None masked=None Feb 04 19:40:24 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm systemd[1]: Starting quadlet-demo-mysql-volume.service... ░░ Subject: A start job for unit quadlet-demo-mysql-volume.service has begun execution ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ A start job for unit quadlet-demo-mysql-volume.service has begun execution. ░░ ░░ The job identifier is 2718. Feb 04 19:40:24 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm podman[7288]: 2024-02-04 19:40:24.981398446 +0000 UTC m=+0.185820977 system refresh Feb 04 19:40:24 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm podman[7288]: 2024-02-04 19:40:24.989791907 +0000 UTC m=+0.194214409 volume create systemd-quadlet-demo-mysql Feb 04 19:40:24 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm quadlet-demo-mysql-volume[7288]: systemd-quadlet-demo-mysql Feb 04 19:40:24 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Feb 04 19:40:25 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm systemd[1]: Finished quadlet-demo-mysql-volume.service. ░░ Subject: A start job for unit quadlet-demo-mysql-volume.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ A start job for unit quadlet-demo-mysql-volume.service has finished successfully. ░░ ░░ The job identifier is 2718. Feb 04 19:40:24 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:init_t:s0 msg='unit=quadlet-demo-mysql-volume comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 04 19:40:25 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_END pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:25 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGOUT pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:25 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGIN pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:25 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_START pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:25 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: CRYPTO_KEY_USER pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:4d:0e:a0:55:e6:42:83:ce:eb:49:54:e8:e7:6b:94:70:b8:87:2b:24:6c:0c:40:0b:23:42:6b:40:30:1b:d1:22 direction=? spid=7298 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=? res=success' Feb 04 19:40:25 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm python3[7314]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None Feb 04 19:40:25 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_END pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:25 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGOUT pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:25 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGIN pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:25 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_START pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:25 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: CRYPTO_KEY_USER pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:4d:0e:a0:55:e6:42:83:ce:eb:49:54:e8:e7:6b:94:70:b8:87:2b:24:6c:0c:40:0b:23:42:6b:40:30:1b:d1:22 direction=? spid=7316 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=? res=success' Feb 04 19:40:25 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm python3[7332]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Feb 04 19:40:25 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_END pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:25 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGOUT pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:26 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGIN pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:26 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_START pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:26 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: CRYPTO_KEY_USER pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:4d:0e:a0:55:e6:42:83:ce:eb:49:54:e8:e7:6b:94:70:b8:87:2b:24:6c:0c:40:0b:23:42:6b:40:30:1b:d1:22 direction=? spid=7335 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=? res=success' Feb 04 19:40:26 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm python3[7351]: ansible-file Invoked with path=/tmp/quadlet_demo state=directory owner=root group=root mode=0777 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Feb 04 19:40:26 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_END pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:26 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGOUT pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:26 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGIN pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:26 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_START pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:26 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: CRYPTO_KEY_USER pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:4d:0e:a0:55:e6:42:83:ce:eb:49:54:e8:e7:6b:94:70:b8:87:2b:24:6c:0c:40:0b:23:42:6b:40:30:1b:d1:22 direction=? spid=7352 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=? res=success' Feb 04 19:40:26 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm python3[7368]: ansible-containers.podman.podman_image Invoked with name=quay.io/linux-system-roles/mysql:5.6 force=True tag=latest pull=True push=False state=present executable=podman build={'force_rm': False, 'format': 'oci', 'cache': True, 'rm': True, 'annotation': None, 'file': None, 'volume': None, 'extra_args': None, 'target': None} push_args={'compress': None, 'format': None, 'remove_signatures': None, 'sign_by': None, 'dest': None, 'transport': None} arch=None path=None validate_certs=None auth_file=None username=None password=NOT_LOGGING_PARAMETER ca_cert_dir=None Feb 04 19:40:26 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm python3[7368]: ansible-containers.podman.podman_image PODMAN-IMAGE-DEBUG: /usr/bin/podman image ls quay.io/linux-system-roles/mysql:5.6 --format json Feb 04 19:40:27 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Feb 04 19:40:27 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm python3[7368]: ansible-containers.podman.podman_image PODMAN-IMAGE-DEBUG: /usr/bin/podman image exists quay.io/linux-system-roles/mysql:5.6 Feb 04 19:40:27 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Feb 04 19:40:27 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm python3[7368]: ansible-containers.podman.podman_image PODMAN-IMAGE-DEBUG: /usr/bin/podman pull quay.io/linux-system-roles/mysql:5.6 -q Feb 04 19:40:32 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm systemd[1]: var-lib-containers-storage-overlay-compat901343853-lower\x2dmapped.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ The unit var-lib-containers-storage-overlay-compat901343853-lower\x2dmapped.mount has successfully entered the 'dead' state. Feb 04 19:40:36 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm podman[7383]: 2024-02-04 19:40:36.424494361 +0000 UTC m=+9.356104097 image pull dd3b2a5dcb48ff61113592ed5ddd762581be4387c7bc552375a2159422aa6bf5 quay.io/linux-system-roles/mysql:5.6 Feb 04 19:40:36 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Feb 04 19:40:36 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm python3[7368]: ansible-containers.podman.podman_image PODMAN-IMAGE-DEBUG: /usr/bin/podman inspect dd3b2a5dcb48ff61113592ed5ddd762581be4387c7bc552375a2159422aa6bf5 --format json Feb 04 19:40:36 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Feb 04 19:40:36 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_END pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:36 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGOUT pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:36 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGIN pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:36 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_START pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:36 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: CRYPTO_KEY_USER pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:4d:0e:a0:55:e6:42:83:ce:eb:49:54:e8:e7:6b:94:70:b8:87:2b:24:6c:0c:40:0b:23:42:6b:40:30:1b:d1:22 direction=? spid=7472 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=? res=success' Feb 04 19:40:36 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm python3[7488]: ansible-file Invoked with path=/etc/containers/systemd state=directory owner=root group=0 mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Feb 04 19:40:36 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_END pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:36 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGOUT pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:36 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGIN pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:36 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_START pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:36 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: CRYPTO_KEY_USER pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:4d:0e:a0:55:e6:42:83:ce:eb:49:54:e8:e7:6b:94:70:b8:87:2b:24:6c:0c:40:0b:23:42:6b:40:30:1b:d1:22 direction=? spid=7489 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=? res=success' Feb 04 19:40:36 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_END pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:36 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGOUT pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:36 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGIN pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:36 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_START pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:36 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: CRYPTO_KEY_USER pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:4d:0e:a0:55:e6:42:83:ce:eb:49:54:e8:e7:6b:94:70:b8:87:2b:24:6c:0c:40:0b:23:42:6b:40:30:1b:d1:22 direction=? spid=7505 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=? res=success' Feb 04 19:40:36 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_END pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:36 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGOUT pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:36 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGIN pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:36 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_START pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:36 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: CRYPTO_KEY_USER pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:4d:0e:a0:55:e6:42:83:ce:eb:49:54:e8:e7:6b:94:70:b8:87:2b:24:6c:0c:40:0b:23:42:6b:40:30:1b:d1:22 direction=? spid=7527 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=? res=success' Feb 04 19:40:37 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm python3[7543]: ansible-ansible.legacy.stat Invoked with path=/etc/containers/systemd/quadlet-demo-mysql.container follow=False get_checksum=True checksum_algorithm=sha1 get_mime=True get_attributes=True Feb 04 19:40:37 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_END pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:37 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGOUT pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:37 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGIN pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:37 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_START pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:37 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: CRYPTO_KEY_USER pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:4d:0e:a0:55:e6:42:83:ce:eb:49:54:e8:e7:6b:94:70:b8:87:2b:24:6c:0c:40:0b:23:42:6b:40:30:1b:d1:22 direction=? spid=7544 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=? res=success' Feb 04 19:40:37 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_END pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:37 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGOUT pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:37 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGIN pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:37 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_START pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:37 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: CRYPTO_KEY_USER pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:4d:0e:a0:55:e6:42:83:ce:eb:49:54:e8:e7:6b:94:70:b8:87:2b:24:6c:0c:40:0b:23:42:6b:40:30:1b:d1:22 direction=? spid=7560 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=? res=success' Feb 04 19:40:37 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_END pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:37 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGOUT pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:37 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGIN pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:37 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_START pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:37 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: CRYPTO_KEY_USER pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:4d:0e:a0:55:e6:42:83:ce:eb:49:54:e8:e7:6b:94:70:b8:87:2b:24:6c:0c:40:0b:23:42:6b:40:30:1b:d1:22 direction=? spid=7577 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=? res=success' Feb 04 19:40:37 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm python3[7593]: ansible-ansible.legacy.copy Invoked with dest=/etc/containers/systemd/quadlet-demo-mysql.container owner=root group=0 mode=0644 src=/root/.ansible/tmp/ansible-tmp-1707075636.8441758-8013-248821270230451/source _original_basename=tmpq25kuzpd follow=False checksum=7bf9312d57274d4dabbd222b01717e2cc0cfa0ed backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Feb 04 19:40:37 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_END pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:37 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGOUT pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:37 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGIN pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:37 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_START pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:37 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: CRYPTO_KEY_USER pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:4d:0e:a0:55:e6:42:83:ce:eb:49:54:e8:e7:6b:94:70:b8:87:2b:24:6c:0c:40:0b:23:42:6b:40:30:1b:d1:22 direction=? spid=7594 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=? res=success' Feb 04 19:40:37 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_END pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:37 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGOUT pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:37 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGIN pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:37 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_START pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:37 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: CRYPTO_KEY_USER pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:4d:0e:a0:55:e6:42:83:ce:eb:49:54:e8:e7:6b:94:70:b8:87:2b:24:6c:0c:40:0b:23:42:6b:40:30:1b:d1:22 direction=? spid=7611 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=? res=success' Feb 04 19:40:37 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm python3[7627]: ansible-systemd Invoked with daemon_reload=True scope=system daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None Feb 04 19:40:37 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm systemd[1]: Reloading requested from client PID 7628 ('systemctl') (unit session-6.scope)... Feb 04 19:40:37 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm systemd[1]: Reloading... Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm systemd[1]: Reloading finished in 425 ms. Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=222 op=LOAD Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=203 op=UNLOAD Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=223 op=LOAD Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=224 op=LOAD Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=204 op=UNLOAD Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=205 op=UNLOAD Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=225 op=LOAD Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=226 op=LOAD Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=206 op=UNLOAD Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=207 op=UNLOAD Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=227 op=LOAD Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=208 op=UNLOAD Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=228 op=LOAD Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=209 op=UNLOAD Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=229 op=LOAD Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=230 op=LOAD Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=210 op=UNLOAD Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=211 op=UNLOAD Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=231 op=LOAD Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=212 op=UNLOAD Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=232 op=LOAD Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=233 op=LOAD Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=213 op=UNLOAD Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=214 op=UNLOAD Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=234 op=LOAD Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=219 op=UNLOAD Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=235 op=LOAD Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=236 op=LOAD Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=220 op=UNLOAD Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=221 op=UNLOAD Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=237 op=LOAD Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=217 op=UNLOAD Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=238 op=LOAD Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=216 op=UNLOAD Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=239 op=LOAD Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=218 op=UNLOAD Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=240 op=LOAD Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=215 op=UNLOAD Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=241 op=LOAD Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=240 op=UNLOAD Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=242 op=LOAD Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=234 op=UNLOAD Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=243 op=LOAD Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=244 op=LOAD Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=235 op=UNLOAD Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=236 op=UNLOAD Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=245 op=LOAD Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=237 op=UNLOAD Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=246 op=LOAD Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=238 op=UNLOAD Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=247 op=LOAD Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=239 op=UNLOAD Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=248 op=LOAD Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=246 op=UNLOAD Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=249 op=LOAD Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=245 op=UNLOAD Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=250 op=LOAD Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=247 op=UNLOAD Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=251 op=LOAD Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=242 op=UNLOAD Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=252 op=LOAD Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=253 op=LOAD Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=243 op=UNLOAD Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit: BPF prog-id=244 op=UNLOAD Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_END pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGOUT pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGIN pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_START pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: CRYPTO_KEY_USER pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:4d:0e:a0:55:e6:42:83:ce:eb:49:54:e8:e7:6b:94:70:b8:87:2b:24:6c:0c:40:0b:23:42:6b:40:30:1b:d1:22 direction=? spid=7678 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=? res=success' Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm python3[7694]: ansible-systemd Invoked with name=quadlet-demo-mysql.service scope=system state=started daemon_reload=False daemon_reexec=False no_block=False enabled=None force=None masked=None Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm systemd[1]: Starting quadlet-demo-network.service... ░░ Subject: A start job for unit quadlet-demo-network.service has begun execution ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ A start job for unit quadlet-demo-network.service has begun execution. ░░ ░░ The job identifier is 2805. Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm quadlet-demo-network[7697]: systemd-quadlet-demo Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm systemd[1]: Finished quadlet-demo-network.service. ░░ Subject: A start job for unit quadlet-demo-network.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ A start job for unit quadlet-demo-network.service has finished successfully. ░░ ░░ The job identifier is 2805. Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:init_t:s0 msg='unit=quadlet-demo-network comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm systemd[1]: Starting quadlet-demo-mysql.service... ░░ Subject: A start job for unit quadlet-demo-mysql.service has begun execution ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ A start job for unit quadlet-demo-mysql.service has begun execution. ░░ ░░ The job identifier is 2804. Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm quadlet-demo-mysql[7705]: Error: running container create option: no secret with name or id "mysql-root-password-container": no such secret Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm podman[7705]: 2024-02-04 19:40:38.839519034 +0000 UTC m=+0.031994257 image pull dd3b2a5dcb48ff61113592ed5ddd762581be4387c7bc552375a2159422aa6bf5 quay.io/linux-system-roles/mysql:5.6 Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm systemd[1]: quadlet-demo-mysql.service: Main process exited, code=exited, status=125/n/a ░░ Subject: Unit process exited ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ An ExecStart= process belonging to unit quadlet-demo-mysql.service has exited. ░░ ░░ The process' exit code is 'exited' and its exit status is 125. Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm systemd[1]: quadlet-demo-mysql.service: Failed with result 'exit-code'. ░░ Subject: Unit failed ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ The unit quadlet-demo-mysql.service has entered the 'failed' state with result 'exit-code'. Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm systemd[1]: Failed to start quadlet-demo-mysql.service. ░░ Subject: A start job for unit quadlet-demo-mysql.service has failed ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ A start job for unit quadlet-demo-mysql.service has finished with a failure. ░░ ░░ The job identifier is 2804 and the job result is failed. Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:init_t:s0 msg='unit=quadlet-demo-mysql comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_END pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGOUT pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_LOGIN pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: USER_START pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=ssh res=success' Feb 04 19:40:38 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm audit[988]: CRYPTO_KEY_USER pid=988 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:4d:0e:a0:55:e6:42:83:ce:eb:49:54:e8:e7:6b:94:70:b8:87:2b:24:6c:0c:40:0b:23:42:6b:40:30:1b:d1:22 direction=? spid=7721 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.167 terminal=? res=success' Feb 04 19:40:39 5ab0178c-979c-4c30-bc0e-1780f08c3668.testing-farm python3[7737]: ansible-ansible.legacy.command Invoked with _raw_params=journalctl -ex _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None TASK [Check] ******************************************************************* task path: /WORKDIR/git-fix-linger-handlingciq7eztj/tests/tests_quadlet_demo.yml:136 Sunday 04 February 2024 19:40:39 +0000 (0:00:00.374) 0:02:11.153 ******* ok: [sut] => { "changed": false, "cmd": [ "podman", "ps", "-a" ], "delta": "0:00:00.035570", "end": "2024-02-04 19:40:39.561583", "rc": 0, "start": "2024-02-04 19:40:39.526013" } STDOUT: CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES TASK [Check pods] ************************************************************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/tests/tests_quadlet_demo.yml:140 Sunday 04 February 2024 19:40:39 +0000 (0:00:00.280) 0:02:11.434 ******* ok: [sut] => { "changed": false, "cmd": [ "podman", "pod", "ps", "--ctr-ids", "--ctr-names", "--ctr-status" ], "delta": "0:00:00.034743", "end": "2024-02-04 19:40:39.840739", "failed_when_result": false, "rc": 0, "start": "2024-02-04 19:40:39.805996" } STDOUT: POD ID NAME STATUS CREATED INFRA ID IDS NAMES STATUS TASK [Check systemd] *********************************************************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/tests/tests_quadlet_demo.yml:145 Sunday 04 February 2024 19:40:39 +0000 (0:00:00.279) 0:02:11.714 ******* ok: [sut] => { "changed": false, "cmd": "set -euo pipefail; systemctl list-units --all --with-dependencies | grep quadlet", "delta": "0:00:00.015592", "end": "2024-02-04 19:40:40.099301", "failed_when_result": false, "rc": 0, "start": "2024-02-04 19:40:40.083709" } STDOUT: quadlet-demo-mysql-volume.service loaded active exited quadlet-demo-mysql-volume.service ● quadlet-demo-mysql.service loaded failed failed quadlet-demo-mysql.service quadlet-demo-network.service loaded active exited quadlet-demo-network.service TASK [LS] ********************************************************************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/tests/tests_quadlet_demo.yml:153 Sunday 04 February 2024 19:40:40 +0000 (0:00:00.255) 0:02:11.970 ******* ok: [sut] => { "changed": false, "cmd": [ "ls", "-alrtF", "/etc/systemd/system" ], "delta": "0:00:00.004678", "end": "2024-02-04 19:40:40.342028", "failed_when_result": false, "rc": 0, "start": "2024-02-04 19:40:40.337350" } STDOUT: total 56 lrwxrwxrwx. 1 root root 43 Jan 30 09:29 dbus.service -> /usr/lib/systemd/system/dbus-broker.service lrwxrwxrwx. 1 root root 48 Jan 30 09:29 dbus-org.freedesktop.resolve1.service -> /usr/lib/systemd/system/systemd-resolved.service lrwxrwxrwx. 1 root root 44 Jan 30 09:29 dbus-org.freedesktop.oom1.service -> /usr/lib/systemd/system/systemd-oomd.service lrwxrwxrwx. 1 root root 37 Jan 30 09:29 ctrl-alt-del.target -> /usr/lib/systemd/system/reboot.target drwxr-xr-x. 2 root root 4096 Jan 30 09:29 systemd-journald.service.wants/ drwxr-xr-x. 2 root root 4096 Jan 30 09:29 getty.target.wants/ drwxr-xr-x. 5 root root 4096 Jan 30 09:29 ../ lrwxrwxrwx. 1 root root 45 Jan 30 09:29 dbus-org.freedesktop.home1.service -> /usr/lib/systemd/system/systemd-homed.service drwxr-xr-x. 2 root root 4096 Jan 30 09:29 systemd-homed.service.wants/ lrwxrwxrwx. 1 root root 41 Jan 30 09:29 dbus-org.bluez.service -> /usr/lib/systemd/system/bluetooth.service drwxr-xr-x. 2 root root 4096 Jan 30 09:29 bluetooth.target.wants/ drwxr-xr-x. 2 root root 4096 Jan 30 09:29 graphical.target.wants/ drwxr-xr-x. 2 root root 4096 Jan 30 09:30 timers.target.wants/ lrwxrwxrwx. 1 root root 57 Jan 30 09:30 dbus-org.freedesktop.nm-dispatcher.service -> /usr/lib/systemd/system/NetworkManager-dispatcher.service drwxr-xr-x. 2 root root 4096 Jan 30 09:30 network-online.target.wants/ lrwxrwxrwx. 1 root root 41 Jan 30 09:34 default.target -> /usr/lib/systemd/system/multi-user.target drwxr-xr-x. 2 root root 4096 Jan 30 09:47 remote-fs.target.wants/ drwxr-xr-x. 2 root root 4096 Jan 30 09:49 cloud-init.target.wants/ drwxr-xr-x. 2 root root 4096 Jan 30 09:50 sockets.target.wants/ drwxr-xr-x. 2 root root 4096 Feb 4 19:32 sysinit.target.wants/ drwxr-xr-x. 2 root root 4096 Feb 4 19:40 multi-user.target.wants/ lrwxrwxrwx. 1 root root 41 Feb 4 19:40 dbus-org.fedoraproject.FirewallD1.service -> /usr/lib/systemd/system/firewalld.service drwxr-xr-x. 14 root root 4096 Feb 4 19:40 ./ TASK [Cleanup] ***************************************************************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/tests/tests_quadlet_demo.yml:158 Sunday 04 February 2024 19:40:40 +0000 (0:00:00.244) 0:02:12.214 ******* TASK [fedora.linux_system_roles.podman : Set platform/version specific variables] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:3 Sunday 04 February 2024 19:40:40 +0000 (0:00:00.062) 0:02:12.276 ******* included: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml for sut TASK [fedora.linux_system_roles.podman : Ensure ansible_facts used by role] **** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:3 Sunday 04 February 2024 19:40:40 +0000 (0:00:00.037) 0:02:12.314 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_required_facts | difference(ansible_facts.keys() | list) | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if system is ostree] ************ task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:11 Sunday 04 February 2024 19:40:40 +0000 (0:00:00.030) 0:02:12.345 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_is_ostree is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set flag to indicate system is ostree] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:16 Sunday 04 February 2024 19:40:40 +0000 (0:00:00.021) 0:02:12.366 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_is_ostree is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set platform/version specific variables] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:20 Sunday 04 February 2024 19:40:40 +0000 (0:00:00.019) 0:02:12.386 ******* ok: [sut] => (item=RedHat.yml) => { "ansible_facts": { "__podman_packages": [ "podman", "shadow-utils-subid" ] }, "ansible_included_var_files": [ "/WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/vars/RedHat.yml" ], "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml" } skipping: [sut] => (item=Fedora.yml) => { "ansible_loop_var": "item", "changed": false, "false_condition": "__vars_file is file", "item": "Fedora.yml", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=Fedora_39.yml) => { "ansible_loop_var": "item", "changed": false, "false_condition": "__vars_file is file", "item": "Fedora_39.yml", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=Fedora_39.yml) => { "ansible_loop_var": "item", "changed": false, "false_condition": "__vars_file is file", "item": "Fedora_39.yml", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Gather the package facts] ************* task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:6 Sunday 04 February 2024 19:40:40 +0000 (0:00:00.048) 0:02:12.434 ******* ok: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Enable copr if requested] ************* task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:10 Sunday 04 February 2024 19:40:41 +0000 (0:00:00.881) 0:02:13.316 ******* skipping: [sut] => { "changed": false, "false_condition": "podman_use_copr | d(false)", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Ensure required packages are installed] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:14 Sunday 04 February 2024 19:40:41 +0000 (0:00:00.049) 0:02:13.366 ******* skipping: [sut] => { "changed": false, "false_condition": "(__podman_packages | difference(ansible_facts.packages))", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get podman version] ******************* task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:22 Sunday 04 February 2024 19:40:41 +0000 (0:00:00.026) 0:02:13.392 ******* ok: [sut] => { "changed": false, "cmd": [ "podman", "--version" ], "delta": "0:00:00.031252", "end": "2024-02-04 19:40:41.796202", "rc": 0, "start": "2024-02-04 19:40:41.764950" } STDOUT: podman version 4.9.0 TASK [fedora.linux_system_roles.podman : Set podman version] ******************* task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:27 Sunday 04 February 2024 19:40:41 +0000 (0:00:00.277) 0:02:13.670 ******* ok: [sut] => { "ansible_facts": { "podman_version": "4.9.0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Podman package version must be 4.2 or later] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:31 Sunday 04 February 2024 19:40:41 +0000 (0:00:00.024) 0:02:13.694 ******* skipping: [sut] => { "changed": false, "false_condition": "podman_version is version(\"4.2\", \"<\")", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Podman package version must be 4.4 or later for quadlet, secrets] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:38 Sunday 04 February 2024 19:40:41 +0000 (0:00:00.021) 0:02:13.716 ******* skipping: [sut] => { "changed": false, "false_condition": "podman_version is version(\"4.4\", \"<\")", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Podman package version must be 4.4 or later for quadlet, secrets] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:48 Sunday 04 February 2024 19:40:41 +0000 (0:00:00.023) 0:02:13.739 ******* META: end_host conditional evaluated to False, continuing execution for sut skipping: [sut] => { "skip_reason": "end_host conditional evaluated to False, continuing execution for sut" } MSG: end_host conditional evaluated to false, continuing execution for sut TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:55 Sunday 04 February 2024 19:40:41 +0000 (0:00:00.026) 0:02:13.766 ******* included: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for sut TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Sunday 04 February 2024 19:40:41 +0000 (0:00:00.043) 0:02:13.810 ******* skipping: [sut] => { "changed": false, "false_condition": "'getent_passwd' not in ansible_facts or __podman_user not in ansible_facts['getent_passwd']", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Sunday 04 February 2024 19:40:41 +0000 (0:00:00.025) 0:02:13.835 ******* skipping: [sut] => { "changed": false, "false_condition": "not ansible_facts[\"getent_passwd\"][__podman_user]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Sunday 04 February 2024 19:40:42 +0000 (0:00:00.025) 0:02:13.861 ******* ok: [sut] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get group information] **************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:27 Sunday 04 February 2024 19:40:42 +0000 (0:00:00.031) 0:02:13.892 ******* ok: [sut] => { "ansible_facts": { "getent_group": { "root": [ "x", "0", "" ] } }, "changed": false } TASK [fedora.linux_system_roles.podman : Set group name] *********************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:34 Sunday 04 February 2024 19:40:42 +0000 (0:00:00.256) 0:02:14.148 ******* ok: [sut] => { "ansible_facts": { "__podman_group_name": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:38 Sunday 04 February 2024 19:40:42 +0000 (0:00:00.033) 0:02:14.182 ******* ok: [sut] => { "changed": false, "stat": { "atime": 1707075608.5761826, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "72ba28a3be19c467e89ab90f5d21f99b605159a7", "ctime": 1707075555.3834877, "dev": 51714, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 203277, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-pie-executable", "mode": "0755", "mtime": 1696291200.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 16032, "uid": 0, "version": "1460241045", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************ task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:49 Sunday 04 February 2024 19:40:42 +0000 (0:00:00.256) 0:02:14.438 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check group with getsubids] *********** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:53 Sunday 04 February 2024 19:40:42 +0000 (0:00:00.022) 0:02:14.460 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if user is in subuid file] ****** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:63 Sunday 04 February 2024 19:40:42 +0000 (0:00:00.022) 0:02:14.483 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:71 Sunday 04 February 2024 19:40:42 +0000 (0:00:00.021) 0:02:14.504 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if group is in subgid file] ***** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:78 Sunday 04 February 2024 19:40:42 +0000 (0:00:00.051) 0:02:14.556 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] ***** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:86 Sunday 04 February 2024 19:40:42 +0000 (0:00:00.021) 0:02:14.577 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set config file paths] **************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:61 Sunday 04 February 2024 19:40:42 +0000 (0:00:00.020) 0:02:14.598 ******* ok: [sut] => { "ansible_facts": { "__podman_container_conf_file": "/etc/containers/containers.conf.d/50-systemroles.conf", "__podman_policy_json_file": "/etc/containers/policy.json", "__podman_registries_conf_file": "/etc/containers/registries.conf.d/50-systemroles.conf", "__podman_storage_conf_file": "/etc/containers/storage.conf" }, "changed": false } TASK [fedora.linux_system_roles.podman : Handle container.conf.d] ************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:70 Sunday 04 February 2024 19:40:42 +0000 (0:00:00.029) 0:02:14.628 ******* included: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_container_conf_d.yml for sut TASK [fedora.linux_system_roles.podman : Ensure containers.d exists] *********** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_container_conf_d.yml:5 Sunday 04 February 2024 19:40:42 +0000 (0:00:00.043) 0:02:14.671 ******* skipping: [sut] => { "changed": false, "false_condition": "podman_containers_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Update container config file] ********* task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_container_conf_d.yml:13 Sunday 04 February 2024 19:40:42 +0000 (0:00:00.022) 0:02:14.694 ******* skipping: [sut] => { "changed": false, "false_condition": "podman_containers_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Handle registries.conf.d] ************* task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:73 Sunday 04 February 2024 19:40:42 +0000 (0:00:00.032) 0:02:14.727 ******* included: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_registries_conf_d.yml for sut TASK [fedora.linux_system_roles.podman : Ensure registries.d exists] *********** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_registries_conf_d.yml:5 Sunday 04 February 2024 19:40:42 +0000 (0:00:00.044) 0:02:14.771 ******* skipping: [sut] => { "changed": false, "false_condition": "podman_registries_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Update registries config file] ******** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_registries_conf_d.yml:13 Sunday 04 February 2024 19:40:42 +0000 (0:00:00.023) 0:02:14.794 ******* skipping: [sut] => { "changed": false, "false_condition": "podman_registries_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Handle storage.conf] ****************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:76 Sunday 04 February 2024 19:40:42 +0000 (0:00:00.023) 0:02:14.817 ******* included: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_storage_conf.yml for sut TASK [fedora.linux_system_roles.podman : Ensure storage.conf parent dir exists] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_storage_conf.yml:5 Sunday 04 February 2024 19:40:43 +0000 (0:00:00.044) 0:02:14.861 ******* skipping: [sut] => { "changed": false, "false_condition": "podman_storage_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Update storage config file] *********** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_storage_conf.yml:13 Sunday 04 February 2024 19:40:43 +0000 (0:00:00.022) 0:02:14.884 ******* skipping: [sut] => { "changed": false, "false_condition": "podman_storage_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Handle policy.json] ******************* task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:79 Sunday 04 February 2024 19:40:43 +0000 (0:00:00.023) 0:02:14.907 ******* included: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml for sut TASK [fedora.linux_system_roles.podman : Ensure policy.json parent dir exists] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:6 Sunday 04 February 2024 19:40:43 +0000 (0:00:00.045) 0:02:14.953 ******* skipping: [sut] => { "changed": false, "false_condition": "podman_policy_json | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Stat the policy.json file] ************ task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:14 Sunday 04 February 2024 19:40:43 +0000 (0:00:00.053) 0:02:15.006 ******* skipping: [sut] => { "changed": false, "false_condition": "podman_policy_json | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get the existing policy.json] ********* task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:19 Sunday 04 February 2024 19:40:43 +0000 (0:00:00.030) 0:02:15.037 ******* skipping: [sut] => { "changed": false, "false_condition": "podman_policy_json | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Write new policy.json file] *********** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:25 Sunday 04 February 2024 19:40:43 +0000 (0:00:00.033) 0:02:15.070 ******* skipping: [sut] => { "changed": false, "false_condition": "podman_policy_json | length > 0", "skip_reason": "Conditional result was False" } TASK [Manage firewall for specified ports] ************************************* task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:85 Sunday 04 February 2024 19:40:43 +0000 (0:00:00.035) 0:02:15.105 ******* TASK [fedora.linux_system_roles.firewall : Setup firewalld] ******************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:2 Sunday 04 February 2024 19:40:43 +0000 (0:00:00.080) 0:02:15.186 ******* included: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml for sut TASK [fedora.linux_system_roles.firewall : Ensure ansible_facts used by role] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:2 Sunday 04 February 2024 19:40:43 +0000 (0:00:00.038) 0:02:15.224 ******* skipping: [sut] => { "changed": false, "false_condition": "__firewall_required_facts | difference(ansible_facts.keys() | list) | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Check if system is ostree] ********** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:10 Sunday 04 February 2024 19:40:43 +0000 (0:00:00.029) 0:02:15.254 ******* skipping: [sut] => { "changed": false, "false_condition": "not __firewall_is_ostree is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Set flag to indicate system is ostree] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:15 Sunday 04 February 2024 19:40:43 +0000 (0:00:00.022) 0:02:15.277 ******* skipping: [sut] => { "changed": false, "false_condition": "not __firewall_is_ostree is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Install firewalld] ****************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:19 Sunday 04 February 2024 19:40:43 +0000 (0:00:00.021) 0:02:15.298 ******* ok: [sut] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [fedora.linux_system_roles.firewall : Collect service facts] ************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:5 Sunday 04 February 2024 19:40:45 +0000 (0:00:01.988) 0:02:17.286 ******* skipping: [sut] => { "changed": false, "false_condition": "firewall_disable_conflicting_services | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Attempt to stop and disable conflicting services] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:9 Sunday 04 February 2024 19:40:45 +0000 (0:00:00.021) 0:02:17.308 ******* skipping: [sut] => (item=nftables) => { "ansible_loop_var": "item", "changed": false, "false_condition": "firewall_disable_conflicting_services | bool", "item": "nftables", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=iptables) => { "ansible_loop_var": "item", "changed": false, "false_condition": "firewall_disable_conflicting_services | bool", "item": "iptables", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=ufw) => { "ansible_loop_var": "item", "changed": false, "false_condition": "firewall_disable_conflicting_services | bool", "item": "ufw", "skip_reason": "Conditional result was False" } skipping: [sut] => { "changed": false } MSG: All items skipped TASK [fedora.linux_system_roles.firewall : Unmask firewalld service] *********** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:22 Sunday 04 February 2024 19:40:45 +0000 (0:00:00.026) 0:02:17.334 ******* ok: [sut] => { "changed": false, "name": "firewalld", "status": { "AccessSELinuxContext": "system_u:object_r:firewalld_unit_file_t:s0", "ActiveEnterTimestamp": "Sun 2024-02-04 19:40:14 UTC", "ActiveEnterTimestampMonotonic": "740184677", "ActiveExitTimestampMonotonic": "0", "ActiveState": "active", "After": "dbus-broker.service dbus.socket polkit.service basic.target sysinit.target system.slice", "AllowIsolate": "no", "AssertResult": "yes", "AssertTimestamp": "Sun 2024-02-04 19:40:13 UTC", "AssertTimestampMonotonic": "739023197", "Before": "network-pre.target shutdown.target multi-user.target", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "BusName": "org.fedoraproject.FirewallD1", "CPUAccounting": "yes", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "635480000", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "yes", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore", "CleanResult": "success", "CollectMode": "inactive", "ConditionResult": "yes", "ConditionTimestamp": "Sun 2024-02-04 19:40:13 UTC", "ConditionTimestampMonotonic": "739023194", "ConfigurationDirectoryMode": "0755", "Conflicts": "nftables.service iptables.service ebtables.service ipset.service ip6tables.service shutdown.target", "ControlGroup": "/system.slice/firewalld.service", "ControlGroupId": "6105", "ControlPID": "0", "CoredumpFilter": "0x33", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "DefaultStartupMemoryLow": "0", "Delegate": "no", "Description": "firewalld - dynamic firewall daemon", "DevicePolicy": "auto", "Documentation": "\"man:firewalld(1)\"", "DropInPaths": "/usr/lib/systemd/system/service.d/10-timeout-abort.conf", "DynamicUser": "no", "EffectiveCPUs": "0-1", "EffectiveMemoryNodes": "0", "EnvironmentFiles": "/etc/sysconfig/firewalld (ignore_errors=yes)", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "6625", "ExecMainStartTimestamp": "Sun 2024-02-04 19:40:13 UTC", "ExecMainStartTimestampMonotonic": "739034144", "ExecMainStatus": "0", "ExecReload": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecReloadEx": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStart": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStartEx": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExitType": "main", "ExtensionImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FileDescriptorStorePreserve": "restart", "FinalKillSignal": "9", "FragmentPath": "/usr/lib/systemd/system/firewalld.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOReadBytes": "18446744073709551615", "IOReadOperations": "18446744073709551615", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "IOWeight": "[not set]", "IOWriteBytes": "18446744073709551615", "IOWriteOperations": "18446744073709551615", "IPAccounting": "no", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "Id": "firewalld.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestamp": "Sun 2024-02-04 19:40:13 UTC", "InactiveExitTimestampMonotonic": "739034492", "InvocationID": "5b2d1a3e8acc4f9d82d144128c51d0be", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "mixed", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "infinity", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "8388608", "LimitMEMLOCKSoft": "8388608", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "524288", "LimitNOFILESoft": "1024", "LimitNPROC": "14778", "LimitNPROCSoft": "14778", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "14778", "LimitSIGPENDINGSoft": "14778", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "6625", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "ManagedOOMSwap": "auto", "MemoryAccounting": "yes", "MemoryAvailable": "infinity", "MemoryCurrent": "33116160", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryKSM": "no", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemoryPressureThresholdUSec": "200ms", "MemoryPressureWatch": "auto", "MemorySwapMax": "infinity", "MemoryZSwapMax": "infinity", "MountAPIVFS": "no", "MountImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAPolicy": "n/a", "Names": "firewalld.service dbus-org.fedoraproject.FirewallD1.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMPolicy": "stop", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "OnSuccessJobMode": "fail", "Perpetual": "no", "PrivateDevices": "no", "PrivateIPC": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProcSubset": "all", "ProtectClock": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectHostname": "no", "ProtectKernelLogs": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectProc": "default", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "ReloadResult": "success", "ReloadSignal": "1", "RemainAfterExit": "no", "RemoveIPC": "no", "Requires": "system.slice dbus.socket sysinit.target", "Restart": "no", "RestartKillSignal": "15", "RestartMaxDelayUSec": "infinity", "RestartMode": "normal", "RestartSteps": "0", "RestartUSec": "100ms", "RestartUSecNext": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RootEphemeral": "no", "RootImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "Slice": "system.slice", "StandardError": "null", "StandardInput": "null", "StandardOutput": "null", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StartupMemoryHigh": "18446744073709551615", "StartupMemoryLow": "0", "StartupMemoryMax": "18446744073709551615", "StartupMemorySwapMax": "18446744073709551615", "StartupMemoryZSwapMax": "18446744073709551615", "StateChangeTimestamp": "Sun 2024-02-04 19:40:38 UTC", "StateChangeTimestampMonotonic": "764242582", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "running", "SuccessAction": "none", "SyslogFacility": "3", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "2147483646", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "2", "TasksMax": "4433", "TimeoutAbortUSec": "45s", "TimeoutCleanUSec": "infinity", "TimeoutStartFailureMode": "terminate", "TimeoutStartUSec": "45s", "TimeoutStopFailureMode": "abort", "TimeoutStopUSec": "45s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "dbus", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "enabled", "UnitFileState": "enabled", "UtmpMode": "init", "WantedBy": "multi-user.target", "Wants": "network-pre.target", "WatchdogSignal": "6", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "0" } } TASK [fedora.linux_system_roles.firewall : Enable and start firewalld service] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:28 Sunday 04 February 2024 19:40:45 +0000 (0:00:00.435) 0:02:17.770 ******* ok: [sut] => { "changed": false, "enabled": true, "name": "firewalld", "state": "started", "status": { "AccessSELinuxContext": "system_u:object_r:firewalld_unit_file_t:s0", "ActiveEnterTimestamp": "Sun 2024-02-04 19:40:14 UTC", "ActiveEnterTimestampMonotonic": "740184677", "ActiveExitTimestampMonotonic": "0", "ActiveState": "active", "After": "dbus-broker.service dbus.socket polkit.service basic.target sysinit.target system.slice", "AllowIsolate": "no", "AssertResult": "yes", "AssertTimestamp": "Sun 2024-02-04 19:40:13 UTC", "AssertTimestampMonotonic": "739023197", "Before": "network-pre.target shutdown.target multi-user.target", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "BusName": "org.fedoraproject.FirewallD1", "CPUAccounting": "yes", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "635480000", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "yes", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore", "CleanResult": "success", "CollectMode": "inactive", "ConditionResult": "yes", "ConditionTimestamp": "Sun 2024-02-04 19:40:13 UTC", "ConditionTimestampMonotonic": "739023194", "ConfigurationDirectoryMode": "0755", "Conflicts": "nftables.service iptables.service ebtables.service ipset.service ip6tables.service shutdown.target", "ControlGroup": "/system.slice/firewalld.service", "ControlGroupId": "6105", "ControlPID": "0", "CoredumpFilter": "0x33", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "DefaultStartupMemoryLow": "0", "Delegate": "no", "Description": "firewalld - dynamic firewall daemon", "DevicePolicy": "auto", "Documentation": "\"man:firewalld(1)\"", "DropInPaths": "/usr/lib/systemd/system/service.d/10-timeout-abort.conf", "DynamicUser": "no", "EffectiveCPUs": "0-1", "EffectiveMemoryNodes": "0", "EnvironmentFiles": "/etc/sysconfig/firewalld (ignore_errors=yes)", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "6625", "ExecMainStartTimestamp": "Sun 2024-02-04 19:40:13 UTC", "ExecMainStartTimestampMonotonic": "739034144", "ExecMainStatus": "0", "ExecReload": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecReloadEx": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStart": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStartEx": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExitType": "main", "ExtensionImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FileDescriptorStorePreserve": "restart", "FinalKillSignal": "9", "FragmentPath": "/usr/lib/systemd/system/firewalld.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOReadBytes": "18446744073709551615", "IOReadOperations": "18446744073709551615", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "IOWeight": "[not set]", "IOWriteBytes": "18446744073709551615", "IOWriteOperations": "18446744073709551615", "IPAccounting": "no", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "Id": "firewalld.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestamp": "Sun 2024-02-04 19:40:13 UTC", "InactiveExitTimestampMonotonic": "739034492", "InvocationID": "5b2d1a3e8acc4f9d82d144128c51d0be", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "mixed", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "infinity", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "8388608", "LimitMEMLOCKSoft": "8388608", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "524288", "LimitNOFILESoft": "1024", "LimitNPROC": "14778", "LimitNPROCSoft": "14778", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "14778", "LimitSIGPENDINGSoft": "14778", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "6625", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "ManagedOOMSwap": "auto", "MemoryAccounting": "yes", "MemoryAvailable": "infinity", "MemoryCurrent": "33116160", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryKSM": "no", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemoryPressureThresholdUSec": "200ms", "MemoryPressureWatch": "auto", "MemorySwapMax": "infinity", "MemoryZSwapMax": "infinity", "MountAPIVFS": "no", "MountImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAPolicy": "n/a", "Names": "firewalld.service dbus-org.fedoraproject.FirewallD1.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMPolicy": "stop", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "OnSuccessJobMode": "fail", "Perpetual": "no", "PrivateDevices": "no", "PrivateIPC": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProcSubset": "all", "ProtectClock": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectHostname": "no", "ProtectKernelLogs": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectProc": "default", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "ReloadResult": "success", "ReloadSignal": "1", "RemainAfterExit": "no", "RemoveIPC": "no", "Requires": "system.slice dbus.socket sysinit.target", "Restart": "no", "RestartKillSignal": "15", "RestartMaxDelayUSec": "infinity", "RestartMode": "normal", "RestartSteps": "0", "RestartUSec": "100ms", "RestartUSecNext": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RootEphemeral": "no", "RootImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "Slice": "system.slice", "StandardError": "null", "StandardInput": "null", "StandardOutput": "null", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StartupMemoryHigh": "18446744073709551615", "StartupMemoryLow": "0", "StartupMemoryMax": "18446744073709551615", "StartupMemorySwapMax": "18446744073709551615", "StartupMemoryZSwapMax": "18446744073709551615", "StateChangeTimestamp": "Sun 2024-02-04 19:40:38 UTC", "StateChangeTimestampMonotonic": "764242582", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "running", "SuccessAction": "none", "SyslogFacility": "3", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "2147483646", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "2", "TasksMax": "4433", "TimeoutAbortUSec": "45s", "TimeoutCleanUSec": "infinity", "TimeoutStartFailureMode": "terminate", "TimeoutStartUSec": "45s", "TimeoutStopFailureMode": "abort", "TimeoutStopUSec": "45s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "dbus", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "enabled", "UnitFileState": "enabled", "UtmpMode": "init", "WantedBy": "multi-user.target", "Wants": "network-pre.target", "WatchdogSignal": "6", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "0" } } TASK [fedora.linux_system_roles.firewall : Check if previous replaced is defined] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:34 Sunday 04 February 2024 19:40:46 +0000 (0:00:00.436) 0:02:18.206 ******* ok: [sut] => { "ansible_facts": { "__firewall_previous_replaced": false, "__firewall_python_cmd": "/usr/bin/python3", "__firewall_report_changed": true }, "changed": false } TASK [fedora.linux_system_roles.firewall : Get config files, checksums before and remove] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:43 Sunday 04 February 2024 19:40:46 +0000 (0:00:00.033) 0:02:18.240 ******* skipping: [sut] => { "changed": false, "false_condition": "__firewall_previous_replaced | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Tell firewall module it is able to report changed] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:55 Sunday 04 February 2024 19:40:46 +0000 (0:00:00.020) 0:02:18.260 ******* skipping: [sut] => { "changed": false, "false_condition": "__firewall_previous_replaced | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Configure firewall] ***************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:71 Sunday 04 February 2024 19:40:46 +0000 (0:00:00.056) 0:02:18.317 ******* ok: [sut] => (item={'port': '8000/tcp', 'state': 'enabled'}) => { "__firewall_changed": false, "ansible_loop_var": "item", "changed": false, "item": { "port": "8000/tcp", "state": "enabled" } } ok: [sut] => (item={'port': '9000/tcp', 'state': 'enabled'}) => { "__firewall_changed": false, "ansible_loop_var": "item", "changed": false, "item": { "port": "9000/tcp", "state": "enabled" } } TASK [fedora.linux_system_roles.firewall : Gather firewall config information] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:120 Sunday 04 February 2024 19:40:47 +0000 (0:00:00.792) 0:02:19.109 ******* skipping: [sut] => (item={'port': '8000/tcp', 'state': 'enabled'}) => { "ansible_loop_var": "item", "changed": false, "false_condition": "firewall | length == 1", "item": { "port": "8000/tcp", "state": "enabled" }, "skip_reason": "Conditional result was False" } skipping: [sut] => (item={'port': '9000/tcp', 'state': 'enabled'}) => { "ansible_loop_var": "item", "changed": false, "false_condition": "firewall | length == 1", "item": { "port": "9000/tcp", "state": "enabled" }, "skip_reason": "Conditional result was False" } skipping: [sut] => { "changed": false } MSG: All items skipped TASK [fedora.linux_system_roles.firewall : Update firewalld_config fact] ******* task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:130 Sunday 04 February 2024 19:40:47 +0000 (0:00:00.036) 0:02:19.146 ******* skipping: [sut] => { "changed": false, "false_condition": "firewall | length == 1", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Gather firewall config if no arguments] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:139 Sunday 04 February 2024 19:40:47 +0000 (0:00:00.025) 0:02:19.172 ******* skipping: [sut] => { "changed": false, "false_condition": "firewall == None or firewall | length == 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Update firewalld_config fact] ******* task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:144 Sunday 04 February 2024 19:40:47 +0000 (0:00:00.023) 0:02:19.196 ******* skipping: [sut] => { "changed": false, "false_condition": "firewall == None or firewall | length == 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Get config files, checksums after] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:153 Sunday 04 February 2024 19:40:47 +0000 (0:00:00.025) 0:02:19.221 ******* skipping: [sut] => { "changed": false, "false_condition": "__firewall_previous_replaced | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Calculate what has changed] ********* task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:163 Sunday 04 February 2024 19:40:47 +0000 (0:00:00.021) 0:02:19.242 ******* skipping: [sut] => { "changed": false, "false_condition": "__firewall_previous_replaced | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Show diffs] ************************* task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:169 Sunday 04 February 2024 19:40:47 +0000 (0:00:00.023) 0:02:19.266 ******* skipping: [sut] => { "false_condition": "__firewall_previous_replaced | bool" } TASK [Manage selinux for specified ports] ************************************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:92 Sunday 04 February 2024 19:40:47 +0000 (0:00:00.034) 0:02:19.301 ******* skipping: [sut] => { "changed": false, "false_condition": "podman_selinux_ports | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Keep track of users that need to cancel linger] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:99 Sunday 04 February 2024 19:40:47 +0000 (0:00:00.024) 0:02:19.325 ******* ok: [sut] => { "ansible_facts": { "__podman_cancel_user_linger": [] }, "changed": false } TASK [fedora.linux_system_roles.podman : Handle secrets] *********************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:103 Sunday 04 February 2024 19:40:47 +0000 (0:00:00.030) 0:02:19.356 ******* included: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml for sut => (item=(censored due to no_log)) included: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml for sut => (item=(censored due to no_log)) included: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml for sut => (item=(censored due to no_log)) TASK [fedora.linux_system_roles.podman : Set variables part 0] ***************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:3 Sunday 04 February 2024 19:40:47 +0000 (0:00:00.068) 0:02:19.424 ******* ok: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Set variables part 1] ***************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:10 Sunday 04 February 2024 19:40:47 +0000 (0:00:00.029) 0:02:19.453 ******* ok: [sut] => { "ansible_facts": { "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set variables part 2] ***************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:14 Sunday 04 February 2024 19:40:47 +0000 (0:00:00.026) 0:02:19.480 ******* ok: [sut] => { "ansible_facts": { "__podman_rootless": false, "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:20 Sunday 04 February 2024 19:40:47 +0000 (0:00:00.031) 0:02:19.511 ******* included: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for sut TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Sunday 04 February 2024 19:40:47 +0000 (0:00:00.068) 0:02:19.580 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Sunday 04 February 2024 19:40:47 +0000 (0:00:00.022) 0:02:19.602 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Sunday 04 February 2024 19:40:47 +0000 (0:00:00.022) 0:02:19.625 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Manage each secret] ******************* task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:25 Sunday 04 February 2024 19:40:47 +0000 (0:00:00.022) 0:02:19.647 ******* ok: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Set variables part 0] ***************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:3 Sunday 04 February 2024 19:40:48 +0000 (0:00:00.312) 0:02:19.959 ******* ok: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Set variables part 1] ***************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:10 Sunday 04 February 2024 19:40:48 +0000 (0:00:00.029) 0:02:19.989 ******* ok: [sut] => { "ansible_facts": { "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set variables part 2] ***************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:14 Sunday 04 February 2024 19:40:48 +0000 (0:00:00.025) 0:02:20.015 ******* ok: [sut] => { "ansible_facts": { "__podman_rootless": false, "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:20 Sunday 04 February 2024 19:40:48 +0000 (0:00:00.029) 0:02:20.044 ******* included: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for sut TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Sunday 04 February 2024 19:40:48 +0000 (0:00:00.038) 0:02:20.083 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Sunday 04 February 2024 19:40:48 +0000 (0:00:00.021) 0:02:20.104 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Sunday 04 February 2024 19:40:48 +0000 (0:00:00.019) 0:02:20.124 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Manage each secret] ******************* task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:25 Sunday 04 February 2024 19:40:48 +0000 (0:00:00.021) 0:02:20.145 ******* ok: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Set variables part 0] ***************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:3 Sunday 04 February 2024 19:40:48 +0000 (0:00:00.305) 0:02:20.451 ******* ok: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Set variables part 1] ***************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:10 Sunday 04 February 2024 19:40:48 +0000 (0:00:00.029) 0:02:20.481 ******* ok: [sut] => { "ansible_facts": { "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set variables part 2] ***************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:14 Sunday 04 February 2024 19:40:48 +0000 (0:00:00.026) 0:02:20.507 ******* ok: [sut] => { "ansible_facts": { "__podman_rootless": false, "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:20 Sunday 04 February 2024 19:40:48 +0000 (0:00:00.029) 0:02:20.537 ******* included: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for sut TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Sunday 04 February 2024 19:40:48 +0000 (0:00:00.035) 0:02:20.572 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Sunday 04 February 2024 19:40:48 +0000 (0:00:00.055) 0:02:20.627 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Sunday 04 February 2024 19:40:48 +0000 (0:00:00.022) 0:02:20.649 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Manage each secret] ******************* task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:25 Sunday 04 February 2024 19:40:48 +0000 (0:00:00.020) 0:02:20.670 ******* ok: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Handle Kubernetes specifications] ***** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:110 Sunday 04 February 2024 19:40:49 +0000 (0:00:00.299) 0:02:20.969 ******* skipping: [sut] => { "changed": false, "skipped_reason": "No items in the list" } TASK [fedora.linux_system_roles.podman : Handle Quadlet specifications] ******** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:116 Sunday 04 February 2024 19:40:49 +0000 (0:00:00.020) 0:02:20.990 ******* included: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for sut => (item={'file_src': 'quadlet-demo-mysql.volume', 'state': 'absent'}) included: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for sut => (item={'template_src': 'quadlet-demo-mysql.container.j2', 'state': 'absent'}) included: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for sut => (item={'file_src': 'envoy-proxy-configmap.yml', 'state': 'absent'}) included: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for sut => (item={'template_src': 'quadlet-demo.yml.j2', 'state': 'absent'}) included: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for sut => (item={'file_src': 'quadlet-demo.kube', 'state': 'absent'}) included: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for sut => (item={'file_src': 'quadlet-demo.network', 'state': 'absent'}) TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14 Sunday 04 February 2024 19:40:49 +0000 (0:00:00.104) 0:02:21.095 ******* ok: [sut] => { "ansible_facts": { "__podman_quadlet_file_src": "quadlet-demo-mysql.volume", "__podman_quadlet_spec": {}, "__podman_quadlet_str": "[Volume]", "__podman_quadlet_template_src": "" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25 Sunday 04 February 2024 19:40:49 +0000 (0:00:00.035) 0:02:21.130 ******* ok: [sut] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_pull_image": true, "__podman_state": "absent", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] ***** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35 Sunday 04 February 2024 19:40:49 +0000 (0:00:00.031) 0:02:21.162 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_quadlet_file_src", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48 Sunday 04 February 2024 19:40:49 +0000 (0:00:00.024) 0:02:21.186 ******* ok: [sut] => { "ansible_facts": { "__podman_quadlet_name": "quadlet-demo-mysql", "__podman_quadlet_type": "volume", "__podman_rootless": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57 Sunday 04 February 2024 19:40:49 +0000 (0:00:00.036) 0:02:21.223 ******* included: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for sut TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Sunday 04 February 2024 19:40:49 +0000 (0:00:00.043) 0:02:21.266 ******* skipping: [sut] => { "changed": false, "false_condition": "'getent_passwd' not in ansible_facts or __podman_user not in ansible_facts['getent_passwd']", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Sunday 04 February 2024 19:40:49 +0000 (0:00:00.027) 0:02:21.293 ******* skipping: [sut] => { "changed": false, "false_condition": "not ansible_facts[\"getent_passwd\"][__podman_user]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Sunday 04 February 2024 19:40:49 +0000 (0:00:00.026) 0:02:21.320 ******* ok: [sut] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get group information] **************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:27 Sunday 04 February 2024 19:40:49 +0000 (0:00:00.033) 0:02:21.353 ******* ok: [sut] => { "ansible_facts": { "getent_group": { "root": [ "x", "0", "" ] } }, "changed": false } TASK [fedora.linux_system_roles.podman : Set group name] *********************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:34 Sunday 04 February 2024 19:40:49 +0000 (0:00:00.256) 0:02:21.610 ******* ok: [sut] => { "ansible_facts": { "__podman_group_name": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:38 Sunday 04 February 2024 19:40:49 +0000 (0:00:00.062) 0:02:21.673 ******* ok: [sut] => { "changed": false, "stat": { "atime": 1707075608.5761826, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "72ba28a3be19c467e89ab90f5d21f99b605159a7", "ctime": 1707075555.3834877, "dev": 51714, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 203277, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-pie-executable", "mode": "0755", "mtime": 1696291200.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 16032, "uid": 0, "version": "1460241045", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************ task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:49 Sunday 04 February 2024 19:40:50 +0000 (0:00:00.256) 0:02:21.929 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check group with getsubids] *********** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:53 Sunday 04 February 2024 19:40:50 +0000 (0:00:00.022) 0:02:21.952 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if user is in subuid file] ****** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:63 Sunday 04 February 2024 19:40:50 +0000 (0:00:00.022) 0:02:21.974 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:71 Sunday 04 February 2024 19:40:50 +0000 (0:00:00.022) 0:02:21.996 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if group is in subgid file] ***** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:78 Sunday 04 February 2024 19:40:50 +0000 (0:00:00.023) 0:02:22.019 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] ***** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:86 Sunday 04 February 2024 19:40:50 +0000 (0:00:00.022) 0:02:22.042 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62 Sunday 04 February 2024 19:40:50 +0000 (0:00:00.022) 0:02:22.064 ******* ok: [sut] => { "ansible_facts": { "__podman_activate_systemd_unit": true, "__podman_images_found": [], "__podman_kube_yamls_raw": "", "__podman_service_name": "quadlet-demo-mysql-volume.service", "__podman_systemd_scope": "system", "__podman_user_home_dir": "/root", "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73 Sunday 04 February 2024 19:40:50 +0000 (0:00:00.043) 0:02:22.107 ******* ok: [sut] => { "ansible_facts": { "__podman_quadlet_path": "/etc/containers/systemd" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get kube yaml contents] *************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77 Sunday 04 February 2024 19:40:50 +0000 (0:00:00.024) 0:02:22.132 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_state != \"absent\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87 Sunday 04 February 2024 19:40:50 +0000 (0:00:00.021) 0:02:22.154 ******* ok: [sut] => { "ansible_facts": { "__podman_images": [], "__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo-mysql.volume", "__podman_volumes": [] }, "changed": false } TASK [fedora.linux_system_roles.podman : Cleanup quadlets] ********************* task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:103 Sunday 04 February 2024 19:40:50 +0000 (0:00:00.059) 0:02:22.214 ******* included: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml for sut TASK [fedora.linux_system_roles.podman : Stop and disable service] ************* task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:4 Sunday 04 February 2024 19:40:50 +0000 (0:00:00.048) 0:02:22.262 ******* changed: [sut] => { "changed": true, "enabled": false, "failed_when_result": false, "name": "quadlet-demo-mysql-volume.service", "state": "stopped", "status": { "AccessSELinuxContext": "system_u:object_r:systemd_unit_file_t:s0", "ActiveEnterTimestamp": "Sun 2024-02-04 19:40:25 UTC", "ActiveEnterTimestampMonotonic": "750950170", "ActiveExitTimestampMonotonic": "0", "ActiveState": "active", "After": "basic.target systemd-journald.socket -.mount system.slice sysinit.target", "AllowIsolate": "no", "AssertResult": "yes", "AssertTimestamp": "Sun 2024-02-04 19:40:24 UTC", "AssertTimestampMonotonic": "750727467", "Before": "shutdown.target quadlet-demo-mysql.service", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "CPUAccounting": "yes", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "74312000", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "no", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore", "CleanResult": "success", "CollectMode": "inactive", "ConditionResult": "yes", "ConditionTimestamp": "Sun 2024-02-04 19:40:24 UTC", "ConditionTimestampMonotonic": "750727463", "ConfigurationDirectoryMode": "0755", "Conflicts": "shutdown.target", "ControlGroupId": "0", "ControlPID": "0", "CoredumpFilter": "0x33", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "DefaultStartupMemoryLow": "0", "Delegate": "no", "Description": "quadlet-demo-mysql-volume.service", "DevicePolicy": "auto", "DropInPaths": "/usr/lib/systemd/system/service.d/10-timeout-abort.conf", "DynamicUser": "no", "ExecMainCode": "1", "ExecMainExitTimestamp": "Sun 2024-02-04 19:40:24 UTC", "ExecMainExitTimestampMonotonic": "750944902", "ExecMainPID": "7288", "ExecMainStartTimestamp": "Sun 2024-02-04 19:40:24 UTC", "ExecMainStartTimestampMonotonic": "750735154", "ExecMainStatus": "0", "ExecStart": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman volume create --ignore systemd-quadlet-demo-mysql ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStartEx": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman volume create --ignore systemd-quadlet-demo-mysql ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExitType": "main", "ExtensionImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FileDescriptorStorePreserve": "restart", "FinalKillSignal": "9", "FragmentPath": "/run/systemd/generator/quadlet-demo-mysql-volume.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOReadBytes": "18446744073709551615", "IOReadOperations": "18446744073709551615", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "IOWeight": "[not set]", "IOWriteBytes": "18446744073709551615", "IOWriteOperations": "18446744073709551615", "IPAccounting": "no", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "Id": "quadlet-demo-mysql-volume.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestamp": "Sun 2024-02-04 19:40:24 UTC", "InactiveExitTimestampMonotonic": "750735592", "InvocationID": "3ee141b7f9f6421aab16797edf4b2269", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "control-group", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "infinity", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "8388608", "LimitMEMLOCKSoft": "8388608", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "524288", "LimitNOFILESoft": "1024", "LimitNPROC": "14778", "LimitNPROCSoft": "14778", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "14778", "LimitSIGPENDINGSoft": "14778", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "0", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "ManagedOOMSwap": "auto", "MemoryAccounting": "yes", "MemoryAvailable": "infinity", "MemoryCurrent": "[not set]", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryKSM": "no", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemoryPressureThresholdUSec": "200ms", "MemoryPressureWatch": "auto", "MemorySwapMax": "infinity", "MemoryZSwapMax": "infinity", "MountAPIVFS": "no", "MountImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAPolicy": "n/a", "Names": "quadlet-demo-mysql-volume.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMPolicy": "stop", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "OnSuccessJobMode": "fail", "Perpetual": "no", "PrivateDevices": "no", "PrivateIPC": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProcSubset": "all", "ProtectClock": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectHostname": "no", "ProtectKernelLogs": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectProc": "default", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "ReloadResult": "success", "ReloadSignal": "1", "RemainAfterExit": "yes", "RemoveIPC": "no", "RequiredBy": "quadlet-demo-mysql.service", "Requires": "-.mount sysinit.target system.slice", "RequiresMountsFor": "/run/containers", "Restart": "no", "RestartKillSignal": "15", "RestartMaxDelayUSec": "infinity", "RestartMode": "normal", "RestartSteps": "0", "RestartUSec": "100ms", "RestartUSecNext": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RootEphemeral": "no", "RootImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "Slice": "system.slice", "StandardError": "inherit", "StandardInput": "null", "StandardOutput": "journal", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StartupMemoryHigh": "18446744073709551615", "StartupMemoryLow": "0", "StartupMemoryMax": "18446744073709551615", "StartupMemorySwapMax": "18446744073709551615", "StartupMemoryZSwapMax": "18446744073709551615", "StateChangeTimestamp": "Sun 2024-02-04 19:40:25 UTC", "StateChangeTimestampMonotonic": "750950170", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "exited", "SuccessAction": "none", "SyslogFacility": "3", "SyslogIdentifier": "quadlet-demo-mysql-volume", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "2147483646", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "[not set]", "TasksMax": "4433", "TimeoutAbortUSec": "45s", "TimeoutCleanUSec": "infinity", "TimeoutStartFailureMode": "terminate", "TimeoutStartUSec": "infinity", "TimeoutStopFailureMode": "abort", "TimeoutStopUSec": "45s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "oneshot", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "disabled", "UnitFileState": "generated", "UtmpMode": "init", "WatchdogSignal": "6", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "0" } } TASK [fedora.linux_system_roles.podman : Remove quadlet file] ****************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:22 Sunday 04 February 2024 19:40:51 +0000 (0:00:00.922) 0:02:23.185 ******* changed: [sut] => { "changed": true, "path": "/etc/containers/systemd/quadlet-demo-mysql.volume", "state": "absent" } TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:28 Sunday 04 February 2024 19:40:51 +0000 (0:00:00.259) 0:02:23.445 ******* included: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for sut TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Sunday 04 February 2024 19:40:51 +0000 (0:00:00.036) 0:02:23.481 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Sunday 04 February 2024 19:40:51 +0000 (0:00:00.051) 0:02:23.533 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Sunday 04 February 2024 19:40:51 +0000 (0:00:00.021) 0:02:23.554 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Reload systemctl] ********************* task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:36 Sunday 04 February 2024 19:40:51 +0000 (0:00:00.021) 0:02:23.576 ******* ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.podman : Prune images no longer in use] ******** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:45 Sunday 04 February 2024 19:40:52 +0000 (0:00:00.871) 0:02:24.447 ******* changed: [sut] => { "changed": true, "cmd": [ "podman", "image", "prune", "-f" ], "delta": "0:00:00.039258", "end": "2024-02-04 19:40:52.865337", "rc": 0, "start": "2024-02-04 19:40:52.826079" } TASK [fedora.linux_system_roles.podman : Remove network] *********************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:53 Sunday 04 February 2024 19:40:52 +0000 (0:00:00.293) 0:02:24.741 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_quadlet_type == \"network\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create and update quadlets] *********** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:107 Sunday 04 February 2024 19:40:52 +0000 (0:00:00.025) 0:02:24.767 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_state != \"absent\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14 Sunday 04 February 2024 19:40:52 +0000 (0:00:00.020) 0:02:24.787 ******* ok: [sut] => { "ansible_facts": { "__podman_quadlet_file_src": "", "__podman_quadlet_spec": {}, "__podman_quadlet_str": "[Install]\nWantedBy=default.target\n\n[Container]\nImage=quay.io/linux-system-roles/mysql:5.6\nContainerName=quadlet-demo-mysql\nVolume=quadlet-demo-mysql.volume:/var/lib/mysql\nVolume=/tmp/quadlet_demo:/var/lib/quadlet_demo:Z\nNetwork=quadlet-demo.network\nSecret=mysql-root-password-container,type=env,target=MYSQL_ROOT_PASSWORD\nHealthCmd=/usr/bin/true\nHealthOnFailure=kill\n", "__podman_quadlet_template_src": "quadlet-demo-mysql.container.j2" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25 Sunday 04 February 2024 19:40:53 +0000 (0:00:00.073) 0:02:24.861 ******* ok: [sut] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_pull_image": true, "__podman_state": "absent", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] ***** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35 Sunday 04 February 2024 19:40:53 +0000 (0:00:00.029) 0:02:24.890 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_quadlet_str", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48 Sunday 04 February 2024 19:40:53 +0000 (0:00:00.023) 0:02:24.913 ******* ok: [sut] => { "ansible_facts": { "__podman_quadlet_name": "quadlet-demo-mysql", "__podman_quadlet_type": "container", "__podman_rootless": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57 Sunday 04 February 2024 19:40:53 +0000 (0:00:00.035) 0:02:24.949 ******* included: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for sut TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Sunday 04 February 2024 19:40:53 +0000 (0:00:00.041) 0:02:24.991 ******* skipping: [sut] => { "changed": false, "false_condition": "'getent_passwd' not in ansible_facts or __podman_user not in ansible_facts['getent_passwd']", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Sunday 04 February 2024 19:40:53 +0000 (0:00:00.024) 0:02:25.016 ******* skipping: [sut] => { "changed": false, "false_condition": "not ansible_facts[\"getent_passwd\"][__podman_user]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Sunday 04 February 2024 19:40:53 +0000 (0:00:00.025) 0:02:25.042 ******* ok: [sut] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get group information] **************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:27 Sunday 04 February 2024 19:40:53 +0000 (0:00:00.031) 0:02:25.073 ******* ok: [sut] => { "ansible_facts": { "getent_group": { "root": [ "x", "0", "" ] } }, "changed": false } TASK [fedora.linux_system_roles.podman : Set group name] *********************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:34 Sunday 04 February 2024 19:40:53 +0000 (0:00:00.271) 0:02:25.345 ******* ok: [sut] => { "ansible_facts": { "__podman_group_name": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:38 Sunday 04 February 2024 19:40:53 +0000 (0:00:00.031) 0:02:25.377 ******* ok: [sut] => { "changed": false, "stat": { "atime": 1707075608.5761826, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "72ba28a3be19c467e89ab90f5d21f99b605159a7", "ctime": 1707075555.3834877, "dev": 51714, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 203277, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-pie-executable", "mode": "0755", "mtime": 1696291200.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 16032, "uid": 0, "version": "1460241045", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************ task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:49 Sunday 04 February 2024 19:40:53 +0000 (0:00:00.255) 0:02:25.632 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check group with getsubids] *********** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:53 Sunday 04 February 2024 19:40:53 +0000 (0:00:00.055) 0:02:25.687 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if user is in subuid file] ****** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:63 Sunday 04 February 2024 19:40:53 +0000 (0:00:00.021) 0:02:25.709 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:71 Sunday 04 February 2024 19:40:53 +0000 (0:00:00.021) 0:02:25.730 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if group is in subgid file] ***** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:78 Sunday 04 February 2024 19:40:53 +0000 (0:00:00.021) 0:02:25.751 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] ***** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:86 Sunday 04 February 2024 19:40:53 +0000 (0:00:00.020) 0:02:25.772 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62 Sunday 04 February 2024 19:40:53 +0000 (0:00:00.021) 0:02:25.793 ******* ok: [sut] => { "ansible_facts": { "__podman_activate_systemd_unit": true, "__podman_images_found": [ "quay.io/linux-system-roles/mysql:5.6" ], "__podman_kube_yamls_raw": "", "__podman_service_name": "quadlet-demo-mysql.service", "__podman_systemd_scope": "system", "__podman_user_home_dir": "/root", "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73 Sunday 04 February 2024 19:40:54 +0000 (0:00:00.046) 0:02:25.839 ******* ok: [sut] => { "ansible_facts": { "__podman_quadlet_path": "/etc/containers/systemd" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get kube yaml contents] *************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77 Sunday 04 February 2024 19:40:54 +0000 (0:00:00.025) 0:02:25.865 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_state != \"absent\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87 Sunday 04 February 2024 19:40:54 +0000 (0:00:00.020) 0:02:25.885 ******* ok: [sut] => { "ansible_facts": { "__podman_images": [ "quay.io/linux-system-roles/mysql:5.6" ], "__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo-mysql.container", "__podman_volumes": [ "/tmp/quadlet_demo" ] }, "changed": false } TASK [fedora.linux_system_roles.podman : Cleanup quadlets] ********************* task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:103 Sunday 04 February 2024 19:40:54 +0000 (0:00:00.059) 0:02:25.945 ******* included: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml for sut TASK [fedora.linux_system_roles.podman : Stop and disable service] ************* task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:4 Sunday 04 February 2024 19:40:54 +0000 (0:00:00.045) 0:02:25.990 ******* changed: [sut] => { "changed": true, "enabled": false, "failed_when_result": false, "name": "quadlet-demo-mysql.service", "state": "stopped", "status": { "AccessSELinuxContext": "system_u:object_r:systemd_unit_file_t:s0", "ActiveEnterTimestampMonotonic": "0", "ActiveExitTimestampMonotonic": "0", "ActiveState": "failed", "After": "tmp.mount quadlet-demo-mysql-volume.service basic.target sysinit.target system.slice -.mount systemd-journald.socket quadlet-demo-network.service", "AllowIsolate": "no", "AssertResult": "yes", "AssertTimestamp": "Sun 2024-02-04 19:40:38 UTC", "AssertTimestampMonotonic": "764737296", "Before": "shutdown.target multi-user.target", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "CPUAccounting": "yes", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "88616000", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "no", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore", "CleanResult": "success", "CollectMode": "inactive", "ConditionResult": "yes", "ConditionTimestamp": "Sun 2024-02-04 19:40:38 UTC", "ConditionTimestampMonotonic": "764737292", "ConfigurationDirectoryMode": "0755", "Conflicts": "shutdown.target", "ControlGroupId": "0", "ControlPID": "0", "CoredumpFilter": "0x33", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "DefaultStartupMemoryLow": "0", "Delegate": "yes", "DelegateControllers": "cpu cpuacct cpuset io blkio memory devices pids bpf-firewall bpf-devices bpf-foreign bpf-socket-bind bpf-restrict-network-interfaces", "Description": "quadlet-demo-mysql.service", "DevicePolicy": "auto", "DropInPaths": "/usr/lib/systemd/system/service.d/10-timeout-abort.conf", "DynamicUser": "no", "Environment": "PODMAN_SYSTEMD_UNIT=quadlet-demo-mysql.service", "ExecMainCode": "1", "ExecMainExitTimestamp": "Sun 2024-02-04 19:40:38 UTC", "ExecMainExitTimestampMonotonic": "764800104", "ExecMainPID": "7705", "ExecMainStartTimestamp": "Sun 2024-02-04 19:40:38 UTC", "ExecMainStartTimestampMonotonic": "764749181", "ExecMainStatus": "125", "ExecStart": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman run --name=quadlet-demo-mysql --cidfile=/run/quadlet-demo-mysql.cid --replace --rm --cgroups=split --network=systemd-quadlet-demo --sdnotify=conmon -d -v systemd-quadlet-demo-mysql:/var/lib/mysql -v /tmp/quadlet_demo:/var/lib/quadlet_demo:Z --secret mysql-root-password-container,type=env,target=MYSQL_ROOT_PASSWORD --health-cmd /usr/bin/true --health-on-failure kill quay.io/linux-system-roles/mysql:5.6 ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStartEx": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman run --name=quadlet-demo-mysql --cidfile=/run/quadlet-demo-mysql.cid --replace --rm --cgroups=split --network=systemd-quadlet-demo --sdnotify=conmon -d -v systemd-quadlet-demo-mysql:/var/lib/mysql -v /tmp/quadlet_demo:/var/lib/quadlet_demo:Z --secret mysql-root-password-container,type=env,target=MYSQL_ROOT_PASSWORD --health-cmd /usr/bin/true --health-on-failure kill quay.io/linux-system-roles/mysql:5.6 ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStop": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman rm -v -f -i --cidfile=/run/quadlet-demo-mysql.cid ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStopEx": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman rm -v -f -i --cidfile=/run/quadlet-demo-mysql.cid ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStopPost": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman rm -v -f -i --cidfile=/run/quadlet-demo-mysql.cid ; ignore_errors=yes ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStopPostEx": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman rm -v -f -i --cidfile=/run/quadlet-demo-mysql.cid ; flags=ignore-failure ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExitType": "main", "ExtensionImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FileDescriptorStorePreserve": "restart", "FinalKillSignal": "9", "FragmentPath": "/run/systemd/generator/quadlet-demo-mysql.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOReadBytes": "18446744073709551615", "IOReadOperations": "18446744073709551615", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "IOWeight": "[not set]", "IOWriteBytes": "18446744073709551615", "IOWriteOperations": "18446744073709551615", "IPAccounting": "no", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "Id": "quadlet-demo-mysql.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestamp": "Sun 2024-02-04 19:40:38 UTC", "InactiveEnterTimestampMonotonic": "764842302", "InactiveExitTimestamp": "Sun 2024-02-04 19:40:38 UTC", "InactiveExitTimestampMonotonic": "764749619", "InvocationID": "e4f076362da4402c9911ed0304c98cb4", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "mixed", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "infinity", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "8388608", "LimitMEMLOCKSoft": "8388608", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "524288", "LimitNOFILESoft": "1024", "LimitNPROC": "14778", "LimitNPROCSoft": "14778", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "14778", "LimitSIGPENDINGSoft": "14778", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "0", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "ManagedOOMSwap": "auto", "MemoryAccounting": "yes", "MemoryAvailable": "infinity", "MemoryCurrent": "[not set]", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryKSM": "no", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemoryPressureThresholdUSec": "200ms", "MemoryPressureWatch": "auto", "MemorySwapMax": "infinity", "MemoryZSwapMax": "infinity", "MountAPIVFS": "no", "MountImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAPolicy": "n/a", "Names": "quadlet-demo-mysql.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "all", "OOMPolicy": "continue", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "OnSuccessJobMode": "fail", "Perpetual": "no", "PrivateDevices": "no", "PrivateIPC": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProcSubset": "all", "ProtectClock": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectHostname": "no", "ProtectKernelLogs": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectProc": "default", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "ReloadResult": "success", "ReloadSignal": "1", "RemainAfterExit": "no", "RemoveIPC": "no", "Requires": "tmp.mount -.mount quadlet-demo-mysql-volume.service sysinit.target quadlet-demo-network.service system.slice", "RequiresMountsFor": "/tmp/quadlet_demo /run/containers", "Restart": "no", "RestartKillSignal": "15", "RestartMaxDelayUSec": "infinity", "RestartMode": "normal", "RestartSteps": "0", "RestartUSec": "100ms", "RestartUSecNext": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "exit-code", "RootDirectoryStartOnly": "no", "RootEphemeral": "no", "RootImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "Slice": "system.slice", "SourcePath": "/etc/containers/systemd/quadlet-demo-mysql.container", "StandardError": "inherit", "StandardInput": "null", "StandardOutput": "journal", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StartupMemoryHigh": "18446744073709551615", "StartupMemoryLow": "0", "StartupMemoryMax": "18446744073709551615", "StartupMemorySwapMax": "18446744073709551615", "StartupMemoryZSwapMax": "18446744073709551615", "StateChangeTimestamp": "Sun 2024-02-04 19:40:38 UTC", "StateChangeTimestampMonotonic": "764842302", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "failed", "SuccessAction": "none", "SyslogFacility": "3", "SyslogIdentifier": "quadlet-demo-mysql", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "2147483646", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "[not set]", "TasksMax": "4433", "TimeoutAbortUSec": "45s", "TimeoutCleanUSec": "infinity", "TimeoutStartFailureMode": "terminate", "TimeoutStartUSec": "45s", "TimeoutStopFailureMode": "abort", "TimeoutStopUSec": "45s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "notify", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "disabled", "UnitFileState": "generated", "UtmpMode": "init", "WantedBy": "multi-user.target", "WatchdogSignal": "6", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "0" } } TASK [fedora.linux_system_roles.podman : Remove quadlet file] ****************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:22 Sunday 04 February 2024 19:40:55 +0000 (0:00:00.909) 0:02:26.899 ******* changed: [sut] => { "changed": true, "path": "/etc/containers/systemd/quadlet-demo-mysql.container", "state": "absent" } TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:28 Sunday 04 February 2024 19:40:55 +0000 (0:00:00.259) 0:02:27.159 ******* included: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for sut TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Sunday 04 February 2024 19:40:55 +0000 (0:00:00.035) 0:02:27.194 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Sunday 04 February 2024 19:40:55 +0000 (0:00:00.021) 0:02:27.215 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Sunday 04 February 2024 19:40:55 +0000 (0:00:00.020) 0:02:27.235 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Reload systemctl] ********************* task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:36 Sunday 04 February 2024 19:40:55 +0000 (0:00:00.052) 0:02:27.288 ******* ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.podman : Prune images no longer in use] ******** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:45 Sunday 04 February 2024 19:40:56 +0000 (0:00:00.933) 0:02:28.221 ******* changed: [sut] => { "changed": true, "cmd": [ "podman", "image", "prune", "-f" ], "delta": "0:00:00.038702", "end": "2024-02-04 19:40:56.640330", "rc": 0, "start": "2024-02-04 19:40:56.601628" } TASK [fedora.linux_system_roles.podman : Remove network] *********************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:53 Sunday 04 February 2024 19:40:56 +0000 (0:00:00.295) 0:02:28.517 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_quadlet_type == \"network\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create and update quadlets] *********** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:107 Sunday 04 February 2024 19:40:56 +0000 (0:00:00.024) 0:02:28.541 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_state != \"absent\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14 Sunday 04 February 2024 19:40:56 +0000 (0:00:00.021) 0:02:28.562 ******* ok: [sut] => { "ansible_facts": { "__podman_quadlet_file_src": "envoy-proxy-configmap.yml", "__podman_quadlet_spec": {}, "__podman_quadlet_str": "---\napiVersion: v1\nkind: ConfigMap\nmetadata:\n name: envoy-proxy-config\ndata:\n envoy.yaml: |\n admin:\n address:\n socket_address:\n address: 0.0.0.0\n port_value: 9901\n\n static_resources:\n listeners:\n - name: listener_0\n address:\n socket_address:\n address: 0.0.0.0\n port_value: 8080\n filter_chains:\n - filters:\n - name: envoy.filters.network.http_connection_manager\n typed_config:\n \"@type\": type.googleapis.com/envoy.extensions.filters.network.http_connection_manager.v3.HttpConnectionManager\n stat_prefix: ingress_http\n codec_type: AUTO\n route_config:\n name: local_route\n virtual_hosts:\n - name: local_service\n domains: [\"*\"]\n routes:\n - match:\n prefix: \"/\"\n route:\n cluster: backend\n http_filters:\n - name: envoy.filters.http.router\n typed_config:\n \"@type\": type.googleapis.com/envoy.extensions.filters.http.router.v3.Router\n transport_socket:\n name: envoy.transport_sockets.tls\n typed_config:\n \"@type\": type.googleapis.com/envoy.extensions.transport_sockets.tls.v3.DownstreamTlsContext\n common_tls_context:\n tls_certificates:\n - certificate_chain:\n filename: /etc/envoy-certificates/certificate.pem\n private_key:\n filename: /etc/envoy-certificates/certificate.key\n clusters:\n - name: backend\n connect_timeout: 5s\n type: STATIC\n dns_refresh_rate: 1800s\n lb_policy: ROUND_ROBIN\n load_assignment:\n cluster_name: backend\n endpoints:\n - lb_endpoints:\n - endpoint:\n address:\n socket_address:\n address: 127.0.0.1\n port_value: 80", "__podman_quadlet_template_src": "" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25 Sunday 04 February 2024 19:40:56 +0000 (0:00:00.036) 0:02:28.599 ******* ok: [sut] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_pull_image": true, "__podman_state": "absent", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] ***** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35 Sunday 04 February 2024 19:40:56 +0000 (0:00:00.030) 0:02:28.629 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_quadlet_file_src", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48 Sunday 04 February 2024 19:40:56 +0000 (0:00:00.023) 0:02:28.653 ******* ok: [sut] => { "ansible_facts": { "__podman_quadlet_name": "envoy-proxy-configmap", "__podman_quadlet_type": "yml", "__podman_rootless": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57 Sunday 04 February 2024 19:40:56 +0000 (0:00:00.035) 0:02:28.688 ******* included: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for sut TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Sunday 04 February 2024 19:40:56 +0000 (0:00:00.042) 0:02:28.731 ******* skipping: [sut] => { "changed": false, "false_condition": "'getent_passwd' not in ansible_facts or __podman_user not in ansible_facts['getent_passwd']", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Sunday 04 February 2024 19:40:56 +0000 (0:00:00.025) 0:02:28.757 ******* skipping: [sut] => { "changed": false, "false_condition": "not ansible_facts[\"getent_passwd\"][__podman_user]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Sunday 04 February 2024 19:40:56 +0000 (0:00:00.025) 0:02:28.782 ******* ok: [sut] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get group information] **************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:27 Sunday 04 February 2024 19:40:56 +0000 (0:00:00.031) 0:02:28.813 ******* ok: [sut] => { "ansible_facts": { "getent_group": { "root": [ "x", "0", "" ] } }, "changed": false } TASK [fedora.linux_system_roles.podman : Set group name] *********************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:34 Sunday 04 February 2024 19:40:57 +0000 (0:00:00.254) 0:02:29.068 ******* ok: [sut] => { "ansible_facts": { "__podman_group_name": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:38 Sunday 04 February 2024 19:40:57 +0000 (0:00:00.031) 0:02:29.099 ******* ok: [sut] => { "changed": false, "stat": { "atime": 1707075608.5761826, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "72ba28a3be19c467e89ab90f5d21f99b605159a7", "ctime": 1707075555.3834877, "dev": 51714, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 203277, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-pie-executable", "mode": "0755", "mtime": 1696291200.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 16032, "uid": 0, "version": "1460241045", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************ task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:49 Sunday 04 February 2024 19:40:57 +0000 (0:00:00.254) 0:02:29.354 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check group with getsubids] *********** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:53 Sunday 04 February 2024 19:40:57 +0000 (0:00:00.022) 0:02:29.376 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if user is in subuid file] ****** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:63 Sunday 04 February 2024 19:40:57 +0000 (0:00:00.019) 0:02:29.396 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:71 Sunday 04 February 2024 19:40:57 +0000 (0:00:00.058) 0:02:29.454 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if group is in subgid file] ***** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:78 Sunday 04 February 2024 19:40:57 +0000 (0:00:00.023) 0:02:29.478 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] ***** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:86 Sunday 04 February 2024 19:40:57 +0000 (0:00:00.022) 0:02:29.501 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62 Sunday 04 February 2024 19:40:57 +0000 (0:00:00.021) 0:02:29.522 ******* ok: [sut] => { "ansible_facts": { "__podman_activate_systemd_unit": true, "__podman_images_found": [], "__podman_kube_yamls_raw": "", "__podman_service_name": "", "__podman_systemd_scope": "system", "__podman_user_home_dir": "/root", "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73 Sunday 04 February 2024 19:40:57 +0000 (0:00:00.044) 0:02:29.566 ******* ok: [sut] => { "ansible_facts": { "__podman_quadlet_path": "/etc/containers/systemd" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get kube yaml contents] *************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77 Sunday 04 February 2024 19:40:57 +0000 (0:00:00.026) 0:02:29.593 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_state != \"absent\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87 Sunday 04 February 2024 19:40:57 +0000 (0:00:00.021) 0:02:29.615 ******* ok: [sut] => { "ansible_facts": { "__podman_images": [], "__podman_quadlet_file": "/etc/containers/systemd/envoy-proxy-configmap.yml", "__podman_volumes": [] }, "changed": false } TASK [fedora.linux_system_roles.podman : Cleanup quadlets] ********************* task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:103 Sunday 04 February 2024 19:40:57 +0000 (0:00:00.059) 0:02:29.675 ******* included: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml for sut TASK [fedora.linux_system_roles.podman : Stop and disable service] ************* task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:4 Sunday 04 February 2024 19:40:57 +0000 (0:00:00.046) 0:02:29.721 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_service_name | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Remove quadlet file] ****************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:22 Sunday 04 February 2024 19:40:57 +0000 (0:00:00.023) 0:02:29.745 ******* ok: [sut] => { "changed": false, "path": "/etc/containers/systemd/envoy-proxy-configmap.yml", "state": "absent" } TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:28 Sunday 04 February 2024 19:40:58 +0000 (0:00:00.257) 0:02:30.002 ******* included: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for sut TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Sunday 04 February 2024 19:40:58 +0000 (0:00:00.036) 0:02:30.039 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Sunday 04 February 2024 19:40:58 +0000 (0:00:00.021) 0:02:30.060 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Sunday 04 February 2024 19:40:58 +0000 (0:00:00.021) 0:02:30.082 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Reload systemctl] ********************* task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:36 Sunday 04 February 2024 19:40:58 +0000 (0:00:00.020) 0:02:30.103 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_file_removed is changed", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Prune images no longer in use] ******** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:45 Sunday 04 February 2024 19:40:58 +0000 (0:00:00.024) 0:02:30.127 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_file_removed is changed", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Remove network] *********************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:53 Sunday 04 February 2024 19:40:58 +0000 (0:00:00.022) 0:02:30.149 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_file_removed is changed", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create and update quadlets] *********** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:107 Sunday 04 February 2024 19:40:58 +0000 (0:00:00.057) 0:02:30.207 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_state != \"absent\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14 Sunday 04 February 2024 19:40:58 +0000 (0:00:00.030) 0:02:30.237 ******* ok: [sut] => { "ansible_facts": { "__podman_quadlet_file_src": "", "__podman_quadlet_spec": {}, "__podman_quadlet_str": "---\napiVersion: v1\nkind: PersistentVolumeClaim\nmetadata:\n name: wp-pv-claim\n labels:\n app: wordpress\nspec:\n accessModes:\n - ReadWriteOnce\n resources:\n requests:\n storage: 20Gi\n---\napiVersion: v1\nkind: Pod\nmetadata:\n name: quadlet-demo\nspec:\n containers:\n - name: wordpress\n image: quay.io/linux-system-roles/wordpress:4.8-apache\n env:\n - name: WORDPRESS_DB_HOST\n value: quadlet-demo-mysql\n - name: WORDPRESS_DB_PASSWORD\n valueFrom:\n secretKeyRef:\n name: mysql-root-password-kube\n key: password\n volumeMounts:\n - name: wordpress-persistent-storage\n mountPath: /var/www/html\n resources:\n requests:\n memory: \"64Mi\"\n cpu: \"250m\"\n limits:\n memory: \"128Mi\"\n cpu: \"500m\"\n - name: envoy\n image: quay.io/linux-system-roles/envoyproxy:v1.25.0\n volumeMounts:\n - name: config-volume\n mountPath: /etc/envoy\n - name: certificates\n mountPath: /etc/envoy-certificates\n env:\n - name: ENVOY_UID\n value: \"0\"\n resources:\n requests:\n memory: \"64Mi\"\n cpu: \"250m\"\n limits:\n memory: \"128Mi\"\n cpu: \"500m\"\n volumes:\n - name: config-volume\n configMap:\n name: envoy-proxy-config\n - name: certificates\n secret:\n secretName: envoy-certificates\n - name: wordpress-persistent-storage\n persistentVolumeClaim:\n claimName: wp-pv-claim\n - name: www # not used - for testing hostpath\n hostPath:\n path: /tmp/httpd3\n - name: create # not used - for testing hostpath\n hostPath:\n path: /tmp/httpd3-create\n", "__podman_quadlet_template_src": "quadlet-demo.yml.j2" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25 Sunday 04 February 2024 19:40:58 +0000 (0:00:00.072) 0:02:30.309 ******* ok: [sut] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_pull_image": true, "__podman_state": "absent", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] ***** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35 Sunday 04 February 2024 19:40:58 +0000 (0:00:00.032) 0:02:30.341 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_quadlet_str", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48 Sunday 04 February 2024 19:40:58 +0000 (0:00:00.035) 0:02:30.377 ******* ok: [sut] => { "ansible_facts": { "__podman_quadlet_name": "quadlet-demo", "__podman_quadlet_type": "yml", "__podman_rootless": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57 Sunday 04 February 2024 19:40:58 +0000 (0:00:00.038) 0:02:30.415 ******* included: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for sut TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Sunday 04 February 2024 19:40:58 +0000 (0:00:00.043) 0:02:30.459 ******* skipping: [sut] => { "changed": false, "false_condition": "'getent_passwd' not in ansible_facts or __podman_user not in ansible_facts['getent_passwd']", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Sunday 04 February 2024 19:40:58 +0000 (0:00:00.027) 0:02:30.486 ******* skipping: [sut] => { "changed": false, "false_condition": "not ansible_facts[\"getent_passwd\"][__podman_user]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Sunday 04 February 2024 19:40:58 +0000 (0:00:00.026) 0:02:30.512 ******* ok: [sut] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get group information] **************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:27 Sunday 04 February 2024 19:40:58 +0000 (0:00:00.032) 0:02:30.545 ******* ok: [sut] => { "ansible_facts": { "getent_group": { "root": [ "x", "0", "" ] } }, "changed": false } TASK [fedora.linux_system_roles.podman : Set group name] *********************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:34 Sunday 04 February 2024 19:40:58 +0000 (0:00:00.257) 0:02:30.802 ******* ok: [sut] => { "ansible_facts": { "__podman_group_name": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:38 Sunday 04 February 2024 19:40:58 +0000 (0:00:00.033) 0:02:30.836 ******* ok: [sut] => { "changed": false, "stat": { "atime": 1707075608.5761826, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "72ba28a3be19c467e89ab90f5d21f99b605159a7", "ctime": 1707075555.3834877, "dev": 51714, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 203277, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-pie-executable", "mode": "0755", "mtime": 1696291200.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 16032, "uid": 0, "version": "1460241045", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************ task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:49 Sunday 04 February 2024 19:40:59 +0000 (0:00:00.279) 0:02:31.115 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check group with getsubids] *********** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:53 Sunday 04 February 2024 19:40:59 +0000 (0:00:00.021) 0:02:31.136 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if user is in subuid file] ****** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:63 Sunday 04 February 2024 19:40:59 +0000 (0:00:00.021) 0:02:31.158 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:71 Sunday 04 February 2024 19:40:59 +0000 (0:00:00.021) 0:02:31.179 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if group is in subgid file] ***** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:78 Sunday 04 February 2024 19:40:59 +0000 (0:00:00.021) 0:02:31.200 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] ***** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:86 Sunday 04 February 2024 19:40:59 +0000 (0:00:00.021) 0:02:31.222 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62 Sunday 04 February 2024 19:40:59 +0000 (0:00:00.057) 0:02:31.280 ******* ok: [sut] => { "ansible_facts": { "__podman_activate_systemd_unit": true, "__podman_images_found": [], "__podman_kube_yamls_raw": "", "__podman_service_name": "", "__podman_systemd_scope": "system", "__podman_user_home_dir": "/root", "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73 Sunday 04 February 2024 19:40:59 +0000 (0:00:00.043) 0:02:31.324 ******* ok: [sut] => { "ansible_facts": { "__podman_quadlet_path": "/etc/containers/systemd" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get kube yaml contents] *************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77 Sunday 04 February 2024 19:40:59 +0000 (0:00:00.025) 0:02:31.349 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_state != \"absent\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87 Sunday 04 February 2024 19:40:59 +0000 (0:00:00.022) 0:02:31.372 ******* ok: [sut] => { "ansible_facts": { "__podman_images": [], "__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo.yml", "__podman_volumes": [] }, "changed": false } TASK [fedora.linux_system_roles.podman : Cleanup quadlets] ********************* task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:103 Sunday 04 February 2024 19:40:59 +0000 (0:00:00.070) 0:02:31.443 ******* included: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml for sut TASK [fedora.linux_system_roles.podman : Stop and disable service] ************* task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:4 Sunday 04 February 2024 19:40:59 +0000 (0:00:00.047) 0:02:31.490 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_service_name | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Remove quadlet file] ****************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:22 Sunday 04 February 2024 19:40:59 +0000 (0:00:00.025) 0:02:31.515 ******* ok: [sut] => { "changed": false, "path": "/etc/containers/systemd/quadlet-demo.yml", "state": "absent" } TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:28 Sunday 04 February 2024 19:40:59 +0000 (0:00:00.262) 0:02:31.778 ******* included: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for sut TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Sunday 04 February 2024 19:40:59 +0000 (0:00:00.036) 0:02:31.815 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Sunday 04 February 2024 19:40:59 +0000 (0:00:00.022) 0:02:31.838 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Sunday 04 February 2024 19:41:00 +0000 (0:00:00.021) 0:02:31.859 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Reload systemctl] ********************* task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:36 Sunday 04 February 2024 19:41:00 +0000 (0:00:00.022) 0:02:31.881 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_file_removed is changed", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Prune images no longer in use] ******** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:45 Sunday 04 February 2024 19:41:00 +0000 (0:00:00.023) 0:02:31.904 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_file_removed is changed", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Remove network] *********************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:53 Sunday 04 February 2024 19:41:00 +0000 (0:00:00.023) 0:02:31.927 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_file_removed is changed", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create and update quadlets] *********** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:107 Sunday 04 February 2024 19:41:00 +0000 (0:00:00.023) 0:02:31.951 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_state != \"absent\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14 Sunday 04 February 2024 19:41:00 +0000 (0:00:00.021) 0:02:31.972 ******* ok: [sut] => { "ansible_facts": { "__podman_quadlet_file_src": "quadlet-demo.kube", "__podman_quadlet_spec": {}, "__podman_quadlet_str": "[Install]\nWantedBy=default.target\n\n[Unit]\nRequires=quadlet-demo-mysql.service\nAfter=quadlet-demo-mysql.service\n\n[Kube]\n# Point to the yaml file in the same directory\nYaml=quadlet-demo.yml\n# Use the quadlet-demo network\nNetwork=quadlet-demo.network\n# Publish the envoy proxy data port\nPublishPort=8000:8080\n# Publish the envoy proxy admin port\nPublishPort=9000:9901\n# Use the envoy proxy config map in the same directory\nConfigMap=envoy-proxy-configmap.yml", "__podman_quadlet_template_src": "" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25 Sunday 04 February 2024 19:41:00 +0000 (0:00:00.069) 0:02:32.042 ******* ok: [sut] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_pull_image": true, "__podman_state": "absent", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] ***** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35 Sunday 04 February 2024 19:41:00 +0000 (0:00:00.031) 0:02:32.073 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_quadlet_file_src", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48 Sunday 04 February 2024 19:41:00 +0000 (0:00:00.024) 0:02:32.098 ******* ok: [sut] => { "ansible_facts": { "__podman_quadlet_name": "quadlet-demo", "__podman_quadlet_type": "kube", "__podman_rootless": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57 Sunday 04 February 2024 19:41:00 +0000 (0:00:00.037) 0:02:32.135 ******* included: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for sut TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Sunday 04 February 2024 19:41:00 +0000 (0:00:00.044) 0:02:32.179 ******* skipping: [sut] => { "changed": false, "false_condition": "'getent_passwd' not in ansible_facts or __podman_user not in ansible_facts['getent_passwd']", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Sunday 04 February 2024 19:41:00 +0000 (0:00:00.027) 0:02:32.207 ******* skipping: [sut] => { "changed": false, "false_condition": "not ansible_facts[\"getent_passwd\"][__podman_user]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Sunday 04 February 2024 19:41:00 +0000 (0:00:00.027) 0:02:32.234 ******* ok: [sut] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get group information] **************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:27 Sunday 04 February 2024 19:41:00 +0000 (0:00:00.032) 0:02:32.266 ******* ok: [sut] => { "ansible_facts": { "getent_group": { "root": [ "x", "0", "" ] } }, "changed": false } TASK [fedora.linux_system_roles.podman : Set group name] *********************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:34 Sunday 04 February 2024 19:41:00 +0000 (0:00:00.264) 0:02:32.531 ******* ok: [sut] => { "ansible_facts": { "__podman_group_name": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:38 Sunday 04 February 2024 19:41:00 +0000 (0:00:00.034) 0:02:32.565 ******* ok: [sut] => { "changed": false, "stat": { "atime": 1707075608.5761826, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "72ba28a3be19c467e89ab90f5d21f99b605159a7", "ctime": 1707075555.3834877, "dev": 51714, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 203277, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-pie-executable", "mode": "0755", "mtime": 1696291200.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 16032, "uid": 0, "version": "1460241045", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************ task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:49 Sunday 04 February 2024 19:41:00 +0000 (0:00:00.262) 0:02:32.827 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check group with getsubids] *********** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:53 Sunday 04 February 2024 19:41:01 +0000 (0:00:00.023) 0:02:32.851 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if user is in subuid file] ****** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:63 Sunday 04 February 2024 19:41:01 +0000 (0:00:00.021) 0:02:32.872 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:71 Sunday 04 February 2024 19:41:01 +0000 (0:00:00.022) 0:02:32.895 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if group is in subgid file] ***** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:78 Sunday 04 February 2024 19:41:01 +0000 (0:00:00.022) 0:02:32.918 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] ***** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:86 Sunday 04 February 2024 19:41:01 +0000 (0:00:00.022) 0:02:32.940 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62 Sunday 04 February 2024 19:41:01 +0000 (0:00:00.021) 0:02:32.962 ******* ok: [sut] => { "ansible_facts": { "__podman_activate_systemd_unit": true, "__podman_images_found": [], "__podman_kube_yamls_raw": [ "quadlet-demo.yml" ], "__podman_service_name": "quadlet-demo.service", "__podman_systemd_scope": "system", "__podman_user_home_dir": "/root", "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73 Sunday 04 February 2024 19:41:01 +0000 (0:00:00.040) 0:02:33.002 ******* ok: [sut] => { "ansible_facts": { "__podman_quadlet_path": "/etc/containers/systemd" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get kube yaml contents] *************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77 Sunday 04 February 2024 19:41:01 +0000 (0:00:00.058) 0:02:33.061 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_state != \"absent\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87 Sunday 04 February 2024 19:41:01 +0000 (0:00:00.021) 0:02:33.083 ******* ok: [sut] => { "ansible_facts": { "__podman_images": [], "__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo.kube", "__podman_volumes": [] }, "changed": false } TASK [fedora.linux_system_roles.podman : Cleanup quadlets] ********************* task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:103 Sunday 04 February 2024 19:41:01 +0000 (0:00:00.058) 0:02:33.141 ******* included: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml for sut TASK [fedora.linux_system_roles.podman : Stop and disable service] ************* task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:4 Sunday 04 February 2024 19:41:01 +0000 (0:00:00.046) 0:02:33.188 ******* ok: [sut] => { "changed": false, "failed_when_result": false } MSG: Could not find the requested service quadlet-demo.service: host TASK [fedora.linux_system_roles.podman : Remove quadlet file] ****************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:22 Sunday 04 February 2024 19:41:01 +0000 (0:00:00.436) 0:02:33.624 ******* ok: [sut] => { "changed": false, "path": "/etc/containers/systemd/quadlet-demo.kube", "state": "absent" } TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:28 Sunday 04 February 2024 19:41:02 +0000 (0:00:00.261) 0:02:33.886 ******* included: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for sut TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Sunday 04 February 2024 19:41:02 +0000 (0:00:00.037) 0:02:33.923 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Sunday 04 February 2024 19:41:02 +0000 (0:00:00.022) 0:02:33.946 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Sunday 04 February 2024 19:41:02 +0000 (0:00:00.023) 0:02:33.970 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Reload systemctl] ********************* task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:36 Sunday 04 February 2024 19:41:02 +0000 (0:00:00.023) 0:02:33.994 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_file_removed is changed", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Prune images no longer in use] ******** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:45 Sunday 04 February 2024 19:41:02 +0000 (0:00:00.024) 0:02:34.018 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_file_removed is changed", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Remove network] *********************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:53 Sunday 04 February 2024 19:41:02 +0000 (0:00:00.024) 0:02:34.042 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_file_removed is changed", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create and update quadlets] *********** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:107 Sunday 04 February 2024 19:41:02 +0000 (0:00:00.025) 0:02:34.068 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_state != \"absent\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14 Sunday 04 February 2024 19:41:02 +0000 (0:00:00.021) 0:02:34.090 ******* ok: [sut] => { "ansible_facts": { "__podman_quadlet_file_src": "quadlet-demo.network", "__podman_quadlet_spec": {}, "__podman_quadlet_str": "[Network]\nSubnet=192.168.30.0/24\nGateway=192.168.30.1\nLabel=app=wordpress", "__podman_quadlet_template_src": "" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25 Sunday 04 February 2024 19:41:02 +0000 (0:00:00.033) 0:02:34.123 ******* ok: [sut] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_pull_image": true, "__podman_state": "absent", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] ***** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35 Sunday 04 February 2024 19:41:02 +0000 (0:00:00.030) 0:02:34.153 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_quadlet_file_src", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48 Sunday 04 February 2024 19:41:02 +0000 (0:00:00.023) 0:02:34.177 ******* ok: [sut] => { "ansible_facts": { "__podman_quadlet_name": "quadlet-demo", "__podman_quadlet_type": "network", "__podman_rootless": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57 Sunday 04 February 2024 19:41:02 +0000 (0:00:00.072) 0:02:34.249 ******* included: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for sut TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Sunday 04 February 2024 19:41:02 +0000 (0:00:00.047) 0:02:34.297 ******* skipping: [sut] => { "changed": false, "false_condition": "'getent_passwd' not in ansible_facts or __podman_user not in ansible_facts['getent_passwd']", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Sunday 04 February 2024 19:41:02 +0000 (0:00:00.027) 0:02:34.324 ******* skipping: [sut] => { "changed": false, "false_condition": "not ansible_facts[\"getent_passwd\"][__podman_user]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Sunday 04 February 2024 19:41:02 +0000 (0:00:00.027) 0:02:34.352 ******* ok: [sut] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get group information] **************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:27 Sunday 04 February 2024 19:41:02 +0000 (0:00:00.033) 0:02:34.385 ******* ok: [sut] => { "ansible_facts": { "getent_group": { "root": [ "x", "0", "" ] } }, "changed": false } TASK [fedora.linux_system_roles.podman : Set group name] *********************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:34 Sunday 04 February 2024 19:41:02 +0000 (0:00:00.259) 0:02:34.645 ******* ok: [sut] => { "ansible_facts": { "__podman_group_name": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:38 Sunday 04 February 2024 19:41:02 +0000 (0:00:00.031) 0:02:34.676 ******* ok: [sut] => { "changed": false, "stat": { "atime": 1707075608.5761826, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "72ba28a3be19c467e89ab90f5d21f99b605159a7", "ctime": 1707075555.3834877, "dev": 51714, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 203277, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-pie-executable", "mode": "0755", "mtime": 1696291200.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 16032, "uid": 0, "version": "1460241045", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************ task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:49 Sunday 04 February 2024 19:41:03 +0000 (0:00:00.253) 0:02:34.929 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check group with getsubids] *********** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:53 Sunday 04 February 2024 19:41:03 +0000 (0:00:00.021) 0:02:34.951 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if user is in subuid file] ****** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:63 Sunday 04 February 2024 19:41:03 +0000 (0:00:00.020) 0:02:34.972 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:71 Sunday 04 February 2024 19:41:03 +0000 (0:00:00.020) 0:02:34.993 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if group is in subgid file] ***** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:78 Sunday 04 February 2024 19:41:03 +0000 (0:00:00.020) 0:02:35.013 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] ***** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:86 Sunday 04 February 2024 19:41:03 +0000 (0:00:00.022) 0:02:35.036 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62 Sunday 04 February 2024 19:41:03 +0000 (0:00:00.020) 0:02:35.056 ******* ok: [sut] => { "ansible_facts": { "__podman_activate_systemd_unit": true, "__podman_images_found": [], "__podman_kube_yamls_raw": "", "__podman_service_name": "", "__podman_systemd_scope": "system", "__podman_user_home_dir": "/root", "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73 Sunday 04 February 2024 19:41:03 +0000 (0:00:00.043) 0:02:35.099 ******* ok: [sut] => { "ansible_facts": { "__podman_quadlet_path": "/etc/containers/systemd" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get kube yaml contents] *************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77 Sunday 04 February 2024 19:41:03 +0000 (0:00:00.023) 0:02:35.123 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_state != \"absent\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87 Sunday 04 February 2024 19:41:03 +0000 (0:00:00.020) 0:02:35.143 ******* ok: [sut] => { "ansible_facts": { "__podman_images": [], "__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo.network", "__podman_volumes": [] }, "changed": false } TASK [fedora.linux_system_roles.podman : Cleanup quadlets] ********************* task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:103 Sunday 04 February 2024 19:41:03 +0000 (0:00:00.091) 0:02:35.235 ******* included: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml for sut TASK [fedora.linux_system_roles.podman : Stop and disable service] ************* task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:4 Sunday 04 February 2024 19:41:03 +0000 (0:00:00.047) 0:02:35.282 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_service_name | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Remove quadlet file] ****************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:22 Sunday 04 February 2024 19:41:03 +0000 (0:00:00.024) 0:02:35.307 ******* changed: [sut] => { "changed": true, "path": "/etc/containers/systemd/quadlet-demo.network", "state": "absent" } TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:28 Sunday 04 February 2024 19:41:03 +0000 (0:00:00.254) 0:02:35.561 ******* included: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for sut TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Sunday 04 February 2024 19:41:03 +0000 (0:00:00.036) 0:02:35.598 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Sunday 04 February 2024 19:41:03 +0000 (0:00:00.022) 0:02:35.620 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Sunday 04 February 2024 19:41:03 +0000 (0:00:00.021) 0:02:35.641 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Reload systemctl] ********************* task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:36 Sunday 04 February 2024 19:41:03 +0000 (0:00:00.021) 0:02:35.663 ******* ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.podman : Prune images no longer in use] ******** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:45 Sunday 04 February 2024 19:41:04 +0000 (0:00:00.871) 0:02:36.535 ******* changed: [sut] => { "changed": true, "cmd": [ "podman", "image", "prune", "-f" ], "delta": "0:00:00.040393", "end": "2024-02-04 19:41:04.951752", "rc": 0, "start": "2024-02-04 19:41:04.911359" } TASK [fedora.linux_system_roles.podman : Remove network] *********************** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:53 Sunday 04 February 2024 19:41:04 +0000 (0:00:00.290) 0:02:36.825 ******* changed: [sut] => { "changed": true, "cmd": [ "podman", "network", "rm", "systemd-quadlet-demo" ], "delta": "0:00:00.037810", "end": "2024-02-04 19:41:05.239190", "rc": 0, "start": "2024-02-04 19:41:05.201380" } STDOUT: systemd-quadlet-demo TASK [fedora.linux_system_roles.podman : Create and update quadlets] *********** task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:107 Sunday 04 February 2024 19:41:05 +0000 (0:00:00.288) 0:02:37.113 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_state != \"absent\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Cancel linger] ************************ task path: /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:122 Sunday 04 February 2024 19:41:05 +0000 (0:00:00.021) 0:02:37.135 ******* skipping: [sut] => { "changed": false, "skipped_reason": "No items in the list" } PLAY RECAP ********************************************************************* sut : ok=264 changed=25 unreachable=0 failed=0 skipped=249 rescued=1 ignored=0 Sunday 04 February 2024 19:41:05 +0000 (0:00:00.066) 0:02:37.201 ******* =============================================================================== fedora.linux_system_roles.podman : Ensure required packages are installed -- 80.59s /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:14 fedora.linux_system_roles.podman : Ensure container images are present --- 9.88s /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:20 fedora.linux_system_roles.certificate : Ensure certificate role dependencies are installed --- 6.16s /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:5 fedora.linux_system_roles.certificate : Ensure provider packages are installed --- 4.82s /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:23 fedora.linux_system_roles.firewall : Enable and start firewalld service --- 2.08s /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:28 fedora.linux_system_roles.firewall : Install firewalld ------------------ 2.02s /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:19 fedora.linux_system_roles.firewall : Install firewalld ------------------ 1.99s /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:19 fedora.linux_system_roles.firewall : Configure firewall ----------------- 1.66s /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:71 fedora.linux_system_roles.certificate : Ensure provider service is running --- 1.12s /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:90 fedora.linux_system_roles.podman : Gather the package facts ------------- 1.00s /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:6 Gathering Facts --------------------------------------------------------- 0.94s /WORKDIR/git-fix-linger-handlingciq7eztj/tests/tests_quadlet_demo.yml:9 ------- fedora.linux_system_roles.podman : Reload systemctl --------------------- 0.93s /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:36 fedora.linux_system_roles.podman : Stop and disable service ------------- 0.92s /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:4 fedora.linux_system_roles.podman : Stop and disable service ------------- 0.91s /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:4 fedora.linux_system_roles.podman : Reload systemctl --------------------- 0.89s /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:80 fedora.linux_system_roles.podman : Gather the package facts ------------- 0.88s /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:6 fedora.linux_system_roles.podman : Reload systemctl --------------------- 0.88s /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:80 fedora.linux_system_roles.podman : Reload systemctl --------------------- 0.88s /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:80 fedora.linux_system_roles.podman : Reload systemctl --------------------- 0.87s /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:36 fedora.linux_system_roles.podman : Reload systemctl --------------------- 0.87s /WORKDIR/git-fix-linger-handlingciq7eztj/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:36 ---^---^---^---^---^--- # STDERR: ---v---v---v---v---v--- [DEPRECATION WARNING]: ANSIBLE_COLLECTIONS_PATHS option, does not fit var naming standard, use the singular form ANSIBLE_COLLECTIONS_PATH instead. This feature will be removed from ansible-core in version 2.19. Deprecation warnings can be disabled by setting deprecation_warnings=False in ansible.cfg. [WARNING]: Using a variable for a task's 'args' is unsafe in some situations (see https://docs.ansible.com/ansible/devel/reference_appendices/faq.html#argsplat- unsafe) ---^---^---^---^---^---