# STDOUT: ---v---v---v---v---v--- ansible-playbook [core 2.16.0] config file = /etc/ansible/ansible.cfg configured module search path = ['/home/jenkins/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules'] ansible python module location = /opt/ansible-2.16/lib/python3.11/site-packages/ansible ansible collection location = /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection executable location = /opt/ansible-2.16/bin/ansible-playbook python version = 3.11.5 (main, Sep 7 2023, 00:00:00) [GCC 11.4.1 20230605 (Red Hat 11.4.1-2)] (/opt/ansible-2.16/bin/python) jinja version = 3.1.2 libyaml = True Using /etc/ansible/ansible.cfg as config file Skipping callback 'debug', as we already have a stdout callback. Skipping callback 'default', as we already have a stdout callback. Skipping callback 'minimal', as we already have a stdout callback. Skipping callback 'oneline', as we already have a stdout callback. PLAYBOOK: tests_quadlet_demo.yml *********************************************** 2 plays in /WORKDIR/git-handle-quadlet-cleanup35__kigx/tests/tests_quadlet_demo.yml PLAY [all] ********************************************************************* TASK [Include vault variables] ************************************************* task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/tests/tests_quadlet_demo.yml:4 Wednesday 10 July 2024 05:11:55 +0000 (0:00:00.011) 0:00:00.011 ******** ok: [sut] => { "ansible_facts": { "__podman_test_password": { "__ansible_vault": "$ANSIBLE_VAULT;1.1;AES256\n35383939616163653333633431363463313831383037386236646138333162396161356130303461\n3932623930643263313563336163316337643562333936360a363538636631313039343233383732\n38666530383538656639363465313230343533386130303833336434303438333161656262346562\n3362626538613031640a663330613638366132356534363534353239616666653466353961323533\n6565\n" }, "mysql_container_root_password": { "__ansible_vault": "$ANSIBLE_VAULT;1.1;AES256\n61333932373230333539663035366431326163363166363036323963623131363530326231303634\n6635326161643165363366323062333334363730376631660a393566366139353861656364656661\n38653463363837336639363032646433666361646535366137303464623261313663643336306465\n6264663730656337310a343962353137386238383064646533366433333437303566656433386233\n34343235326665646661623131643335313236313131353661386338343366316261643634653633\n3832313034366536616531323963333234326461353130303532\n" } }, "ansible_included_var_files": [ "/WORKDIR/git-handle-quadlet-cleanup35__kigx/tests/vars/vault-variables.yml" ], "changed": false } PLAY [Deploy the quadlet demo app] ********************************************* TASK [Gathering Facts] ********************************************************* task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/tests/tests_quadlet_demo.yml:9 Wednesday 10 July 2024 05:11:55 +0000 (0:00:00.012) 0:00:00.024 ******** ok: [sut] TASK [Generate certificates] *************************************************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/tests/tests_quadlet_demo.yml:39 Wednesday 10 July 2024 05:11:56 +0000 (0:00:00.974) 0:00:00.998 ******** TASK [fedora.linux_system_roles.certificate : Set version specific variables] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:2 Wednesday 10 July 2024 05:11:56 +0000 (0:00:00.023) 0:00:01.022 ******** included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/set_vars.yml for sut TASK [fedora.linux_system_roles.certificate : Ensure ansible_facts used by role] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/set_vars.yml:2 Wednesday 10 July 2024 05:11:56 +0000 (0:00:00.012) 0:00:01.034 ******** skipping: [sut] => { "changed": false, "false_condition": "__certificate_required_facts | difference(ansible_facts.keys() | list) | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.certificate : Check if system is ostree] ******* task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/set_vars.yml:10 Wednesday 10 July 2024 05:11:56 +0000 (0:00:00.017) 0:00:01.052 ******** ok: [sut] => { "changed": false, "stat": { "exists": false } } TASK [fedora.linux_system_roles.certificate : Set flag to indicate system is ostree] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/set_vars.yml:15 Wednesday 10 July 2024 05:11:56 +0000 (0:00:00.275) 0:00:01.327 ******** ok: [sut] => { "ansible_facts": { "__certificate_is_ostree": false }, "changed": false } TASK [fedora.linux_system_roles.certificate : Set platform/version specific variables] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/set_vars.yml:19 Wednesday 10 July 2024 05:11:56 +0000 (0:00:00.014) 0:00:01.342 ******** skipping: [sut] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "false_condition": "__vars_file is file", "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [sut] => (item=Fedora.yml) => { "ansible_facts": { "__certificate_certmonger_packages": [ "certmonger", "python3-packaging" ] }, "ansible_included_var_files": [ "/WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/vars/Fedora.yml" ], "ansible_loop_var": "item", "changed": false, "item": "Fedora.yml" } skipping: [sut] => (item=Fedora_39.yml) => { "ansible_loop_var": "item", "changed": false, "false_condition": "__vars_file is file", "item": "Fedora_39.yml", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=Fedora_39.yml) => { "ansible_loop_var": "item", "changed": false, "false_condition": "__vars_file is file", "item": "Fedora_39.yml", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.certificate : Ensure certificate role dependencies are installed] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:5 Wednesday 10 July 2024 05:11:56 +0000 (0:00:00.023) 0:00:01.366 ******** changed: [sut] => { "changed": true, "rc": 0, "results": [ "Installed: python3-ply-3.11-20.fc39.noarch", "Installed: python3-pycparser-2.20-11.fc39.noarch", "Installed: python3-cryptography-41.0.7-1.fc39.x86_64", "Installed: python3-cffi-1.15.1-6.fc39.x86_64", "Installed: python3-pyasn1-0.5.1-1.fc39.noarch" ] } TASK [fedora.linux_system_roles.certificate : Ensure provider packages are installed] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:23 Wednesday 10 July 2024 05:12:01 +0000 (0:00:04.834) 0:00:06.200 ******** changed: [sut] => (item=certmonger) => { "__certificate_provider": "certmonger", "ansible_loop_var": "__certificate_provider", "changed": true, "rc": 0, "results": [ "Installed: certmonger-0.79.18-2.fc39.x86_64", "Installed: dbus-tools-1:1.14.10-1.fc39.x86_64", "Installed: python3-packaging-23.1-4.fc39.noarch" ] } TASK [fedora.linux_system_roles.certificate : Ensure pre-scripts hooks directory exists] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:35 Wednesday 10 July 2024 05:12:06 +0000 (0:00:04.469) 0:00:10.670 ******** changed: [sut] => (item=certmonger) => { "__certificate_provider": "certmonger", "ansible_loop_var": "__certificate_provider", "changed": true, "gid": 0, "group": "root", "mode": "0700", "owner": "root", "path": "/etc/certmonger//pre-scripts", "secontext": "unconfined_u:object_r:etc_t:s0", "size": 4096, "state": "directory", "uid": 0 } TASK [fedora.linux_system_roles.certificate : Ensure post-scripts hooks directory exists] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:61 Wednesday 10 July 2024 05:12:06 +0000 (0:00:00.399) 0:00:11.069 ******** changed: [sut] => (item=certmonger) => { "__certificate_provider": "certmonger", "ansible_loop_var": "__certificate_provider", "changed": true, "gid": 0, "group": "root", "mode": "0700", "owner": "root", "path": "/etc/certmonger//post-scripts", "secontext": "unconfined_u:object_r:etc_t:s0", "size": 4096, "state": "directory", "uid": 0 } TASK [fedora.linux_system_roles.certificate : Ensure provider service is running] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:90 Wednesday 10 July 2024 05:12:06 +0000 (0:00:00.248) 0:00:11.318 ******** changed: [sut] => (item=certmonger) => { "__certificate_provider": "certmonger", "ansible_loop_var": "__certificate_provider", "changed": true, "enabled": true, "name": "certmonger", "state": "started", "status": { "AccessSELinuxContext": "system_u:object_r:certmonger_unit_file_t:s0", "ActiveEnterTimestampMonotonic": "0", "ActiveExitTimestampMonotonic": "0", "ActiveState": "inactive", "After": "syslog.target systemd-journald.socket dbus-broker.service sysinit.target system.slice basic.target network.target dbus.socket", "AllowIsolate": "no", "AssertResult": "no", "AssertTimestampMonotonic": "0", "Before": "shutdown.target", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "BusName": "org.fedorahosted.certmonger", "CPUAccounting": "yes", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "[not set]", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "no", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore", "CleanResult": "success", "CollectMode": "inactive", "ConditionResult": "no", "ConditionTimestampMonotonic": "0", "ConfigurationDirectoryMode": "0755", "Conflicts": "shutdown.target", "ControlGroupId": "0", "ControlPID": "0", "CoredumpFilter": "0x33", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "DefaultStartupMemoryLow": "0", "Delegate": "no", "Description": "Certificate monitoring and PKI enrollment", "DevicePolicy": "auto", "DropInPaths": "/usr/lib/systemd/system/service.d/10-timeout-abort.conf", "DynamicUser": "no", "EnvironmentFiles": "/etc/sysconfig/certmonger (ignore_errors=yes)", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "0", "ExecMainStartTimestampMonotonic": "0", "ExecMainStatus": "0", "ExecStart": "{ path=/usr/sbin/certmonger ; argv[]=/usr/sbin/certmonger -S -p /run/certmonger.pid -n $OPTS ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStartEx": "{ path=/usr/sbin/certmonger ; argv[]=/usr/sbin/certmonger -S -p /run/certmonger.pid -n $OPTS ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExitType": "main", "ExtensionImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FileDescriptorStorePreserve": "restart", "FinalKillSignal": "9", "FragmentPath": "/usr/lib/systemd/system/certmonger.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOReadBytes": "18446744073709551615", "IOReadOperations": "18446744073709551615", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "IOWeight": "[not set]", "IOWriteBytes": "18446744073709551615", "IOWriteOperations": "18446744073709551615", "IPAccounting": "no", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "Id": "certmonger.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestampMonotonic": "0", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "control-group", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "infinity", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "8388608", "LimitMEMLOCKSoft": "8388608", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "524288", "LimitNOFILESoft": "1024", "LimitNPROC": "14756", "LimitNPROCSoft": "14756", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "14756", "LimitSIGPENDINGSoft": "14756", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "0", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "ManagedOOMSwap": "auto", "MemoryAccounting": "yes", "MemoryAvailable": "infinity", "MemoryCurrent": "[not set]", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryKSM": "no", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemoryPressureThresholdUSec": "200ms", "MemoryPressureWatch": "auto", "MemorySwapMax": "infinity", "MemoryZSwapMax": "infinity", "MountAPIVFS": "no", "MountImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAPolicy": "n/a", "Names": "certmonger.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMPolicy": "stop", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "OnSuccessJobMode": "fail", "PIDFile": "/run/certmonger.pid", "PartOf": "dbus-broker.service", "Perpetual": "no", "PrivateDevices": "no", "PrivateIPC": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProcSubset": "all", "ProtectClock": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectHostname": "no", "ProtectKernelLogs": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectProc": "default", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "ReloadResult": "success", "ReloadSignal": "1", "RemainAfterExit": "no", "RemoveIPC": "no", "Requires": "dbus.socket sysinit.target system.slice", "Restart": "no", "RestartKillSignal": "15", "RestartMaxDelayUSec": "infinity", "RestartMode": "normal", "RestartSteps": "0", "RestartUSec": "100ms", "RestartUSecNext": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RootEphemeral": "no", "RootImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "Slice": "system.slice", "StandardError": "inherit", "StandardInput": "null", "StandardOutput": "journal", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StartupMemoryHigh": "18446744073709551615", "StartupMemoryLow": "0", "StartupMemoryMax": "18446744073709551615", "StartupMemorySwapMax": "18446744073709551615", "StartupMemoryZSwapMax": "18446744073709551615", "StateChangeTimestampMonotonic": "0", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "dead", "SuccessAction": "none", "SyslogFacility": "3", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "2147483646", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "[not set]", "TasksMax": "4426", "TimeoutAbortUSec": "45s", "TimeoutCleanUSec": "infinity", "TimeoutStartFailureMode": "terminate", "TimeoutStartUSec": "45s", "TimeoutStopFailureMode": "abort", "TimeoutStopUSec": "45s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "dbus", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "disabled", "UnitFileState": "disabled", "UtmpMode": "init", "WatchdogSignal": "6", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "infinity" } } TASK [fedora.linux_system_roles.certificate : Ensure certificate requests] ***** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:101 Wednesday 10 July 2024 05:12:07 +0000 (0:00:01.009) 0:00:12.328 ******** changed: [sut] => (item={'name': 'quadlet_demo', 'dns': ['localhost'], 'ca': 'self-sign'}) => { "ansible_loop_var": "item", "changed": true, "item": { "ca": "self-sign", "dns": [ "localhost" ], "name": "quadlet_demo" } } MSG: Certificate requested (new). TASK [fedora.linux_system_roles.certificate : Slurp the contents of the files] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:152 Wednesday 10 July 2024 05:12:08 +0000 (0:00:00.733) 0:00:13.061 ******** ok: [sut] => (item=['cert', {'name': 'quadlet_demo', 'dns': ['localhost'], 'ca': 'self-sign'}]) => { "ansible_loop_var": "item", "changed": false, "content": "LS0tLS1CRUdJTiBDRVJUSUZJQ0FURS0tLS0tCk1JSURnakNDQW1xZ0F3SUJBZ0lRZDR1WUh2UExTSlc4ekdlQWtEWEV0akFOQmdrcWhraUc5dzBCQVFzRkFEQlEKTVNBd0hnWURWUVFEREJkTWIyTmhiQ0JUYVdkdWFXNW5JRUYxZEdodmNtbDBlVEVzTUNvR0ExVUVBd3dqTnpjNApZams0TVdVdFpqTmpZalE0T1RVdFltTmpZelkzT0RBdE9UQXpOV00wWWpVd0hoY05NalF3TnpFd01EVXhNakE0CldoY05NalV3TnpFd01EVXhNakEzV2pBVU1SSXdFQVlEVlFRREV3bHNiMk5oYkdodmMzUXdnZ0VpTUEwR0NTcUcKU0liM0RRRUJBUVVBQTRJQkR3QXdnZ0VLQW9JQkFRQ1lHWktkbWltVkVBNVJwRmhhVjRBTlBUWURkb0Rqd2gxNQprZGxOWmZHQW50dVJBT24wNTA4NSsrNTBRZmhvbU1yeVJsOVRwd0poZEoxT3Q2RkVPYU9pZUZadGFhSGluRmhZCk1WY0lQUFYyUmdGMVd6SXlQQzYzUTFqQ2xIVHQvMWdvRk1lTG52a3VKdkpGZUZvVk9scG0zckJUYllGT252enEKYkVyT0NDNVNxK0p5V3dYYkJtY25xUjRzczBYeGRUZzV5S0FxU3dtMXNTK0dtYVByVDJoYnpqaVNoQXpzaEZEWgpsOUJGNHc0NW1KL2pWc3JDc2k3OXBYL0VyOXgwMVZTeGdvaFJuTzMvaDBGUHUxTzVsbFBENm1HU1oxbUpIMzUvCkZ4WGlZcDRKZmw1V1Y0QkROOFBQWnJiay9uNGErYmNrRkh3cm1Ib3pLdnVUMkVMUlpxVGhBZ01CQUFHamdaTXcKZ1pBd0N3WURWUjBQQkFRREFnV2dNQlFHQTFVZEVRUU5NQXVDQ1d4dlkyRnNhRzl6ZERBZEJnTlZIU1VFRmpBVQpCZ2dyQmdFRkJRY0RBUVlJS3dZQkJRVUhBd0l3REFZRFZSMFRBUUgvQkFJd0FEQWRCZ05WSFE0RUZnUVV0OE5lCkdTSHVxOGcrakZTL3h2YS9FK3BPNXJzd0h3WURWUjBqQkJnd0ZvQVU2bGdLVGpnSGJSTEhxK3E0WWw4SGVtVmEKVzZnd0RRWUpLb1pJaHZjTkFRRUxCUUFEZ2dFQkFFYkhOb1dFVWZkVVZwTDBlMVhxMGRSclVtaUoraXJHRG5BYwpjWUxaUXdCQjRLSW9INkQzQ051ZUZydXVSNmxEN3d0bVNPYmZDU3ZLNGFmYXdST01nWE1yVVErQnBkZDdpSVRJClRCeUp6bDVpQ0M0NWY1c21kNS9nR01QN1Rqd2RGeVZhT25RV0ZyNHV5bzhkVHVxamNGcVZaWksxekVZZlQ3aGMKMFh6OGJJRkhzRWJIZ01SaEhxYzlQQ29MRHcxem0reDlVY2JLd3pKUWRNczZESlo5c2NSRWg5cmgxY2NCMW4xbQpVZHVrK2FUdWhRdTQ3Z2NHVUtiMzhqaWxYVllWbys4SE1BZDA1bnZTeHlOa2RrWUIzNVBEM2llQzBaSng5VzJSCk4xQkk5cXJ2emVLKzhpSElxZEU0MktlYksya1EyWmdOaktjSTQ0aEJLSnMvaW5EZGRFND0KLS0tLS1FTkQgQ0VSVElGSUNBVEUtLS0tLQo=", "encoding": "base64", "item": [ "cert", { "ca": "self-sign", "dns": [ "localhost" ], "name": "quadlet_demo" } ], "source": "/etc/pki/tls/certs/quadlet_demo.crt" } ok: [sut] => (item=['key', {'name': 'quadlet_demo', 'dns': ['localhost'], 'ca': 'self-sign'}]) => { "ansible_loop_var": "item", "changed": false, "content": "LS0tLS1CRUdJTiBQUklWQVRFIEtFWS0tLS0tCk1JSUV2UUlCQURBTkJna3Foa2lHOXcwQkFRRUZBQVNDQktjd2dnU2pBZ0VBQW9JQkFRQ1lHWktkbWltVkVBNVIKcEZoYVY0QU5QVFlEZG9EandoMTVrZGxOWmZHQW50dVJBT24wNTA4NSsrNTBRZmhvbU1yeVJsOVRwd0poZEoxTwp0NkZFT2FPaWVGWnRhYUhpbkZoWU1WY0lQUFYyUmdGMVd6SXlQQzYzUTFqQ2xIVHQvMWdvRk1lTG52a3VKdkpGCmVGb1ZPbHBtM3JCVGJZRk9udnpxYkVyT0NDNVNxK0p5V3dYYkJtY25xUjRzczBYeGRUZzV5S0FxU3dtMXNTK0cKbWFQclQyaGJ6amlTaEF6c2hGRFpsOUJGNHc0NW1KL2pWc3JDc2k3OXBYL0VyOXgwMVZTeGdvaFJuTzMvaDBGUAp1MU81bGxQRDZtR1NaMW1KSDM1L0Z4WGlZcDRKZmw1V1Y0QkROOFBQWnJiay9uNGErYmNrRkh3cm1Ib3pLdnVUCjJFTFJacVRoQWdNQkFBRUNnZ0VBQUxTNkxpeDJjZEpkTWFadzBVc0hqWEF0dTNDQjdzU3ljck54U2ZOZG5vaG0KT2g4aUp0eUp2T241SUcyanRiY1RWcEpxYS9yS0Ixb2xab1dTVEZ6MDRDcTlNblR2Znh2MUNrUkEzRms5bGF6agpvZmdBVkxTOExnWTBUOU5rODB1WWpQZVRPVWY5cmJmWkx5YTh5czFPZVB5aXZvRmpMTEp1MTdiMDFpM2FLckN2CkZ0RHE1NmZNMGIxb2t1bDVBS1RhRUluU09MV3FyK0NBYnFaS2dKSDRHQ3ZuWURuRmVzb1daSC9iSWpQbmtOU2oKUE5EUGVRMGJKbFBJUVRmZ2JZM242U3hZckY3TWl5dWdVczROWXYyKzV4NTdhdkNnL3BFckt0c2xhczN2QUV5QgprOUlUbTdXQVBna09FaTBHMVB3TnFzbFpXUW02NUdIRXhZQVNlRFBRZ1FLQmdRRFFYbGpkN3poREF5QXdPVG5XCkNPUWVUcTVFSDBqUW82Ni80YytjdXlieWhzditlR1A0OUJnSkd0Ty84SllieHM2NFZkdTY0Y0E1NHRzMUwwMGcKK2xkLzBxVG5sbEJKWHRTOGFjOEp3N0VjMFROV2RHeWQraTJpSHlxK3BMYXJRbFhiUDFKSlE3MWdudUJtVUlkNwp4WktXTldJcitBbFpudFo1aFNzYmFSbitJUUtCZ1FDNjNtZDhmZHU3cGE4NGVqRzJGVEF5SjE5Y2ZZNzF3bCtYCmkxcU1Sb2xyZFBiVFgwWS8vMDlBQzQvVGVxcWNUYkZqZG9IUW5Qa20rTjVRcjBKUU8veGNuVzBrYWxIc1Bhc0EKcTFPeUJJTCsvY0NqV2gwVklLbzczN3o3WkV5LzdlcnEyeU5jaGhlektZcXlPRldNcU92clFocnMxTjRqS3ZHbQpmaHVWdldCT3dRS0JnUUM5QVlmTXNLeXhKSW16RklXZzMybkZFQXZGSVRlVWdaVUcydGJPelhpenI1QWs0S0dXCmRodkRUb0ZhN3VuS2dsaUxIdDUwczBnQmVnVHJSbC9KRlN2L2Qxdzc4YjRiLzVncTNET0FJUEZzT1ZzL0JkWnMKcWoxd2ZLTHlVVWlkS2t2azE1aFZlc3V5ZEFzSzQrc3hXM0x2SzdkRU9QYUtRNms2aTVBaGlnR1k0UUtCZ0JiRApKcG5zZnNxUnVSRkpHU3NUMDZZWVU2NWhnUHVBck1JNThGS3IrT3RFK1Jmd1FBOWxteEt6RmRidDEwQm8ybFhDClJyZlJseXZmSHIxM1M3MllRUHBMRnROYnZCVHB3aDIzd044clc0V1NCVmdnanZCVzVVV2IwRnBTOG1KaXFiM24KNlp2SW5iUnVOLzNUak9Bd1oyUmx6Zi9SUWluSWZsMEllWFc0OXJWQkFvR0FYWFQ3UkxualZDdlVCbzhtaWpUMwplUm5wNHVtSm9CUTk0Tk11NzBxV0pXbmIxWk1yUmlMNHZEd3JMZk11YUhYM2xjUUk3YklmSnUydkZqZGZOdjJOCjVUYlVMcy91aE51ZVhwaFlGMGlXb05VNmwveTVOaTYyN3pQUTdxMWphekU0RjhOaktxazZSOWhZY1U4NW1JZ3YKa2QwYkVYckRtM1lHem5SWmxWQW04RHM9Ci0tLS0tRU5EIFBSSVZBVEUgS0VZLS0tLS0K", "encoding": "base64", "item": [ "key", { "ca": "self-sign", "dns": [ "localhost" ], "name": "quadlet_demo" } ], "source": "/etc/pki/tls/private/quadlet_demo.key" } ok: [sut] => (item=['ca', {'name': 'quadlet_demo', 'dns': ['localhost'], 'ca': 'self-sign'}]) => { "ansible_loop_var": "item", "changed": false, "content": "LS0tLS1CRUdJTiBDRVJUSUZJQ0FURS0tLS0tCk1JSURnakNDQW1xZ0F3SUJBZ0lRZDR1WUh2UExTSlc4ekdlQWtEWEV0akFOQmdrcWhraUc5dzBCQVFzRkFEQlEKTVNBd0hnWURWUVFEREJkTWIyTmhiQ0JUYVdkdWFXNW5JRUYxZEdodmNtbDBlVEVzTUNvR0ExVUVBd3dqTnpjNApZams0TVdVdFpqTmpZalE0T1RVdFltTmpZelkzT0RBdE9UQXpOV00wWWpVd0hoY05NalF3TnpFd01EVXhNakE0CldoY05NalV3TnpFd01EVXhNakEzV2pBVU1SSXdFQVlEVlFRREV3bHNiMk5oYkdodmMzUXdnZ0VpTUEwR0NTcUcKU0liM0RRRUJBUVVBQTRJQkR3QXdnZ0VLQW9JQkFRQ1lHWktkbWltVkVBNVJwRmhhVjRBTlBUWURkb0Rqd2gxNQprZGxOWmZHQW50dVJBT24wNTA4NSsrNTBRZmhvbU1yeVJsOVRwd0poZEoxT3Q2RkVPYU9pZUZadGFhSGluRmhZCk1WY0lQUFYyUmdGMVd6SXlQQzYzUTFqQ2xIVHQvMWdvRk1lTG52a3VKdkpGZUZvVk9scG0zckJUYllGT252enEKYkVyT0NDNVNxK0p5V3dYYkJtY25xUjRzczBYeGRUZzV5S0FxU3dtMXNTK0dtYVByVDJoYnpqaVNoQXpzaEZEWgpsOUJGNHc0NW1KL2pWc3JDc2k3OXBYL0VyOXgwMVZTeGdvaFJuTzMvaDBGUHUxTzVsbFBENm1HU1oxbUpIMzUvCkZ4WGlZcDRKZmw1V1Y0QkROOFBQWnJiay9uNGErYmNrRkh3cm1Ib3pLdnVUMkVMUlpxVGhBZ01CQUFHamdaTXcKZ1pBd0N3WURWUjBQQkFRREFnV2dNQlFHQTFVZEVRUU5NQXVDQ1d4dlkyRnNhRzl6ZERBZEJnTlZIU1VFRmpBVQpCZ2dyQmdFRkJRY0RBUVlJS3dZQkJRVUhBd0l3REFZRFZSMFRBUUgvQkFJd0FEQWRCZ05WSFE0RUZnUVV0OE5lCkdTSHVxOGcrakZTL3h2YS9FK3BPNXJzd0h3WURWUjBqQkJnd0ZvQVU2bGdLVGpnSGJSTEhxK3E0WWw4SGVtVmEKVzZnd0RRWUpLb1pJaHZjTkFRRUxCUUFEZ2dFQkFFYkhOb1dFVWZkVVZwTDBlMVhxMGRSclVtaUoraXJHRG5BYwpjWUxaUXdCQjRLSW9INkQzQ051ZUZydXVSNmxEN3d0bVNPYmZDU3ZLNGFmYXdST01nWE1yVVErQnBkZDdpSVRJClRCeUp6bDVpQ0M0NWY1c21kNS9nR01QN1Rqd2RGeVZhT25RV0ZyNHV5bzhkVHVxamNGcVZaWksxekVZZlQ3aGMKMFh6OGJJRkhzRWJIZ01SaEhxYzlQQ29MRHcxem0reDlVY2JLd3pKUWRNczZESlo5c2NSRWg5cmgxY2NCMW4xbQpVZHVrK2FUdWhRdTQ3Z2NHVUtiMzhqaWxYVllWbys4SE1BZDA1bnZTeHlOa2RrWUIzNVBEM2llQzBaSng5VzJSCk4xQkk5cXJ2emVLKzhpSElxZEU0MktlYksya1EyWmdOaktjSTQ0aEJLSnMvaW5EZGRFND0KLS0tLS1FTkQgQ0VSVElGSUNBVEUtLS0tLQo=", "encoding": "base64", "item": [ "ca", { "ca": "self-sign", "dns": [ "localhost" ], "name": "quadlet_demo" } ], "source": "/etc/pki/tls/certs/quadlet_demo.crt" } TASK [fedora.linux_system_roles.certificate : Create return data] ************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:160 Wednesday 10 July 2024 05:12:09 +0000 (0:00:00.696) 0:00:13.757 ******** ok: [sut] => { "ansible_facts": { "certificate_test_certs": { "quadlet_demo": { "ca": "/etc/pki/tls/certs/quadlet_demo.crt", "ca_content": "-----BEGIN CERTIFICATE-----\nMIIDgjCCAmqgAwIBAgIQd4uYHvPLSJW8zGeAkDXEtjANBgkqhkiG9w0BAQsFADBQ\nMSAwHgYDVQQDDBdMb2NhbCBTaWduaW5nIEF1dGhvcml0eTEsMCoGA1UEAwwjNzc4\nYjk4MWUtZjNjYjQ4OTUtYmNjYzY3ODAtOTAzNWM0YjUwHhcNMjQwNzEwMDUxMjA4\nWhcNMjUwNzEwMDUxMjA3WjAUMRIwEAYDVQQDEwlsb2NhbGhvc3QwggEiMA0GCSqG\nSIb3DQEBAQUAA4IBDwAwggEKAoIBAQCYGZKdmimVEA5RpFhaV4ANPTYDdoDjwh15\nkdlNZfGAntuRAOn05085++50QfhomMryRl9TpwJhdJ1Ot6FEOaOieFZtaaHinFhY\nMVcIPPV2RgF1WzIyPC63Q1jClHTt/1goFMeLnvkuJvJFeFoVOlpm3rBTbYFOnvzq\nbErOCC5Sq+JyWwXbBmcnqR4ss0XxdTg5yKAqSwm1sS+GmaPrT2hbzjiShAzshFDZ\nl9BF4w45mJ/jVsrCsi79pX/Er9x01VSxgohRnO3/h0FPu1O5llPD6mGSZ1mJH35/\nFxXiYp4Jfl5WV4BDN8PPZrbk/n4a+bckFHwrmHozKvuT2ELRZqThAgMBAAGjgZMw\ngZAwCwYDVR0PBAQDAgWgMBQGA1UdEQQNMAuCCWxvY2FsaG9zdDAdBgNVHSUEFjAU\nBggrBgEFBQcDAQYIKwYBBQUHAwIwDAYDVR0TAQH/BAIwADAdBgNVHQ4EFgQUt8Ne\nGSHuq8g+jFS/xva/E+pO5rswHwYDVR0jBBgwFoAU6lgKTjgHbRLHq+q4Yl8HemVa\nW6gwDQYJKoZIhvcNAQELBQADggEBAEbHNoWEUfdUVpL0e1Xq0dRrUmiJ+irGDnAc\ncYLZQwBB4KIoH6D3CNueFruuR6lD7wtmSObfCSvK4afawROMgXMrUQ+Bpdd7iITI\nTByJzl5iCC45f5smd5/gGMP7TjwdFyVaOnQWFr4uyo8dTuqjcFqVZZK1zEYfT7hc\n0Xz8bIFHsEbHgMRhHqc9PCoLDw1zm+x9UcbKwzJQdMs6DJZ9scREh9rh1ccB1n1m\nUduk+aTuhQu47gcGUKb38jilXVYVo+8HMAd05nvSxyNkdkYB35PD3ieC0ZJx9W2R\nN1BI9qrvzeK+8iHIqdE42KebK2kQ2ZgNjKcI44hBKJs/inDddE4=\n-----END CERTIFICATE-----\n", "cert": "/etc/pki/tls/certs/quadlet_demo.crt", "cert_content": "-----BEGIN CERTIFICATE-----\nMIIDgjCCAmqgAwIBAgIQd4uYHvPLSJW8zGeAkDXEtjANBgkqhkiG9w0BAQsFADBQ\nMSAwHgYDVQQDDBdMb2NhbCBTaWduaW5nIEF1dGhvcml0eTEsMCoGA1UEAwwjNzc4\nYjk4MWUtZjNjYjQ4OTUtYmNjYzY3ODAtOTAzNWM0YjUwHhcNMjQwNzEwMDUxMjA4\nWhcNMjUwNzEwMDUxMjA3WjAUMRIwEAYDVQQDEwlsb2NhbGhvc3QwggEiMA0GCSqG\nSIb3DQEBAQUAA4IBDwAwggEKAoIBAQCYGZKdmimVEA5RpFhaV4ANPTYDdoDjwh15\nkdlNZfGAntuRAOn05085++50QfhomMryRl9TpwJhdJ1Ot6FEOaOieFZtaaHinFhY\nMVcIPPV2RgF1WzIyPC63Q1jClHTt/1goFMeLnvkuJvJFeFoVOlpm3rBTbYFOnvzq\nbErOCC5Sq+JyWwXbBmcnqR4ss0XxdTg5yKAqSwm1sS+GmaPrT2hbzjiShAzshFDZ\nl9BF4w45mJ/jVsrCsi79pX/Er9x01VSxgohRnO3/h0FPu1O5llPD6mGSZ1mJH35/\nFxXiYp4Jfl5WV4BDN8PPZrbk/n4a+bckFHwrmHozKvuT2ELRZqThAgMBAAGjgZMw\ngZAwCwYDVR0PBAQDAgWgMBQGA1UdEQQNMAuCCWxvY2FsaG9zdDAdBgNVHSUEFjAU\nBggrBgEFBQcDAQYIKwYBBQUHAwIwDAYDVR0TAQH/BAIwADAdBgNVHQ4EFgQUt8Ne\nGSHuq8g+jFS/xva/E+pO5rswHwYDVR0jBBgwFoAU6lgKTjgHbRLHq+q4Yl8HemVa\nW6gwDQYJKoZIhvcNAQELBQADggEBAEbHNoWEUfdUVpL0e1Xq0dRrUmiJ+irGDnAc\ncYLZQwBB4KIoH6D3CNueFruuR6lD7wtmSObfCSvK4afawROMgXMrUQ+Bpdd7iITI\nTByJzl5iCC45f5smd5/gGMP7TjwdFyVaOnQWFr4uyo8dTuqjcFqVZZK1zEYfT7hc\n0Xz8bIFHsEbHgMRhHqc9PCoLDw1zm+x9UcbKwzJQdMs6DJZ9scREh9rh1ccB1n1m\nUduk+aTuhQu47gcGUKb38jilXVYVo+8HMAd05nvSxyNkdkYB35PD3ieC0ZJx9W2R\nN1BI9qrvzeK+8iHIqdE42KebK2kQ2ZgNjKcI44hBKJs/inDddE4=\n-----END CERTIFICATE-----\n", "key": "/etc/pki/tls/private/quadlet_demo.key", "key_content": "-----BEGIN PRIVATE KEY-----\nMIIEvQIBADANBgkqhkiG9w0BAQEFAASCBKcwggSjAgEAAoIBAQCYGZKdmimVEA5R\npFhaV4ANPTYDdoDjwh15kdlNZfGAntuRAOn05085++50QfhomMryRl9TpwJhdJ1O\nt6FEOaOieFZtaaHinFhYMVcIPPV2RgF1WzIyPC63Q1jClHTt/1goFMeLnvkuJvJF\neFoVOlpm3rBTbYFOnvzqbErOCC5Sq+JyWwXbBmcnqR4ss0XxdTg5yKAqSwm1sS+G\nmaPrT2hbzjiShAzshFDZl9BF4w45mJ/jVsrCsi79pX/Er9x01VSxgohRnO3/h0FP\nu1O5llPD6mGSZ1mJH35/FxXiYp4Jfl5WV4BDN8PPZrbk/n4a+bckFHwrmHozKvuT\n2ELRZqThAgMBAAECggEAALS6Lix2cdJdMaZw0UsHjXAtu3CB7sSycrNxSfNdnohm\nOh8iJtyJvOn5IG2jtbcTVpJqa/rKB1olZoWSTFz04Cq9MnTvfxv1CkRA3Fk9lazj\nofgAVLS8LgY0T9Nk80uYjPeTOUf9rbfZLya8ys1OePyivoFjLLJu17b01i3aKrCv\nFtDq56fM0b1okul5AKTaEInSOLWqr+CAbqZKgJH4GCvnYDnFesoWZH/bIjPnkNSj\nPNDPeQ0bJlPIQTfgbY3n6SxYrF7MiyugUs4NYv2+5x57avCg/pErKtslas3vAEyB\nk9ITm7WAPgkOEi0G1PwNqslZWQm65GHExYASeDPQgQKBgQDQXljd7zhDAyAwOTnW\nCOQeTq5EH0jQo66/4c+cuybyhsv+eGP49BgJGtO/8JYbxs64Vdu64cA54ts1L00g\n+ld/0qTnllBJXtS8ac8Jw7Ec0TNWdGyd+i2iHyq+pLarQlXbP1JJQ71gnuBmUId7\nxZKWNWIr+AlZntZ5hSsbaRn+IQKBgQC63md8fdu7pa84ejG2FTAyJ19cfY71wl+X\ni1qMRolrdPbTX0Y//09AC4/TeqqcTbFjdoHQnPkm+N5Qr0JQO/xcnW0kalHsPasA\nq1OyBIL+/cCjWh0VIKo737z7ZEy/7erq2yNchhezKYqyOFWMqOvrQhrs1N4jKvGm\nfhuVvWBOwQKBgQC9AYfMsKyxJImzFIWg32nFEAvFITeUgZUG2tbOzXizr5Ak4KGW\ndhvDToFa7unKgliLHt50s0gBegTrRl/JFSv/d1w78b4b/5gq3DOAIPFsOVs/BdZs\nqj1wfKLyUUidKkvk15hVesuydAsK4+sxW3LvK7dEOPaKQ6k6i5AhigGY4QKBgBbD\nJpnsfsqRuRFJGSsT06YYU65hgPuArMI58FKr+OtE+RfwQA9lmxKzFdbt10Bo2lXC\nRrfRlyvfHr13S72YQPpLFtNbvBTpwh23wN8rW4WSBVggjvBW5UWb0FpS8mJiqb3n\n6ZvInbRuN/3TjOAwZ2Rlzf/RQinIfl0IeXW49rVBAoGAXXT7RLnjVCvUBo8mijT3\neRnp4umJoBQ94NMu70qWJWnb1ZMrRiL4vDwrLfMuaHX3lcQI7bIfJu2vFjdfNv2N\n5TbULs/uhNueXphYF0iWoNU6l/y5Ni627zPQ7q1jazE4F8NjKqk6R9hYcU85mIgv\nkd0bEXrDm3YGznRZlVAm8Ds=\n-----END PRIVATE KEY-----\n" } } }, "changed": false } TASK [fedora.linux_system_roles.certificate : Stop tracking certificates] ****** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:176 Wednesday 10 July 2024 05:12:09 +0000 (0:00:00.022) 0:00:13.780 ******** ok: [sut] => (item={'cert': '/etc/pki/tls/certs/quadlet_demo.crt', 'cert_content': '-----BEGIN CERTIFICATE-----\nMIIDgjCCAmqgAwIBAgIQd4uYHvPLSJW8zGeAkDXEtjANBgkqhkiG9w0BAQsFADBQ\nMSAwHgYDVQQDDBdMb2NhbCBTaWduaW5nIEF1dGhvcml0eTEsMCoGA1UEAwwjNzc4\nYjk4MWUtZjNjYjQ4OTUtYmNjYzY3ODAtOTAzNWM0YjUwHhcNMjQwNzEwMDUxMjA4\nWhcNMjUwNzEwMDUxMjA3WjAUMRIwEAYDVQQDEwlsb2NhbGhvc3QwggEiMA0GCSqG\nSIb3DQEBAQUAA4IBDwAwggEKAoIBAQCYGZKdmimVEA5RpFhaV4ANPTYDdoDjwh15\nkdlNZfGAntuRAOn05085++50QfhomMryRl9TpwJhdJ1Ot6FEOaOieFZtaaHinFhY\nMVcIPPV2RgF1WzIyPC63Q1jClHTt/1goFMeLnvkuJvJFeFoVOlpm3rBTbYFOnvzq\nbErOCC5Sq+JyWwXbBmcnqR4ss0XxdTg5yKAqSwm1sS+GmaPrT2hbzjiShAzshFDZ\nl9BF4w45mJ/jVsrCsi79pX/Er9x01VSxgohRnO3/h0FPu1O5llPD6mGSZ1mJH35/\nFxXiYp4Jfl5WV4BDN8PPZrbk/n4a+bckFHwrmHozKvuT2ELRZqThAgMBAAGjgZMw\ngZAwCwYDVR0PBAQDAgWgMBQGA1UdEQQNMAuCCWxvY2FsaG9zdDAdBgNVHSUEFjAU\nBggrBgEFBQcDAQYIKwYBBQUHAwIwDAYDVR0TAQH/BAIwADAdBgNVHQ4EFgQUt8Ne\nGSHuq8g+jFS/xva/E+pO5rswHwYDVR0jBBgwFoAU6lgKTjgHbRLHq+q4Yl8HemVa\nW6gwDQYJKoZIhvcNAQELBQADggEBAEbHNoWEUfdUVpL0e1Xq0dRrUmiJ+irGDnAc\ncYLZQwBB4KIoH6D3CNueFruuR6lD7wtmSObfCSvK4afawROMgXMrUQ+Bpdd7iITI\nTByJzl5iCC45f5smd5/gGMP7TjwdFyVaOnQWFr4uyo8dTuqjcFqVZZK1zEYfT7hc\n0Xz8bIFHsEbHgMRhHqc9PCoLDw1zm+x9UcbKwzJQdMs6DJZ9scREh9rh1ccB1n1m\nUduk+aTuhQu47gcGUKb38jilXVYVo+8HMAd05nvSxyNkdkYB35PD3ieC0ZJx9W2R\nN1BI9qrvzeK+8iHIqdE42KebK2kQ2ZgNjKcI44hBKJs/inDddE4=\n-----END CERTIFICATE-----\n', 'key': '/etc/pki/tls/private/quadlet_demo.key', 'key_content': '-----BEGIN PRIVATE KEY-----\nMIIEvQIBADANBgkqhkiG9w0BAQEFAASCBKcwggSjAgEAAoIBAQCYGZKdmimVEA5R\npFhaV4ANPTYDdoDjwh15kdlNZfGAntuRAOn05085++50QfhomMryRl9TpwJhdJ1O\nt6FEOaOieFZtaaHinFhYMVcIPPV2RgF1WzIyPC63Q1jClHTt/1goFMeLnvkuJvJF\neFoVOlpm3rBTbYFOnvzqbErOCC5Sq+JyWwXbBmcnqR4ss0XxdTg5yKAqSwm1sS+G\nmaPrT2hbzjiShAzshFDZl9BF4w45mJ/jVsrCsi79pX/Er9x01VSxgohRnO3/h0FP\nu1O5llPD6mGSZ1mJH35/FxXiYp4Jfl5WV4BDN8PPZrbk/n4a+bckFHwrmHozKvuT\n2ELRZqThAgMBAAECggEAALS6Lix2cdJdMaZw0UsHjXAtu3CB7sSycrNxSfNdnohm\nOh8iJtyJvOn5IG2jtbcTVpJqa/rKB1olZoWSTFz04Cq9MnTvfxv1CkRA3Fk9lazj\nofgAVLS8LgY0T9Nk80uYjPeTOUf9rbfZLya8ys1OePyivoFjLLJu17b01i3aKrCv\nFtDq56fM0b1okul5AKTaEInSOLWqr+CAbqZKgJH4GCvnYDnFesoWZH/bIjPnkNSj\nPNDPeQ0bJlPIQTfgbY3n6SxYrF7MiyugUs4NYv2+5x57avCg/pErKtslas3vAEyB\nk9ITm7WAPgkOEi0G1PwNqslZWQm65GHExYASeDPQgQKBgQDQXljd7zhDAyAwOTnW\nCOQeTq5EH0jQo66/4c+cuybyhsv+eGP49BgJGtO/8JYbxs64Vdu64cA54ts1L00g\n+ld/0qTnllBJXtS8ac8Jw7Ec0TNWdGyd+i2iHyq+pLarQlXbP1JJQ71gnuBmUId7\nxZKWNWIr+AlZntZ5hSsbaRn+IQKBgQC63md8fdu7pa84ejG2FTAyJ19cfY71wl+X\ni1qMRolrdPbTX0Y//09AC4/TeqqcTbFjdoHQnPkm+N5Qr0JQO/xcnW0kalHsPasA\nq1OyBIL+/cCjWh0VIKo737z7ZEy/7erq2yNchhezKYqyOFWMqOvrQhrs1N4jKvGm\nfhuVvWBOwQKBgQC9AYfMsKyxJImzFIWg32nFEAvFITeUgZUG2tbOzXizr5Ak4KGW\ndhvDToFa7unKgliLHt50s0gBegTrRl/JFSv/d1w78b4b/5gq3DOAIPFsOVs/BdZs\nqj1wfKLyUUidKkvk15hVesuydAsK4+sxW3LvK7dEOPaKQ6k6i5AhigGY4QKBgBbD\nJpnsfsqRuRFJGSsT06YYU65hgPuArMI58FKr+OtE+RfwQA9lmxKzFdbt10Bo2lXC\nRrfRlyvfHr13S72YQPpLFtNbvBTpwh23wN8rW4WSBVggjvBW5UWb0FpS8mJiqb3n\n6ZvInbRuN/3TjOAwZ2Rlzf/RQinIfl0IeXW49rVBAoGAXXT7RLnjVCvUBo8mijT3\neRnp4umJoBQ94NMu70qWJWnb1ZMrRiL4vDwrLfMuaHX3lcQI7bIfJu2vFjdfNv2N\n5TbULs/uhNueXphYF0iWoNU6l/y5Ni627zPQ7q1jazE4F8NjKqk6R9hYcU85mIgv\nkd0bEXrDm3YGznRZlVAm8Ds=\n-----END PRIVATE KEY-----\n', 'ca': '/etc/pki/tls/certs/quadlet_demo.crt', 'ca_content': '-----BEGIN CERTIFICATE-----\nMIIDgjCCAmqgAwIBAgIQd4uYHvPLSJW8zGeAkDXEtjANBgkqhkiG9w0BAQsFADBQ\nMSAwHgYDVQQDDBdMb2NhbCBTaWduaW5nIEF1dGhvcml0eTEsMCoGA1UEAwwjNzc4\nYjk4MWUtZjNjYjQ4OTUtYmNjYzY3ODAtOTAzNWM0YjUwHhcNMjQwNzEwMDUxMjA4\nWhcNMjUwNzEwMDUxMjA3WjAUMRIwEAYDVQQDEwlsb2NhbGhvc3QwggEiMA0GCSqG\nSIb3DQEBAQUAA4IBDwAwggEKAoIBAQCYGZKdmimVEA5RpFhaV4ANPTYDdoDjwh15\nkdlNZfGAntuRAOn05085++50QfhomMryRl9TpwJhdJ1Ot6FEOaOieFZtaaHinFhY\nMVcIPPV2RgF1WzIyPC63Q1jClHTt/1goFMeLnvkuJvJFeFoVOlpm3rBTbYFOnvzq\nbErOCC5Sq+JyWwXbBmcnqR4ss0XxdTg5yKAqSwm1sS+GmaPrT2hbzjiShAzshFDZ\nl9BF4w45mJ/jVsrCsi79pX/Er9x01VSxgohRnO3/h0FPu1O5llPD6mGSZ1mJH35/\nFxXiYp4Jfl5WV4BDN8PPZrbk/n4a+bckFHwrmHozKvuT2ELRZqThAgMBAAGjgZMw\ngZAwCwYDVR0PBAQDAgWgMBQGA1UdEQQNMAuCCWxvY2FsaG9zdDAdBgNVHSUEFjAU\nBggrBgEFBQcDAQYIKwYBBQUHAwIwDAYDVR0TAQH/BAIwADAdBgNVHQ4EFgQUt8Ne\nGSHuq8g+jFS/xva/E+pO5rswHwYDVR0jBBgwFoAU6lgKTjgHbRLHq+q4Yl8HemVa\nW6gwDQYJKoZIhvcNAQELBQADggEBAEbHNoWEUfdUVpL0e1Xq0dRrUmiJ+irGDnAc\ncYLZQwBB4KIoH6D3CNueFruuR6lD7wtmSObfCSvK4afawROMgXMrUQ+Bpdd7iITI\nTByJzl5iCC45f5smd5/gGMP7TjwdFyVaOnQWFr4uyo8dTuqjcFqVZZK1zEYfT7hc\n0Xz8bIFHsEbHgMRhHqc9PCoLDw1zm+x9UcbKwzJQdMs6DJZ9scREh9rh1ccB1n1m\nUduk+aTuhQu47gcGUKb38jilXVYVo+8HMAd05nvSxyNkdkYB35PD3ieC0ZJx9W2R\nN1BI9qrvzeK+8iHIqdE42KebK2kQ2ZgNjKcI44hBKJs/inDddE4=\n-----END CERTIFICATE-----\n'}) => { "ansible_loop_var": "item", "changed": false, "cmd": [ "getcert", "stop-tracking", "-f", "/etc/pki/tls/certs/quadlet_demo.crt" ], "delta": "0:00:00.030336", "end": "2024-07-10 05:12:09.403801", "item": { "ca": "/etc/pki/tls/certs/quadlet_demo.crt", "ca_content": "-----BEGIN CERTIFICATE-----\nMIIDgjCCAmqgAwIBAgIQd4uYHvPLSJW8zGeAkDXEtjANBgkqhkiG9w0BAQsFADBQ\nMSAwHgYDVQQDDBdMb2NhbCBTaWduaW5nIEF1dGhvcml0eTEsMCoGA1UEAwwjNzc4\nYjk4MWUtZjNjYjQ4OTUtYmNjYzY3ODAtOTAzNWM0YjUwHhcNMjQwNzEwMDUxMjA4\nWhcNMjUwNzEwMDUxMjA3WjAUMRIwEAYDVQQDEwlsb2NhbGhvc3QwggEiMA0GCSqG\nSIb3DQEBAQUAA4IBDwAwggEKAoIBAQCYGZKdmimVEA5RpFhaV4ANPTYDdoDjwh15\nkdlNZfGAntuRAOn05085++50QfhomMryRl9TpwJhdJ1Ot6FEOaOieFZtaaHinFhY\nMVcIPPV2RgF1WzIyPC63Q1jClHTt/1goFMeLnvkuJvJFeFoVOlpm3rBTbYFOnvzq\nbErOCC5Sq+JyWwXbBmcnqR4ss0XxdTg5yKAqSwm1sS+GmaPrT2hbzjiShAzshFDZ\nl9BF4w45mJ/jVsrCsi79pX/Er9x01VSxgohRnO3/h0FPu1O5llPD6mGSZ1mJH35/\nFxXiYp4Jfl5WV4BDN8PPZrbk/n4a+bckFHwrmHozKvuT2ELRZqThAgMBAAGjgZMw\ngZAwCwYDVR0PBAQDAgWgMBQGA1UdEQQNMAuCCWxvY2FsaG9zdDAdBgNVHSUEFjAU\nBggrBgEFBQcDAQYIKwYBBQUHAwIwDAYDVR0TAQH/BAIwADAdBgNVHQ4EFgQUt8Ne\nGSHuq8g+jFS/xva/E+pO5rswHwYDVR0jBBgwFoAU6lgKTjgHbRLHq+q4Yl8HemVa\nW6gwDQYJKoZIhvcNAQELBQADggEBAEbHNoWEUfdUVpL0e1Xq0dRrUmiJ+irGDnAc\ncYLZQwBB4KIoH6D3CNueFruuR6lD7wtmSObfCSvK4afawROMgXMrUQ+Bpdd7iITI\nTByJzl5iCC45f5smd5/gGMP7TjwdFyVaOnQWFr4uyo8dTuqjcFqVZZK1zEYfT7hc\n0Xz8bIFHsEbHgMRhHqc9PCoLDw1zm+x9UcbKwzJQdMs6DJZ9scREh9rh1ccB1n1m\nUduk+aTuhQu47gcGUKb38jilXVYVo+8HMAd05nvSxyNkdkYB35PD3ieC0ZJx9W2R\nN1BI9qrvzeK+8iHIqdE42KebK2kQ2ZgNjKcI44hBKJs/inDddE4=\n-----END CERTIFICATE-----\n", "cert": "/etc/pki/tls/certs/quadlet_demo.crt", "cert_content": "-----BEGIN CERTIFICATE-----\nMIIDgjCCAmqgAwIBAgIQd4uYHvPLSJW8zGeAkDXEtjANBgkqhkiG9w0BAQsFADBQ\nMSAwHgYDVQQDDBdMb2NhbCBTaWduaW5nIEF1dGhvcml0eTEsMCoGA1UEAwwjNzc4\nYjk4MWUtZjNjYjQ4OTUtYmNjYzY3ODAtOTAzNWM0YjUwHhcNMjQwNzEwMDUxMjA4\nWhcNMjUwNzEwMDUxMjA3WjAUMRIwEAYDVQQDEwlsb2NhbGhvc3QwggEiMA0GCSqG\nSIb3DQEBAQUAA4IBDwAwggEKAoIBAQCYGZKdmimVEA5RpFhaV4ANPTYDdoDjwh15\nkdlNZfGAntuRAOn05085++50QfhomMryRl9TpwJhdJ1Ot6FEOaOieFZtaaHinFhY\nMVcIPPV2RgF1WzIyPC63Q1jClHTt/1goFMeLnvkuJvJFeFoVOlpm3rBTbYFOnvzq\nbErOCC5Sq+JyWwXbBmcnqR4ss0XxdTg5yKAqSwm1sS+GmaPrT2hbzjiShAzshFDZ\nl9BF4w45mJ/jVsrCsi79pX/Er9x01VSxgohRnO3/h0FPu1O5llPD6mGSZ1mJH35/\nFxXiYp4Jfl5WV4BDN8PPZrbk/n4a+bckFHwrmHozKvuT2ELRZqThAgMBAAGjgZMw\ngZAwCwYDVR0PBAQDAgWgMBQGA1UdEQQNMAuCCWxvY2FsaG9zdDAdBgNVHSUEFjAU\nBggrBgEFBQcDAQYIKwYBBQUHAwIwDAYDVR0TAQH/BAIwADAdBgNVHQ4EFgQUt8Ne\nGSHuq8g+jFS/xva/E+pO5rswHwYDVR0jBBgwFoAU6lgKTjgHbRLHq+q4Yl8HemVa\nW6gwDQYJKoZIhvcNAQELBQADggEBAEbHNoWEUfdUVpL0e1Xq0dRrUmiJ+irGDnAc\ncYLZQwBB4KIoH6D3CNueFruuR6lD7wtmSObfCSvK4afawROMgXMrUQ+Bpdd7iITI\nTByJzl5iCC45f5smd5/gGMP7TjwdFyVaOnQWFr4uyo8dTuqjcFqVZZK1zEYfT7hc\n0Xz8bIFHsEbHgMRhHqc9PCoLDw1zm+x9UcbKwzJQdMs6DJZ9scREh9rh1ccB1n1m\nUduk+aTuhQu47gcGUKb38jilXVYVo+8HMAd05nvSxyNkdkYB35PD3ieC0ZJx9W2R\nN1BI9qrvzeK+8iHIqdE42KebK2kQ2ZgNjKcI44hBKJs/inDddE4=\n-----END CERTIFICATE-----\n", "key": "/etc/pki/tls/private/quadlet_demo.key", "key_content": "-----BEGIN PRIVATE KEY-----\nMIIEvQIBADANBgkqhkiG9w0BAQEFAASCBKcwggSjAgEAAoIBAQCYGZKdmimVEA5R\npFhaV4ANPTYDdoDjwh15kdlNZfGAntuRAOn05085++50QfhomMryRl9TpwJhdJ1O\nt6FEOaOieFZtaaHinFhYMVcIPPV2RgF1WzIyPC63Q1jClHTt/1goFMeLnvkuJvJF\neFoVOlpm3rBTbYFOnvzqbErOCC5Sq+JyWwXbBmcnqR4ss0XxdTg5yKAqSwm1sS+G\nmaPrT2hbzjiShAzshFDZl9BF4w45mJ/jVsrCsi79pX/Er9x01VSxgohRnO3/h0FP\nu1O5llPD6mGSZ1mJH35/FxXiYp4Jfl5WV4BDN8PPZrbk/n4a+bckFHwrmHozKvuT\n2ELRZqThAgMBAAECggEAALS6Lix2cdJdMaZw0UsHjXAtu3CB7sSycrNxSfNdnohm\nOh8iJtyJvOn5IG2jtbcTVpJqa/rKB1olZoWSTFz04Cq9MnTvfxv1CkRA3Fk9lazj\nofgAVLS8LgY0T9Nk80uYjPeTOUf9rbfZLya8ys1OePyivoFjLLJu17b01i3aKrCv\nFtDq56fM0b1okul5AKTaEInSOLWqr+CAbqZKgJH4GCvnYDnFesoWZH/bIjPnkNSj\nPNDPeQ0bJlPIQTfgbY3n6SxYrF7MiyugUs4NYv2+5x57avCg/pErKtslas3vAEyB\nk9ITm7WAPgkOEi0G1PwNqslZWQm65GHExYASeDPQgQKBgQDQXljd7zhDAyAwOTnW\nCOQeTq5EH0jQo66/4c+cuybyhsv+eGP49BgJGtO/8JYbxs64Vdu64cA54ts1L00g\n+ld/0qTnllBJXtS8ac8Jw7Ec0TNWdGyd+i2iHyq+pLarQlXbP1JJQ71gnuBmUId7\nxZKWNWIr+AlZntZ5hSsbaRn+IQKBgQC63md8fdu7pa84ejG2FTAyJ19cfY71wl+X\ni1qMRolrdPbTX0Y//09AC4/TeqqcTbFjdoHQnPkm+N5Qr0JQO/xcnW0kalHsPasA\nq1OyBIL+/cCjWh0VIKo737z7ZEy/7erq2yNchhezKYqyOFWMqOvrQhrs1N4jKvGm\nfhuVvWBOwQKBgQC9AYfMsKyxJImzFIWg32nFEAvFITeUgZUG2tbOzXizr5Ak4KGW\ndhvDToFa7unKgliLHt50s0gBegTrRl/JFSv/d1w78b4b/5gq3DOAIPFsOVs/BdZs\nqj1wfKLyUUidKkvk15hVesuydAsK4+sxW3LvK7dEOPaKQ6k6i5AhigGY4QKBgBbD\nJpnsfsqRuRFJGSsT06YYU65hgPuArMI58FKr+OtE+RfwQA9lmxKzFdbt10Bo2lXC\nRrfRlyvfHr13S72YQPpLFtNbvBTpwh23wN8rW4WSBVggjvBW5UWb0FpS8mJiqb3n\n6ZvInbRuN/3TjOAwZ2Rlzf/RQinIfl0IeXW49rVBAoGAXXT7RLnjVCvUBo8mijT3\neRnp4umJoBQ94NMu70qWJWnb1ZMrRiL4vDwrLfMuaHX3lcQI7bIfJu2vFjdfNv2N\n5TbULs/uhNueXphYF0iWoNU6l/y5Ni627zPQ7q1jazE4F8NjKqk6R9hYcU85mIgv\nkd0bEXrDm3YGznRZlVAm8Ds=\n-----END PRIVATE KEY-----\n" }, "rc": 0, "start": "2024-07-10 05:12:09.373465" } STDOUT: Request "20240710051208" removed. TASK [fedora.linux_system_roles.certificate : Remove files] ******************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:181 Wednesday 10 July 2024 05:12:09 +0000 (0:00:00.314) 0:00:14.094 ******** changed: [sut] => (item=/etc/pki/tls/certs/quadlet_demo.crt) => { "ansible_loop_var": "item", "changed": true, "item": "/etc/pki/tls/certs/quadlet_demo.crt", "path": "/etc/pki/tls/certs/quadlet_demo.crt", "state": "absent" } changed: [sut] => (item=/etc/pki/tls/private/quadlet_demo.key) => { "ansible_loop_var": "item", "changed": true, "item": "/etc/pki/tls/private/quadlet_demo.key", "path": "/etc/pki/tls/private/quadlet_demo.key", "state": "absent" } ok: [sut] => (item=/etc/pki/tls/certs/quadlet_demo.crt) => { "ansible_loop_var": "item", "changed": false, "item": "/etc/pki/tls/certs/quadlet_demo.crt", "path": "/etc/pki/tls/certs/quadlet_demo.crt", "state": "absent" } TASK [Run the role] ************************************************************ task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/tests/tests_quadlet_demo.yml:50 Wednesday 10 July 2024 05:12:10 +0000 (0:00:00.673) 0:00:14.768 ******** TASK [fedora.linux_system_roles.podman : Set platform/version specific variables] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:3 Wednesday 10 July 2024 05:12:10 +0000 (0:00:00.032) 0:00:14.800 ******** included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml for sut TASK [fedora.linux_system_roles.podman : Ensure ansible_facts used by role] **** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:3 Wednesday 10 July 2024 05:12:10 +0000 (0:00:00.015) 0:00:14.816 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_required_facts | difference(ansible_facts.keys() | list) | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if system is ostree] ************ task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:11 Wednesday 10 July 2024 05:12:10 +0000 (0:00:00.017) 0:00:14.833 ******** ok: [sut] => { "changed": false, "stat": { "exists": false } } TASK [fedora.linux_system_roles.podman : Set flag to indicate system is ostree] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:16 Wednesday 10 July 2024 05:12:10 +0000 (0:00:00.222) 0:00:15.056 ******** ok: [sut] => { "ansible_facts": { "__podman_is_ostree": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Set platform/version specific variables] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:20 Wednesday 10 July 2024 05:12:10 +0000 (0:00:00.014) 0:00:15.071 ******** ok: [sut] => (item=RedHat.yml) => { "ansible_facts": { "__podman_packages": [ "podman", "shadow-utils-subid" ] }, "ansible_included_var_files": [ "/WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/vars/RedHat.yml" ], "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml" } ok: [sut] => (item=Fedora.yml) => { "ansible_facts": { "__podman_packages": [ "iptables-nft", "podman", "shadow-utils-subid" ] }, "ansible_included_var_files": [ "/WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/vars/Fedora.yml" ], "ansible_loop_var": "item", "changed": false, "item": "Fedora.yml" } skipping: [sut] => (item=Fedora_39.yml) => { "ansible_loop_var": "item", "changed": false, "false_condition": "__vars_file is file", "item": "Fedora_39.yml", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=Fedora_39.yml) => { "ansible_loop_var": "item", "changed": false, "false_condition": "__vars_file is file", "item": "Fedora_39.yml", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Gather the package facts] ************* task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:6 Wednesday 10 July 2024 05:12:10 +0000 (0:00:00.023) 0:00:15.095 ******** ok: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Enable copr if requested] ************* task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:10 Wednesday 10 July 2024 05:12:11 +0000 (0:00:00.892) 0:00:15.987 ******** skipping: [sut] => { "changed": false, "false_condition": "podman_use_copr | d(false)", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Ensure required packages are installed] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:14 Wednesday 10 July 2024 05:12:11 +0000 (0:00:00.025) 0:00:16.013 ******** changed: [sut] => { "changed": true, "rc": 0, "results": [ "Installed: passt-0^20240624.g1ee2eca-1.fc39.x86_64", "Installed: passt-selinux-0^20240624.g1ee2eca-1.fc39.noarch", "Installed: shadow-utils-subid-2:4.14.0-2.fc39.x86_64", "Installed: catatonit-0.1.7-18.fc39.x86_64", "Installed: fuse-overlayfs-1.13-1.fc39.x86_64", "Installed: libslirp-4.7.0-4.fc39.x86_64", "Installed: fuse-common-3.16.1-1.fc39.x86_64", "Installed: qemu-user-static-2:8.1.3-5.fc39.x86_64", "Installed: qemu-user-static-aarch64-2:8.1.3-5.fc39.x86_64", "Installed: aardvark-dns-1.11.0-1.fc39.x86_64", "Installed: podman-5:4.9.4-1.fc39.x86_64", "Installed: qemu-user-static-alpha-2:8.1.3-5.fc39.x86_64", "Installed: netavark-1.11.0-1.fc39.x86_64", "Installed: qemu-user-static-arm-2:8.1.3-5.fc39.x86_64", "Installed: criu-3.19-2.fc39.x86_64", "Installed: fuse3-3.16.1-1.fc39.x86_64", "Installed: criu-libs-3.19-2.fc39.x86_64", "Installed: qemu-user-static-cris-2:8.1.3-5.fc39.x86_64", "Installed: qemu-user-static-hexagon-2:8.1.3-5.fc39.x86_64", "Installed: fuse3-libs-3.16.1-1.fc39.x86_64", "Installed: qemu-user-static-hppa-2:8.1.3-5.fc39.x86_64", "Installed: qemu-user-static-loongarch64-2:8.1.3-5.fc39.x86_64", "Installed: qemu-user-static-m68k-2:8.1.3-5.fc39.x86_64", "Installed: qemu-user-static-microblaze-2:8.1.3-5.fc39.x86_64", "Installed: qemu-user-static-mips-2:8.1.3-5.fc39.x86_64", "Installed: qemu-user-static-nios2-2:8.1.3-5.fc39.x86_64", "Installed: qemu-user-static-or1k-2:8.1.3-5.fc39.x86_64", "Installed: qemu-user-static-ppc-2:8.1.3-5.fc39.x86_64", "Installed: qemu-user-static-riscv-2:8.1.3-5.fc39.x86_64", "Installed: qemu-user-static-s390x-2:8.1.3-5.fc39.x86_64", "Installed: crun-1.15-1.fc39.x86_64", "Installed: qemu-user-static-sh4-2:8.1.3-5.fc39.x86_64", "Installed: qemu-user-static-sparc-2:8.1.3-5.fc39.x86_64", "Installed: qemu-user-static-x86-2:8.1.3-5.fc39.x86_64", "Installed: qemu-user-static-xtensa-2:8.1.3-5.fc39.x86_64", "Installed: yajl-2.1.0-22.fc39.x86_64", "Installed: libnet-1.3-1.fc39.x86_64", "Installed: conmon-2:2.1.12-1.fc39.x86_64", "Installed: slirp4netns-1.2.2-1.fc39.x86_64", "Installed: container-selinux-2:2.232.1-1.fc39.noarch", "Installed: containers-common-4:1-99.fc39.noarch", "Installed: containers-common-extra-4:1-99.fc39.noarch", "Installed: gvisor-tap-vsock-6:0.7.3-1.fc39.x86_64", "Installed: gvisor-tap-vsock-gvforwarder-6:0.7.3-1.fc39.x86_64", "Installed: libbsd-0.12.2-3.fc39.x86_64", "Installed: libmd-1.1.0-2.fc39.x86_64" ] } TASK [fedora.linux_system_roles.podman : Get podman version] ******************* task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:22 Wednesday 10 July 2024 05:13:46 +0000 (0:01:34.888) 0:01:50.901 ******** ok: [sut] => { "changed": false, "cmd": [ "podman", "--version" ], "delta": "0:00:00.039043", "end": "2024-07-10 05:13:46.499533", "rc": 0, "start": "2024-07-10 05:13:46.460490" } STDOUT: podman version 4.9.4 TASK [fedora.linux_system_roles.podman : Set podman version] ******************* task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:28 Wednesday 10 July 2024 05:13:46 +0000 (0:00:00.294) 0:01:51.196 ******** ok: [sut] => { "ansible_facts": { "podman_version": "4.9.4" }, "changed": false } TASK [fedora.linux_system_roles.podman : Podman package version must be 4.2 or later] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:32 Wednesday 10 July 2024 05:13:46 +0000 (0:00:00.019) 0:01:51.216 ******** skipping: [sut] => { "changed": false, "false_condition": "podman_version is version(\"4.2\", \"<\")", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Podman package version must be 4.4 or later for quadlet, secrets] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:39 Wednesday 10 July 2024 05:13:46 +0000 (0:00:00.017) 0:01:51.234 ******** skipping: [sut] => { "changed": false, "false_condition": "podman_version is version(\"4.4\", \"<\")", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Podman package version must be 4.4 or later for quadlet, secrets] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:49 Wednesday 10 July 2024 05:13:46 +0000 (0:00:00.045) 0:01:51.279 ******** META: end_host conditional evaluated to False, continuing execution for sut skipping: [sut] => { "skip_reason": "end_host conditional evaluated to False, continuing execution for sut" } MSG: end_host conditional evaluated to false, continuing execution for sut TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:56 Wednesday 10 July 2024 05:13:46 +0000 (0:00:00.045) 0:01:51.325 ******** included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for sut TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Wednesday 10 July 2024 05:13:46 +0000 (0:00:00.053) 0:01:51.378 ******** ok: [sut] => { "ansible_facts": { "getent_passwd": { "root": [ "x", "0", "0", "Super User", "/root", "/bin/bash" ] } }, "changed": false } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Wednesday 10 July 2024 05:13:47 +0000 (0:00:00.301) 0:01:51.680 ******** skipping: [sut] => { "changed": false, "false_condition": "not ansible_facts[\"getent_passwd\"][__podman_user]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Wednesday 10 July 2024 05:13:47 +0000 (0:00:00.020) 0:01:51.701 ******** ok: [sut] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get group information] **************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:28 Wednesday 10 July 2024 05:13:47 +0000 (0:00:00.025) 0:01:51.726 ******** ok: [sut] => { "ansible_facts": { "getent_group": { "root": [ "x", "0", "" ] } }, "changed": false } TASK [fedora.linux_system_roles.podman : Set group name] *********************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:35 Wednesday 10 July 2024 05:13:47 +0000 (0:00:00.235) 0:01:51.962 ******** ok: [sut] => { "ansible_facts": { "__podman_group_name": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:39 Wednesday 10 July 2024 05:13:47 +0000 (0:00:00.023) 0:01:51.986 ******** ok: [sut] => { "changed": false, "stat": { "atime": 1696291200.0, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "72ba28a3be19c467e89ab90f5d21f99b605159a7", "ctime": 1720588377.3007717, "dev": 51714, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 203603, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-pie-executable", "mode": "0755", "mtime": 1696291200.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 16032, "uid": 0, "version": "2092964490", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************ task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:50 Wednesday 10 July 2024 05:13:47 +0000 (0:00:00.234) 0:01:52.221 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check group with getsubids] *********** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:55 Wednesday 10 July 2024 05:13:47 +0000 (0:00:00.017) 0:01:52.239 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:60 Wednesday 10 July 2024 05:13:47 +0000 (0:00:00.017) 0:01:52.256 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subuid file] ********************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:74 Wednesday 10 July 2024 05:13:47 +0000 (0:00:00.017) 0:01:52.273 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subgid file] ********************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:79 Wednesday 10 July 2024 05:13:47 +0000 (0:00:00.015) 0:01:52.289 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:84 Wednesday 10 July 2024 05:13:47 +0000 (0:00:00.015) 0:01:52.305 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:94 Wednesday 10 July 2024 05:13:47 +0000 (0:00:00.015) 0:01:52.320 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] ***** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:101 Wednesday 10 July 2024 05:13:47 +0000 (0:00:00.015) 0:01:52.336 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set config file paths] **************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:62 Wednesday 10 July 2024 05:13:47 +0000 (0:00:00.015) 0:01:52.352 ******** ok: [sut] => { "ansible_facts": { "__podman_container_conf_file": "/etc/containers/containers.conf.d/50-systemroles.conf", "__podman_policy_json_file": "/etc/containers/policy.json", "__podman_registries_conf_file": "/etc/containers/registries.conf.d/50-systemroles.conf", "__podman_storage_conf_file": "/etc/containers/storage.conf" }, "changed": false } TASK [fedora.linux_system_roles.podman : Handle container.conf.d] ************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:71 Wednesday 10 July 2024 05:13:47 +0000 (0:00:00.034) 0:01:52.387 ******** included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_container_conf_d.yml for sut TASK [fedora.linux_system_roles.podman : Ensure containers.d exists] *********** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_container_conf_d.yml:5 Wednesday 10 July 2024 05:13:47 +0000 (0:00:00.033) 0:01:52.420 ******** skipping: [sut] => { "changed": false, "false_condition": "podman_containers_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Update container config file] ********* task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_container_conf_d.yml:13 Wednesday 10 July 2024 05:13:47 +0000 (0:00:00.038) 0:01:52.458 ******** skipping: [sut] => { "changed": false, "false_condition": "podman_containers_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Handle registries.conf.d] ************* task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:74 Wednesday 10 July 2024 05:13:47 +0000 (0:00:00.018) 0:01:52.477 ******** included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_registries_conf_d.yml for sut TASK [fedora.linux_system_roles.podman : Ensure registries.d exists] *********** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_registries_conf_d.yml:5 Wednesday 10 July 2024 05:13:47 +0000 (0:00:00.033) 0:01:52.511 ******** skipping: [sut] => { "changed": false, "false_condition": "podman_registries_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Update registries config file] ******** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_registries_conf_d.yml:13 Wednesday 10 July 2024 05:13:47 +0000 (0:00:00.017) 0:01:52.528 ******** skipping: [sut] => { "changed": false, "false_condition": "podman_registries_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Handle storage.conf] ****************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:77 Wednesday 10 July 2024 05:13:47 +0000 (0:00:00.016) 0:01:52.545 ******** included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_storage_conf.yml for sut TASK [fedora.linux_system_roles.podman : Ensure storage.conf parent dir exists] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_storage_conf.yml:5 Wednesday 10 July 2024 05:13:47 +0000 (0:00:00.034) 0:01:52.579 ******** skipping: [sut] => { "changed": false, "false_condition": "podman_storage_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Update storage config file] *********** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_storage_conf.yml:13 Wednesday 10 July 2024 05:13:47 +0000 (0:00:00.017) 0:01:52.596 ******** skipping: [sut] => { "changed": false, "false_condition": "podman_storage_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Handle policy.json] ******************* task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:80 Wednesday 10 July 2024 05:13:47 +0000 (0:00:00.016) 0:01:52.613 ******** included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml for sut TASK [fedora.linux_system_roles.podman : Ensure policy.json parent dir exists] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:6 Wednesday 10 July 2024 05:13:47 +0000 (0:00:00.036) 0:01:52.649 ******** skipping: [sut] => { "changed": false, "false_condition": "podman_policy_json | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Stat the policy.json file] ************ task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:14 Wednesday 10 July 2024 05:13:48 +0000 (0:00:00.016) 0:01:52.666 ******** skipping: [sut] => { "changed": false, "false_condition": "podman_policy_json | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get the existing policy.json] ********* task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:19 Wednesday 10 July 2024 05:13:48 +0000 (0:00:00.017) 0:01:52.683 ******** skipping: [sut] => { "changed": false, "false_condition": "podman_policy_json | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Write new policy.json file] *********** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:25 Wednesday 10 July 2024 05:13:48 +0000 (0:00:00.016) 0:01:52.700 ******** skipping: [sut] => { "changed": false, "false_condition": "podman_policy_json | length > 0", "skip_reason": "Conditional result was False" } TASK [Manage firewall for specified ports] ************************************* task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:86 Wednesday 10 July 2024 05:13:48 +0000 (0:00:00.017) 0:01:52.717 ******** TASK [fedora.linux_system_roles.firewall : Setup firewalld] ******************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:2 Wednesday 10 July 2024 05:13:48 +0000 (0:00:00.082) 0:01:52.799 ******** included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml for sut TASK [fedora.linux_system_roles.firewall : Ensure ansible_facts used by role] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:2 Wednesday 10 July 2024 05:13:48 +0000 (0:00:00.030) 0:01:52.829 ******** skipping: [sut] => { "changed": false, "false_condition": "__firewall_required_facts | difference(ansible_facts.keys() | list) | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Check if system is ostree] ********** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:10 Wednesday 10 July 2024 05:13:48 +0000 (0:00:00.023) 0:01:52.853 ******** ok: [sut] => { "changed": false, "stat": { "exists": false } } TASK [fedora.linux_system_roles.firewall : Set flag to indicate system is ostree] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:15 Wednesday 10 July 2024 05:13:48 +0000 (0:00:00.231) 0:01:53.085 ******** ok: [sut] => { "ansible_facts": { "__firewall_is_ostree": false }, "changed": false } TASK [fedora.linux_system_roles.firewall : Install firewalld] ****************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:19 Wednesday 10 July 2024 05:13:48 +0000 (0:00:00.022) 0:01:53.107 ******** ok: [sut] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [fedora.linux_system_roles.firewall : Collect service facts] ************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:5 Wednesday 10 July 2024 05:13:50 +0000 (0:00:02.180) 0:01:55.288 ******** skipping: [sut] => { "changed": false, "false_condition": "firewall_disable_conflicting_services | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Attempt to stop and disable conflicting services] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:9 Wednesday 10 July 2024 05:13:50 +0000 (0:00:00.017) 0:01:55.305 ******** skipping: [sut] => (item=nftables) => { "ansible_loop_var": "item", "changed": false, "false_condition": "firewall_disable_conflicting_services | bool", "item": "nftables", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=iptables) => { "ansible_loop_var": "item", "changed": false, "false_condition": "firewall_disable_conflicting_services | bool", "item": "iptables", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=ufw) => { "ansible_loop_var": "item", "changed": false, "false_condition": "firewall_disable_conflicting_services | bool", "item": "ufw", "skip_reason": "Conditional result was False" } skipping: [sut] => { "changed": false } MSG: All items skipped TASK [fedora.linux_system_roles.firewall : Unmask firewalld service] *********** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:22 Wednesday 10 July 2024 05:13:50 +0000 (0:00:00.020) 0:01:55.326 ******** ok: [sut] => { "changed": false, "name": "firewalld", "status": { "AccessSELinuxContext": "system_u:object_r:firewalld_unit_file_t:s0", "ActiveEnterTimestampMonotonic": "0", "ActiveExitTimestampMonotonic": "0", "ActiveState": "inactive", "After": "sysinit.target dbus-broker.service basic.target dbus.socket system.slice polkit.service", "AllowIsolate": "no", "AssertResult": "no", "AssertTimestampMonotonic": "0", "Before": "shutdown.target network-pre.target", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "BusName": "org.fedoraproject.FirewallD1", "CPUAccounting": "yes", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "[not set]", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "yes", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore", "CleanResult": "success", "CollectMode": "inactive", "ConditionResult": "no", "ConditionTimestampMonotonic": "0", "ConfigurationDirectoryMode": "0755", "Conflicts": "ip6tables.service iptables.service nftables.service shutdown.target ebtables.service ipset.service", "ControlGroupId": "0", "ControlPID": "0", "CoredumpFilter": "0x33", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "DefaultStartupMemoryLow": "0", "Delegate": "no", "Description": "firewalld - dynamic firewall daemon", "DevicePolicy": "auto", "Documentation": "\"man:firewalld(1)\"", "DropInPaths": "/usr/lib/systemd/system/service.d/10-timeout-abort.conf", "DynamicUser": "no", "EnvironmentFiles": "/etc/sysconfig/firewalld (ignore_errors=yes)", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "0", "ExecMainStartTimestampMonotonic": "0", "ExecMainStatus": "0", "ExecReload": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecReloadEx": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStart": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStartEx": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExitType": "main", "ExtensionImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FileDescriptorStorePreserve": "restart", "FinalKillSignal": "9", "FragmentPath": "/usr/lib/systemd/system/firewalld.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOReadBytes": "18446744073709551615", "IOReadOperations": "18446744073709551615", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "IOWeight": "[not set]", "IOWriteBytes": "18446744073709551615", "IOWriteOperations": "18446744073709551615", "IPAccounting": "no", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "Id": "firewalld.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestampMonotonic": "0", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "mixed", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "infinity", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "8388608", "LimitMEMLOCKSoft": "8388608", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "524288", "LimitNOFILESoft": "1024", "LimitNPROC": "14756", "LimitNPROCSoft": "14756", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "14756", "LimitSIGPENDINGSoft": "14756", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "0", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "ManagedOOMSwap": "auto", "MemoryAccounting": "yes", "MemoryAvailable": "infinity", "MemoryCurrent": "[not set]", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryKSM": "no", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemoryPressureThresholdUSec": "200ms", "MemoryPressureWatch": "auto", "MemorySwapMax": "infinity", "MemoryZSwapMax": "infinity", "MountAPIVFS": "no", "MountImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAPolicy": "n/a", "Names": "firewalld.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMPolicy": "stop", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "OnSuccessJobMode": "fail", "Perpetual": "no", "PrivateDevices": "no", "PrivateIPC": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProcSubset": "all", "ProtectClock": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectHostname": "no", "ProtectKernelLogs": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectProc": "default", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "ReloadResult": "success", "ReloadSignal": "1", "RemainAfterExit": "no", "RemoveIPC": "no", "Requires": "dbus.socket system.slice sysinit.target", "Restart": "no", "RestartKillSignal": "15", "RestartMaxDelayUSec": "infinity", "RestartMode": "normal", "RestartSteps": "0", "RestartUSec": "100ms", "RestartUSecNext": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RootEphemeral": "no", "RootImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "Slice": "system.slice", "StandardError": "null", "StandardInput": "null", "StandardOutput": "null", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StartupMemoryHigh": "18446744073709551615", "StartupMemoryLow": "0", "StartupMemoryMax": "18446744073709551615", "StartupMemorySwapMax": "18446744073709551615", "StartupMemoryZSwapMax": "18446744073709551615", "StateChangeTimestampMonotonic": "0", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "dead", "SuccessAction": "none", "SyslogFacility": "3", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "2147483646", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "[not set]", "TasksMax": "4426", "TimeoutAbortUSec": "45s", "TimeoutCleanUSec": "infinity", "TimeoutStartFailureMode": "terminate", "TimeoutStartUSec": "45s", "TimeoutStopFailureMode": "abort", "TimeoutStopUSec": "45s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "dbus", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "enabled", "UnitFileState": "disabled", "UtmpMode": "init", "Wants": "network-pre.target", "WatchdogSignal": "6", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "infinity" } } TASK [fedora.linux_system_roles.firewall : Enable and start firewalld service] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:28 Wednesday 10 July 2024 05:13:51 +0000 (0:00:00.416) 0:01:55.742 ******** changed: [sut] => { "changed": true, "enabled": true, "name": "firewalld", "state": "started", "status": { "AccessSELinuxContext": "system_u:object_r:firewalld_unit_file_t:s0", "ActiveEnterTimestampMonotonic": "0", "ActiveExitTimestampMonotonic": "0", "ActiveState": "inactive", "After": "system.slice polkit.service dbus-broker.service basic.target dbus.socket sysinit.target", "AllowIsolate": "no", "AssertResult": "no", "AssertTimestampMonotonic": "0", "Before": "shutdown.target network-pre.target", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "BusName": "org.fedoraproject.FirewallD1", "CPUAccounting": "yes", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "[not set]", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "yes", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore", "CleanResult": "success", "CollectMode": "inactive", "ConditionResult": "no", "ConditionTimestampMonotonic": "0", "ConfigurationDirectoryMode": "0755", "Conflicts": "ebtables.service ipset.service nftables.service ip6tables.service shutdown.target iptables.service", "ControlGroupId": "0", "ControlPID": "0", "CoredumpFilter": "0x33", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "DefaultStartupMemoryLow": "0", "Delegate": "no", "Description": "firewalld - dynamic firewall daemon", "DevicePolicy": "auto", "Documentation": "\"man:firewalld(1)\"", "DropInPaths": "/usr/lib/systemd/system/service.d/10-timeout-abort.conf", "DynamicUser": "no", "EnvironmentFiles": "/etc/sysconfig/firewalld (ignore_errors=yes)", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "0", "ExecMainStartTimestampMonotonic": "0", "ExecMainStatus": "0", "ExecReload": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecReloadEx": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStart": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStartEx": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExitType": "main", "ExtensionImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FileDescriptorStorePreserve": "restart", "FinalKillSignal": "9", "FragmentPath": "/usr/lib/systemd/system/firewalld.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOReadBytes": "18446744073709551615", "IOReadOperations": "18446744073709551615", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "IOWeight": "[not set]", "IOWriteBytes": "18446744073709551615", "IOWriteOperations": "18446744073709551615", "IPAccounting": "no", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "Id": "firewalld.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestampMonotonic": "0", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "mixed", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "infinity", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "8388608", "LimitMEMLOCKSoft": "8388608", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "524288", "LimitNOFILESoft": "1024", "LimitNPROC": "14756", "LimitNPROCSoft": "14756", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "14756", "LimitSIGPENDINGSoft": "14756", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "0", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "ManagedOOMSwap": "auto", "MemoryAccounting": "yes", "MemoryAvailable": "infinity", "MemoryCurrent": "[not set]", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryKSM": "no", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemoryPressureThresholdUSec": "200ms", "MemoryPressureWatch": "auto", "MemorySwapMax": "infinity", "MemoryZSwapMax": "infinity", "MountAPIVFS": "no", "MountImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAPolicy": "n/a", "Names": "firewalld.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMPolicy": "stop", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "OnSuccessJobMode": "fail", "Perpetual": "no", "PrivateDevices": "no", "PrivateIPC": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProcSubset": "all", "ProtectClock": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectHostname": "no", "ProtectKernelLogs": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectProc": "default", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "ReloadResult": "success", "ReloadSignal": "1", "RemainAfterExit": "no", "RemoveIPC": "no", "Requires": "system.slice dbus.socket sysinit.target", "Restart": "no", "RestartKillSignal": "15", "RestartMaxDelayUSec": "infinity", "RestartMode": "normal", "RestartSteps": "0", "RestartUSec": "100ms", "RestartUSecNext": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RootEphemeral": "no", "RootImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "Slice": "system.slice", "StandardError": "null", "StandardInput": "null", "StandardOutput": "null", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StartupMemoryHigh": "18446744073709551615", "StartupMemoryLow": "0", "StartupMemoryMax": "18446744073709551615", "StartupMemorySwapMax": "18446744073709551615", "StartupMemoryZSwapMax": "18446744073709551615", "StateChangeTimestampMonotonic": "0", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "dead", "SuccessAction": "none", "SyslogFacility": "3", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "2147483646", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "[not set]", "TasksMax": "4426", "TimeoutAbortUSec": "45s", "TimeoutCleanUSec": "infinity", "TimeoutStartFailureMode": "terminate", "TimeoutStartUSec": "45s", "TimeoutStopFailureMode": "abort", "TimeoutStopUSec": "45s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "dbus", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "enabled", "UnitFileState": "disabled", "UtmpMode": "init", "Wants": "network-pre.target", "WatchdogSignal": "6", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "infinity" } } TASK [fedora.linux_system_roles.firewall : Check if previous replaced is defined] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:34 Wednesday 10 July 2024 05:13:52 +0000 (0:00:01.093) 0:01:56.836 ******** ok: [sut] => { "ansible_facts": { "__firewall_previous_replaced": false, "__firewall_python_cmd": "/usr/bin/python3", "__firewall_report_changed": true }, "changed": false } TASK [fedora.linux_system_roles.firewall : Get config files, checksums before and remove] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:43 Wednesday 10 July 2024 05:13:52 +0000 (0:00:00.025) 0:01:56.861 ******** skipping: [sut] => { "changed": false, "false_condition": "__firewall_previous_replaced | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Tell firewall module it is able to report changed] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:55 Wednesday 10 July 2024 05:13:52 +0000 (0:00:00.017) 0:01:56.878 ******** skipping: [sut] => { "changed": false, "false_condition": "__firewall_previous_replaced | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Configure firewall] ***************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:71 Wednesday 10 July 2024 05:13:52 +0000 (0:00:00.016) 0:01:56.894 ******** changed: [sut] => (item={'port': '8000/tcp', 'state': 'enabled'}) => { "__firewall_changed": true, "ansible_loop_var": "item", "changed": true, "item": { "port": "8000/tcp", "state": "enabled" } } changed: [sut] => (item={'port': '9000/tcp', 'state': 'enabled'}) => { "__firewall_changed": true, "ansible_loop_var": "item", "changed": true, "item": { "port": "9000/tcp", "state": "enabled" } } TASK [fedora.linux_system_roles.firewall : Gather firewall config information] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:120 Wednesday 10 July 2024 05:13:53 +0000 (0:00:01.168) 0:01:58.063 ******** skipping: [sut] => (item={'port': '8000/tcp', 'state': 'enabled'}) => { "ansible_loop_var": "item", "changed": false, "false_condition": "firewall | length == 1", "item": { "port": "8000/tcp", "state": "enabled" }, "skip_reason": "Conditional result was False" } skipping: [sut] => (item={'port': '9000/tcp', 'state': 'enabled'}) => { "ansible_loop_var": "item", "changed": false, "false_condition": "firewall | length == 1", "item": { "port": "9000/tcp", "state": "enabled" }, "skip_reason": "Conditional result was False" } skipping: [sut] => { "changed": false } MSG: All items skipped TASK [fedora.linux_system_roles.firewall : Update firewalld_config fact] ******* task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:130 Wednesday 10 July 2024 05:13:53 +0000 (0:00:00.028) 0:01:58.092 ******** skipping: [sut] => { "changed": false, "false_condition": "firewall | length == 1", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Gather firewall config if no arguments] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:139 Wednesday 10 July 2024 05:13:53 +0000 (0:00:00.019) 0:01:58.111 ******** skipping: [sut] => { "changed": false, "false_condition": "firewall == None or firewall | length == 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Update firewalld_config fact] ******* task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:144 Wednesday 10 July 2024 05:13:53 +0000 (0:00:00.018) 0:01:58.129 ******** skipping: [sut] => { "changed": false, "false_condition": "firewall == None or firewall | length == 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Get config files, checksums after] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:153 Wednesday 10 July 2024 05:13:53 +0000 (0:00:00.017) 0:01:58.147 ******** skipping: [sut] => { "changed": false, "false_condition": "__firewall_previous_replaced | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Calculate what has changed] ********* task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:163 Wednesday 10 July 2024 05:13:53 +0000 (0:00:00.040) 0:01:58.187 ******** skipping: [sut] => { "changed": false, "false_condition": "__firewall_previous_replaced | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Show diffs] ************************* task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:169 Wednesday 10 July 2024 05:13:53 +0000 (0:00:00.017) 0:01:58.205 ******** skipping: [sut] => { "false_condition": "__firewall_previous_replaced | bool" } TASK [Manage selinux for specified ports] ************************************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:93 Wednesday 10 July 2024 05:13:53 +0000 (0:00:00.026) 0:01:58.231 ******** skipping: [sut] => { "changed": false, "false_condition": "podman_selinux_ports | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Keep track of users that need to cancel linger] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:100 Wednesday 10 July 2024 05:13:53 +0000 (0:00:00.017) 0:01:58.249 ******** ok: [sut] => { "ansible_facts": { "__podman_cancel_user_linger": [] }, "changed": false } TASK [fedora.linux_system_roles.podman : Handle certs.d files - present] ******* task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:104 Wednesday 10 July 2024 05:13:53 +0000 (0:00:00.018) 0:01:58.267 ******** skipping: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Handle credential files - present] **** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:113 Wednesday 10 July 2024 05:13:53 +0000 (0:00:00.014) 0:01:58.282 ******** skipping: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Handle secrets] *********************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:122 Wednesday 10 July 2024 05:13:53 +0000 (0:00:00.014) 0:01:58.297 ******** included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml for sut => (item=(censored due to no_log)) included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml for sut => (item=(censored due to no_log)) included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml for sut => (item=(censored due to no_log)) TASK [fedora.linux_system_roles.podman : Set variables part 1] ***************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:3 Wednesday 10 July 2024 05:13:53 +0000 (0:00:00.068) 0:01:58.365 ******** ok: [sut] => { "ansible_facts": { "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set variables part 2] ***************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:7 Wednesday 10 July 2024 05:13:53 +0000 (0:00:00.020) 0:01:58.386 ******** ok: [sut] => { "ansible_facts": { "__podman_rootless": false, "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:13 Wednesday 10 July 2024 05:13:53 +0000 (0:00:00.022) 0:01:58.409 ******** included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for sut TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Wednesday 10 July 2024 05:13:53 +0000 (0:00:00.027) 0:01:58.436 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Wednesday 10 July 2024 05:13:53 +0000 (0:00:00.016) 0:01:58.453 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Wednesday 10 July 2024 05:13:53 +0000 (0:00:00.016) 0:01:58.469 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Stat XDG_RUNTIME_DIR] ***************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:18 Wednesday 10 July 2024 05:13:53 +0000 (0:00:00.015) 0:01:58.485 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Manage each secret] ******************* task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:34 Wednesday 10 July 2024 05:13:53 +0000 (0:00:00.015) 0:01:58.501 ******** changed: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": true } TASK [fedora.linux_system_roles.podman : Set variables part 1] ***************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:3 Wednesday 10 July 2024 05:13:54 +0000 (0:00:00.582) 0:01:59.083 ******** ok: [sut] => { "ansible_facts": { "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set variables part 2] ***************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:7 Wednesday 10 July 2024 05:13:54 +0000 (0:00:00.021) 0:01:59.104 ******** ok: [sut] => { "ansible_facts": { "__podman_rootless": false, "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:13 Wednesday 10 July 2024 05:13:54 +0000 (0:00:00.024) 0:01:59.129 ******** included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for sut TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Wednesday 10 July 2024 05:13:54 +0000 (0:00:00.028) 0:01:59.157 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Wednesday 10 July 2024 05:13:54 +0000 (0:00:00.017) 0:01:59.175 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Wednesday 10 July 2024 05:13:54 +0000 (0:00:00.017) 0:01:59.192 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Stat XDG_RUNTIME_DIR] ***************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:18 Wednesday 10 July 2024 05:13:54 +0000 (0:00:00.017) 0:01:59.209 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Manage each secret] ******************* task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:34 Wednesday 10 July 2024 05:13:54 +0000 (0:00:00.016) 0:01:59.225 ******** changed: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": true } TASK [fedora.linux_system_roles.podman : Set variables part 1] ***************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:3 Wednesday 10 July 2024 05:13:54 +0000 (0:00:00.369) 0:01:59.595 ******** ok: [sut] => { "ansible_facts": { "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set variables part 2] ***************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:7 Wednesday 10 July 2024 05:13:54 +0000 (0:00:00.020) 0:01:59.616 ******** ok: [sut] => { "ansible_facts": { "__podman_rootless": false, "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:13 Wednesday 10 July 2024 05:13:54 +0000 (0:00:00.024) 0:01:59.641 ******** included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for sut TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Wednesday 10 July 2024 05:13:55 +0000 (0:00:00.028) 0:01:59.669 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Wednesday 10 July 2024 05:13:55 +0000 (0:00:00.017) 0:01:59.687 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Wednesday 10 July 2024 05:13:55 +0000 (0:00:00.016) 0:01:59.703 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Stat XDG_RUNTIME_DIR] ***************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:18 Wednesday 10 July 2024 05:13:55 +0000 (0:00:00.016) 0:01:59.720 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Manage each secret] ******************* task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:34 Wednesday 10 July 2024 05:13:55 +0000 (0:00:00.016) 0:01:59.736 ******** changed: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": true } TASK [fedora.linux_system_roles.podman : Handle Kubernetes specifications] ***** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:129 Wednesday 10 July 2024 05:13:55 +0000 (0:00:00.379) 0:02:00.116 ******** skipping: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Handle Quadlet specifications] ******** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:136 Wednesday 10 July 2024 05:13:55 +0000 (0:00:00.042) 0:02:00.159 ******** included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for sut => (item=(censored due to no_log)) included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for sut => (item=(censored due to no_log)) included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for sut => (item=(censored due to no_log)) included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for sut => (item=(censored due to no_log)) included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for sut => (item=(censored due to no_log)) included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for sut => (item=(censored due to no_log)) TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14 Wednesday 10 July 2024 05:13:55 +0000 (0:00:00.084) 0:02:00.243 ******** ok: [sut] => { "ansible_facts": { "__podman_quadlet_file_src": "quadlet-demo.network", "__podman_quadlet_spec": {}, "__podman_quadlet_str": "[Network]\nSubnet=192.168.30.0/24\nGateway=192.168.30.1\nLabel=app=wordpress", "__podman_quadlet_template_src": "" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25 Wednesday 10 July 2024 05:13:55 +0000 (0:00:00.027) 0:02:00.271 ******** ok: [sut] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_pull_image": true, "__podman_state": "created", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] ***** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35 Wednesday 10 July 2024 05:13:55 +0000 (0:00:00.023) 0:02:00.294 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_quadlet_file_src", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48 Wednesday 10 July 2024 05:13:55 +0000 (0:00:00.018) 0:02:00.313 ******** ok: [sut] => { "ansible_facts": { "__podman_quadlet_name": "quadlet-demo", "__podman_quadlet_type": "network", "__podman_rootless": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57 Wednesday 10 July 2024 05:13:55 +0000 (0:00:00.028) 0:02:00.342 ******** included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for sut TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Wednesday 10 July 2024 05:13:55 +0000 (0:00:00.034) 0:02:00.376 ******** skipping: [sut] => { "changed": false, "false_condition": "'getent_passwd' not in ansible_facts or __podman_user not in ansible_facts['getent_passwd']", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Wednesday 10 July 2024 05:13:55 +0000 (0:00:00.021) 0:02:00.397 ******** skipping: [sut] => { "changed": false, "false_condition": "not ansible_facts[\"getent_passwd\"][__podman_user]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Wednesday 10 July 2024 05:13:55 +0000 (0:00:00.020) 0:02:00.418 ******** ok: [sut] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get group information] **************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:28 Wednesday 10 July 2024 05:13:55 +0000 (0:00:00.027) 0:02:00.446 ******** ok: [sut] => { "ansible_facts": { "getent_group": { "root": [ "x", "0", "" ] } }, "changed": false } TASK [fedora.linux_system_roles.podman : Set group name] *********************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:35 Wednesday 10 July 2024 05:13:56 +0000 (0:00:00.239) 0:02:00.685 ******** ok: [sut] => { "ansible_facts": { "__podman_group_name": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:39 Wednesday 10 July 2024 05:13:56 +0000 (0:00:00.024) 0:02:00.710 ******** ok: [sut] => { "changed": false, "stat": { "atime": 1720588427.5164855, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "72ba28a3be19c467e89ab90f5d21f99b605159a7", "ctime": 1720588377.3007717, "dev": 51714, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 203603, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-pie-executable", "mode": "0755", "mtime": 1696291200.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 16032, "uid": 0, "version": "2092964490", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************ task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:50 Wednesday 10 July 2024 05:13:56 +0000 (0:00:00.235) 0:02:00.945 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check group with getsubids] *********** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:55 Wednesday 10 July 2024 05:13:56 +0000 (0:00:00.017) 0:02:00.963 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:60 Wednesday 10 July 2024 05:13:56 +0000 (0:00:00.017) 0:02:00.980 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subuid file] ********************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:74 Wednesday 10 July 2024 05:13:56 +0000 (0:00:00.016) 0:02:00.997 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subgid file] ********************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:79 Wednesday 10 July 2024 05:13:56 +0000 (0:00:00.041) 0:02:01.038 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:84 Wednesday 10 July 2024 05:13:56 +0000 (0:00:00.018) 0:02:01.056 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:94 Wednesday 10 July 2024 05:13:56 +0000 (0:00:00.016) 0:02:01.073 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] ***** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:101 Wednesday 10 July 2024 05:13:56 +0000 (0:00:00.017) 0:02:01.091 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62 Wednesday 10 July 2024 05:13:56 +0000 (0:00:00.016) 0:02:01.107 ******** ok: [sut] => { "ansible_facts": { "__podman_activate_systemd_unit": true, "__podman_images_found": [], "__podman_kube_yamls_raw": "", "__podman_service_name": "quadlet-demo-network.service", "__podman_systemd_scope": "system", "__podman_user_home_dir": "/root", "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73 Wednesday 10 July 2024 05:13:56 +0000 (0:00:00.033) 0:02:01.141 ******** ok: [sut] => { "ansible_facts": { "__podman_quadlet_path": "/etc/containers/systemd" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get kube yaml contents] *************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77 Wednesday 10 July 2024 05:13:56 +0000 (0:00:00.019) 0:02:01.161 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_kube_yamls_raw | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87 Wednesday 10 July 2024 05:13:56 +0000 (0:00:00.018) 0:02:01.179 ******** ok: [sut] => { "ansible_facts": { "__podman_images": [], "__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo.network", "__podman_volumes": [] }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 6] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:103 Wednesday 10 July 2024 05:13:56 +0000 (0:00:00.044) 0:02:01.224 ******** ok: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Cleanup quadlets] ********************* task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:110 Wednesday 10 July 2024 05:13:56 +0000 (0:00:00.022) 0:02:01.246 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_state == \"absent\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create and update quadlets] *********** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:114 Wednesday 10 July 2024 05:13:56 +0000 (0:00:00.017) 0:02:01.263 ******** included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml for sut TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:2 Wednesday 10 July 2024 05:13:56 +0000 (0:00:00.038) 0:02:01.302 ******** included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for sut TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Wednesday 10 July 2024 05:13:56 +0000 (0:00:00.030) 0:02:01.332 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Wednesday 10 July 2024 05:13:56 +0000 (0:00:00.017) 0:02:01.350 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Wednesday 10 July 2024 05:13:56 +0000 (0:00:00.017) 0:02:01.367 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create host directories] ************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:7 Wednesday 10 July 2024 05:13:56 +0000 (0:00:00.016) 0:02:01.383 ******** skipping: [sut] => { "changed": false, "skipped_reason": "No items in the list" } TASK [fedora.linux_system_roles.podman : Ensure container images are present] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:18 Wednesday 10 July 2024 05:13:56 +0000 (0:00:00.042) 0:02:01.425 ******** skipping: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Ensure the quadlet directory is present] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:39 Wednesday 10 July 2024 05:13:56 +0000 (0:00:00.018) 0:02:01.443 ******** ok: [sut] => { "changed": false, "gid": 0, "group": "root", "mode": "0755", "owner": "root", "path": "/etc/containers/systemd", "secontext": "system_u:object_r:etc_t:s0", "size": 4096, "state": "directory", "uid": 0 } TASK [fedora.linux_system_roles.podman : Ensure quadlet file is copied] ******** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:48 Wednesday 10 July 2024 05:13:57 +0000 (0:00:00.250) 0:02:01.694 ******** changed: [sut] => { "changed": true, "checksum": "e57c08d49aff4bae8daab138d913aeddaa8682a0", "dest": "/etc/containers/systemd/quadlet-demo.network", "gid": 0, "group": "root", "md5sum": "061f3cf318cbd8ab5794bb1173831fb8", "mode": "0644", "owner": "root", "secontext": "system_u:object_r:etc_t:s0", "size": 74, "src": "/root/.ansible/tmp/ansible-tmp-1720588437.067682-11949-230791674364504/source", "state": "file", "uid": 0 } TASK [fedora.linux_system_roles.podman : Ensure quadlet file content is present] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:58 Wednesday 10 July 2024 05:13:57 +0000 (0:00:00.652) 0:02:02.347 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_quadlet_file_src", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Ensure quadlet file is present] ******* task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:70 Wednesday 10 July 2024 05:13:57 +0000 (0:00:00.020) 0:02:02.368 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_copy_file is skipped", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Reload systemctl] ********************* task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:82 Wednesday 10 July 2024 05:13:57 +0000 (0:00:00.017) 0:02:02.386 ******** ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.podman : Start service] ************************ task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:110 Wednesday 10 July 2024 05:13:58 +0000 (0:00:00.798) 0:02:03.184 ******** changed: [sut] => { "changed": true, "name": "quadlet-demo-network.service", "state": "started", "status": { "AccessSELinuxContext": "system_u:object_r:systemd_unit_file_t:s0", "ActiveEnterTimestampMonotonic": "0", "ActiveExitTimestampMonotonic": "0", "ActiveState": "inactive", "After": "-.mount systemd-journald.socket basic.target system.slice sysinit.target", "AllowIsolate": "no", "AssertResult": "no", "AssertTimestampMonotonic": "0", "Before": "shutdown.target", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "CPUAccounting": "yes", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "[not set]", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "no", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore", "CleanResult": "success", "CollectMode": "inactive", "ConditionResult": "no", "ConditionTimestampMonotonic": "0", "ConfigurationDirectoryMode": "0755", "Conflicts": "shutdown.target", "ControlGroupId": "0", "ControlPID": "0", "CoredumpFilter": "0x33", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "DefaultStartupMemoryLow": "0", "Delegate": "no", "Description": "quadlet-demo-network.service", "DevicePolicy": "auto", "DropInPaths": "/usr/lib/systemd/system/service.d/10-timeout-abort.conf", "DynamicUser": "no", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "0", "ExecMainStartTimestampMonotonic": "0", "ExecMainStatus": "0", "ExecStart": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman network create --ignore --subnet=192.168.30.0/24 --gateway=192.168.30.1 --label app=wordpress systemd-quadlet-demo ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStartEx": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman network create --ignore --subnet=192.168.30.0/24 --gateway=192.168.30.1 --label app=wordpress systemd-quadlet-demo ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExitType": "main", "ExtensionImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FileDescriptorStorePreserve": "restart", "FinalKillSignal": "9", "FragmentPath": "/run/systemd/generator/quadlet-demo-network.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOReadBytes": "18446744073709551615", "IOReadOperations": "18446744073709551615", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "IOWeight": "[not set]", "IOWriteBytes": "18446744073709551615", "IOWriteOperations": "18446744073709551615", "IPAccounting": "no", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "Id": "quadlet-demo-network.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestampMonotonic": "0", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "control-group", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "infinity", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "8388608", "LimitMEMLOCKSoft": "8388608", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "524288", "LimitNOFILESoft": "1024", "LimitNPROC": "14756", "LimitNPROCSoft": "14756", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "14756", "LimitSIGPENDINGSoft": "14756", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "0", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "ManagedOOMSwap": "auto", "MemoryAccounting": "yes", "MemoryAvailable": "infinity", "MemoryCurrent": "[not set]", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryKSM": "no", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemoryPressureThresholdUSec": "200ms", "MemoryPressureWatch": "auto", "MemorySwapMax": "infinity", "MemoryZSwapMax": "infinity", "MountAPIVFS": "no", "MountImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAPolicy": "n/a", "Names": "quadlet-demo-network.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMPolicy": "stop", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "OnSuccessJobMode": "fail", "Perpetual": "no", "PrivateDevices": "no", "PrivateIPC": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProcSubset": "all", "ProtectClock": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectHostname": "no", "ProtectKernelLogs": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectProc": "default", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "ReloadResult": "success", "ReloadSignal": "1", "RemainAfterExit": "yes", "RemoveIPC": "no", "Requires": "-.mount system.slice sysinit.target", "RequiresMountsFor": "/run/containers", "Restart": "no", "RestartKillSignal": "15", "RestartMaxDelayUSec": "infinity", "RestartMode": "normal", "RestartSteps": "0", "RestartUSec": "100ms", "RestartUSecNext": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RootEphemeral": "no", "RootImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "Slice": "system.slice", "StandardError": "inherit", "StandardInput": "null", "StandardOutput": "journal", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StartupMemoryHigh": "18446744073709551615", "StartupMemoryLow": "0", "StartupMemoryMax": "18446744073709551615", "StartupMemorySwapMax": "18446744073709551615", "StartupMemoryZSwapMax": "18446744073709551615", "StateChangeTimestampMonotonic": "0", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "dead", "SuccessAction": "none", "SyslogFacility": "3", "SyslogIdentifier": "quadlet-demo-network", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "2147483646", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "[not set]", "TasksMax": "4426", "TimeoutAbortUSec": "45s", "TimeoutCleanUSec": "infinity", "TimeoutStartFailureMode": "terminate", "TimeoutStartUSec": "infinity", "TimeoutStopFailureMode": "abort", "TimeoutStopUSec": "45s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "oneshot", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "disabled", "UnitFileState": "generated", "UtmpMode": "init", "WatchdogSignal": "6", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "infinity" } } TASK [fedora.linux_system_roles.podman : Restart service] ********************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:125 Wednesday 10 July 2024 05:13:58 +0000 (0:00:00.464) 0:02:03.648 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_service_started is changed", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14 Wednesday 10 July 2024 05:13:59 +0000 (0:00:00.018) 0:02:03.667 ******** ok: [sut] => { "ansible_facts": { "__podman_quadlet_file_src": "quadlet-demo-mysql.volume", "__podman_quadlet_spec": {}, "__podman_quadlet_str": "[Volume]", "__podman_quadlet_template_src": "" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25 Wednesday 10 July 2024 05:13:59 +0000 (0:00:00.027) 0:02:03.695 ******** ok: [sut] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_pull_image": true, "__podman_state": "created", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] ***** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35 Wednesday 10 July 2024 05:13:59 +0000 (0:00:00.024) 0:02:03.719 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_quadlet_file_src", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48 Wednesday 10 July 2024 05:13:59 +0000 (0:00:00.017) 0:02:03.737 ******** ok: [sut] => { "ansible_facts": { "__podman_quadlet_name": "quadlet-demo-mysql", "__podman_quadlet_type": "volume", "__podman_rootless": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57 Wednesday 10 July 2024 05:13:59 +0000 (0:00:00.028) 0:02:03.765 ******** included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for sut TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Wednesday 10 July 2024 05:13:59 +0000 (0:00:00.034) 0:02:03.800 ******** skipping: [sut] => { "changed": false, "false_condition": "'getent_passwd' not in ansible_facts or __podman_user not in ansible_facts['getent_passwd']", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Wednesday 10 July 2024 05:13:59 +0000 (0:00:00.020) 0:02:03.821 ******** skipping: [sut] => { "changed": false, "false_condition": "not ansible_facts[\"getent_passwd\"][__podman_user]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Wednesday 10 July 2024 05:13:59 +0000 (0:00:00.019) 0:02:03.841 ******** ok: [sut] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get group information] **************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:28 Wednesday 10 July 2024 05:13:59 +0000 (0:00:00.026) 0:02:03.867 ******** ok: [sut] => { "ansible_facts": { "getent_group": { "root": [ "x", "0", "" ] } }, "changed": false } TASK [fedora.linux_system_roles.podman : Set group name] *********************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:35 Wednesday 10 July 2024 05:13:59 +0000 (0:00:00.237) 0:02:04.105 ******** ok: [sut] => { "ansible_facts": { "__podman_group_name": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:39 Wednesday 10 July 2024 05:13:59 +0000 (0:00:00.054) 0:02:04.159 ******** ok: [sut] => { "changed": false, "stat": { "atime": 1720588427.5164855, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "72ba28a3be19c467e89ab90f5d21f99b605159a7", "ctime": 1720588377.3007717, "dev": 51714, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 203603, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-pie-executable", "mode": "0755", "mtime": 1696291200.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 16032, "uid": 0, "version": "2092964490", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************ task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:50 Wednesday 10 July 2024 05:13:59 +0000 (0:00:00.237) 0:02:04.397 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check group with getsubids] *********** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:55 Wednesday 10 July 2024 05:13:59 +0000 (0:00:00.017) 0:02:04.415 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:60 Wednesday 10 July 2024 05:13:59 +0000 (0:00:00.017) 0:02:04.433 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subuid file] ********************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:74 Wednesday 10 July 2024 05:13:59 +0000 (0:00:00.016) 0:02:04.450 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subgid file] ********************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:79 Wednesday 10 July 2024 05:13:59 +0000 (0:00:00.017) 0:02:04.467 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:84 Wednesday 10 July 2024 05:13:59 +0000 (0:00:00.016) 0:02:04.484 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:94 Wednesday 10 July 2024 05:13:59 +0000 (0:00:00.017) 0:02:04.501 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] ***** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:101 Wednesday 10 July 2024 05:13:59 +0000 (0:00:00.017) 0:02:04.518 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62 Wednesday 10 July 2024 05:13:59 +0000 (0:00:00.017) 0:02:04.535 ******** ok: [sut] => { "ansible_facts": { "__podman_activate_systemd_unit": true, "__podman_images_found": [], "__podman_kube_yamls_raw": "", "__podman_service_name": "quadlet-demo-mysql-volume.service", "__podman_systemd_scope": "system", "__podman_user_home_dir": "/root", "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73 Wednesday 10 July 2024 05:13:59 +0000 (0:00:00.033) 0:02:04.569 ******** ok: [sut] => { "ansible_facts": { "__podman_quadlet_path": "/etc/containers/systemd" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get kube yaml contents] *************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77 Wednesday 10 July 2024 05:13:59 +0000 (0:00:00.019) 0:02:04.588 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_kube_yamls_raw | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87 Wednesday 10 July 2024 05:13:59 +0000 (0:00:00.018) 0:02:04.606 ******** ok: [sut] => { "ansible_facts": { "__podman_images": [], "__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo-mysql.volume", "__podman_volumes": [] }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 6] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:103 Wednesday 10 July 2024 05:13:59 +0000 (0:00:00.045) 0:02:04.652 ******** ok: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Cleanup quadlets] ********************* task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:110 Wednesday 10 July 2024 05:14:00 +0000 (0:00:00.023) 0:02:04.675 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_state == \"absent\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create and update quadlets] *********** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:114 Wednesday 10 July 2024 05:14:00 +0000 (0:00:00.017) 0:02:04.693 ******** included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml for sut TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:2 Wednesday 10 July 2024 05:14:00 +0000 (0:00:00.037) 0:02:04.731 ******** included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for sut TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Wednesday 10 July 2024 05:14:00 +0000 (0:00:00.056) 0:02:04.787 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Wednesday 10 July 2024 05:14:00 +0000 (0:00:00.019) 0:02:04.806 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Wednesday 10 July 2024 05:14:00 +0000 (0:00:00.017) 0:02:04.823 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create host directories] ************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:7 Wednesday 10 July 2024 05:14:00 +0000 (0:00:00.017) 0:02:04.841 ******** skipping: [sut] => { "changed": false, "skipped_reason": "No items in the list" } TASK [fedora.linux_system_roles.podman : Ensure container images are present] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:18 Wednesday 10 July 2024 05:14:00 +0000 (0:00:00.016) 0:02:04.858 ******** skipping: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Ensure the quadlet directory is present] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:39 Wednesday 10 July 2024 05:14:00 +0000 (0:00:00.017) 0:02:04.875 ******** ok: [sut] => { "changed": false, "gid": 0, "group": "root", "mode": "0755", "owner": "root", "path": "/etc/containers/systemd", "secontext": "system_u:object_r:etc_t:s0", "size": 4096, "state": "directory", "uid": 0 } TASK [fedora.linux_system_roles.podman : Ensure quadlet file is copied] ******** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:48 Wednesday 10 July 2024 05:14:00 +0000 (0:00:00.249) 0:02:05.125 ******** changed: [sut] => { "changed": true, "checksum": "585f8cbdf0ec73000f9227dcffbef71e9552ea4a", "dest": "/etc/containers/systemd/quadlet-demo-mysql.volume", "gid": 0, "group": "root", "md5sum": "5ddd03a022aeb4502d9bc8ce436b4233", "mode": "0644", "owner": "root", "secontext": "system_u:object_r:etc_t:s0", "size": 9, "src": "/root/.ansible/tmp/ansible-tmp-1720588440.498636-12001-192902405762867/source", "state": "file", "uid": 0 } TASK [fedora.linux_system_roles.podman : Ensure quadlet file content is present] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:58 Wednesday 10 July 2024 05:14:01 +0000 (0:00:00.588) 0:02:05.713 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_quadlet_file_src", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Ensure quadlet file is present] ******* task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:70 Wednesday 10 July 2024 05:14:01 +0000 (0:00:00.020) 0:02:05.734 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_copy_file is skipped", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Reload systemctl] ********************* task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:82 Wednesday 10 July 2024 05:14:01 +0000 (0:00:00.020) 0:02:05.754 ******** ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.podman : Start service] ************************ task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:110 Wednesday 10 July 2024 05:14:01 +0000 (0:00:00.797) 0:02:06.552 ******** changed: [sut] => { "changed": true, "name": "quadlet-demo-mysql-volume.service", "state": "started", "status": { "AccessSELinuxContext": "system_u:object_r:systemd_unit_file_t:s0", "ActiveEnterTimestampMonotonic": "0", "ActiveExitTimestampMonotonic": "0", "ActiveState": "inactive", "After": "-.mount system.slice systemd-journald.socket sysinit.target basic.target", "AllowIsolate": "no", "AssertResult": "no", "AssertTimestampMonotonic": "0", "Before": "shutdown.target", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "CPUAccounting": "yes", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "[not set]", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "no", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore", "CleanResult": "success", "CollectMode": "inactive", "ConditionResult": "no", "ConditionTimestampMonotonic": "0", "ConfigurationDirectoryMode": "0755", "Conflicts": "shutdown.target", "ControlGroupId": "0", "ControlPID": "0", "CoredumpFilter": "0x33", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "DefaultStartupMemoryLow": "0", "Delegate": "no", "Description": "quadlet-demo-mysql-volume.service", "DevicePolicy": "auto", "DropInPaths": "/usr/lib/systemd/system/service.d/10-timeout-abort.conf", "DynamicUser": "no", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "0", "ExecMainStartTimestampMonotonic": "0", "ExecMainStatus": "0", "ExecStart": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman volume create --ignore systemd-quadlet-demo-mysql ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStartEx": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman volume create --ignore systemd-quadlet-demo-mysql ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExitType": "main", "ExtensionImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FileDescriptorStorePreserve": "restart", "FinalKillSignal": "9", "FragmentPath": "/run/systemd/generator/quadlet-demo-mysql-volume.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOReadBytes": "18446744073709551615", "IOReadOperations": "18446744073709551615", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "IOWeight": "[not set]", "IOWriteBytes": "18446744073709551615", "IOWriteOperations": "18446744073709551615", "IPAccounting": "no", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "Id": "quadlet-demo-mysql-volume.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestampMonotonic": "0", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "control-group", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "infinity", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "8388608", "LimitMEMLOCKSoft": "8388608", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "524288", "LimitNOFILESoft": "1024", "LimitNPROC": "14756", "LimitNPROCSoft": "14756", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "14756", "LimitSIGPENDINGSoft": "14756", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "0", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "ManagedOOMSwap": "auto", "MemoryAccounting": "yes", "MemoryAvailable": "infinity", "MemoryCurrent": "[not set]", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryKSM": "no", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemoryPressureThresholdUSec": "200ms", "MemoryPressureWatch": "auto", "MemorySwapMax": "infinity", "MemoryZSwapMax": "infinity", "MountAPIVFS": "no", "MountImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAPolicy": "n/a", "Names": "quadlet-demo-mysql-volume.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMPolicy": "stop", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "OnSuccessJobMode": "fail", "Perpetual": "no", "PrivateDevices": "no", "PrivateIPC": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProcSubset": "all", "ProtectClock": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectHostname": "no", "ProtectKernelLogs": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectProc": "default", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "ReloadResult": "success", "ReloadSignal": "1", "RemainAfterExit": "yes", "RemoveIPC": "no", "Requires": "sysinit.target -.mount system.slice", "RequiresMountsFor": "/run/containers", "Restart": "no", "RestartKillSignal": "15", "RestartMaxDelayUSec": "infinity", "RestartMode": "normal", "RestartSteps": "0", "RestartUSec": "100ms", "RestartUSecNext": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RootEphemeral": "no", "RootImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "Slice": "system.slice", "StandardError": "inherit", "StandardInput": "null", "StandardOutput": "journal", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StartupMemoryHigh": "18446744073709551615", "StartupMemoryLow": "0", "StartupMemoryMax": "18446744073709551615", "StartupMemorySwapMax": "18446744073709551615", "StartupMemoryZSwapMax": "18446744073709551615", "StateChangeTimestampMonotonic": "0", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "dead", "SuccessAction": "none", "SyslogFacility": "3", "SyslogIdentifier": "quadlet-demo-mysql-volume", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "2147483646", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "[not set]", "TasksMax": "4426", "TimeoutAbortUSec": "45s", "TimeoutCleanUSec": "infinity", "TimeoutStartFailureMode": "terminate", "TimeoutStartUSec": "infinity", "TimeoutStopFailureMode": "abort", "TimeoutStopUSec": "45s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "oneshot", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "disabled", "UnitFileState": "generated", "UtmpMode": "init", "WatchdogSignal": "6", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "infinity" } } TASK [fedora.linux_system_roles.podman : Restart service] ********************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:125 Wednesday 10 July 2024 05:14:02 +0000 (0:00:00.462) 0:02:07.014 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_service_started is changed", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14 Wednesday 10 July 2024 05:14:02 +0000 (0:00:00.021) 0:02:07.035 ******** ok: [sut] => { "ansible_facts": { "__podman_quadlet_file_src": "", "__podman_quadlet_spec": {}, "__podman_quadlet_str": "[Install]\nWantedBy=default.target\n\n[Container]\nImage=quay.io/linux-system-roles/mysql:5.6\nContainerName=quadlet-demo-mysql\nVolume=quadlet-demo-mysql.volume:/var/lib/mysql\nVolume=/tmp/quadlet_demo:/var/lib/quadlet_demo:Z\nNetwork=quadlet-demo.network\nSecret=mysql-root-password-container,type=env,target=MYSQL_ROOT_PASSWORD\nHealthCmd=/usr/bin/true\nHealthOnFailure=kill\n", "__podman_quadlet_template_src": "quadlet-demo-mysql.container.j2" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25 Wednesday 10 July 2024 05:14:02 +0000 (0:00:00.058) 0:02:07.094 ******** ok: [sut] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_pull_image": true, "__podman_state": "created", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] ***** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35 Wednesday 10 July 2024 05:14:02 +0000 (0:00:00.024) 0:02:07.118 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_quadlet_str", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48 Wednesday 10 July 2024 05:14:02 +0000 (0:00:00.019) 0:02:07.138 ******** ok: [sut] => { "ansible_facts": { "__podman_quadlet_name": "quadlet-demo-mysql", "__podman_quadlet_type": "container", "__podman_rootless": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57 Wednesday 10 July 2024 05:14:02 +0000 (0:00:00.028) 0:02:07.166 ******** included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for sut TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Wednesday 10 July 2024 05:14:02 +0000 (0:00:00.062) 0:02:07.229 ******** skipping: [sut] => { "changed": false, "false_condition": "'getent_passwd' not in ansible_facts or __podman_user not in ansible_facts['getent_passwd']", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Wednesday 10 July 2024 05:14:02 +0000 (0:00:00.022) 0:02:07.252 ******** skipping: [sut] => { "changed": false, "false_condition": "not ansible_facts[\"getent_passwd\"][__podman_user]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Wednesday 10 July 2024 05:14:02 +0000 (0:00:00.021) 0:02:07.273 ******** ok: [sut] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get group information] **************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:28 Wednesday 10 July 2024 05:14:02 +0000 (0:00:00.026) 0:02:07.299 ******** ok: [sut] => { "ansible_facts": { "getent_group": { "root": [ "x", "0", "" ] } }, "changed": false } TASK [fedora.linux_system_roles.podman : Set group name] *********************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:35 Wednesday 10 July 2024 05:14:02 +0000 (0:00:00.239) 0:02:07.539 ******** ok: [sut] => { "ansible_facts": { "__podman_group_name": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:39 Wednesday 10 July 2024 05:14:02 +0000 (0:00:00.026) 0:02:07.565 ******** ok: [sut] => { "changed": false, "stat": { "atime": 1720588427.5164855, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "72ba28a3be19c467e89ab90f5d21f99b605159a7", "ctime": 1720588377.3007717, "dev": 51714, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 203603, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-pie-executable", "mode": "0755", "mtime": 1696291200.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 16032, "uid": 0, "version": "2092964490", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************ task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:50 Wednesday 10 July 2024 05:14:03 +0000 (0:00:00.234) 0:02:07.800 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check group with getsubids] *********** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:55 Wednesday 10 July 2024 05:14:03 +0000 (0:00:00.018) 0:02:07.818 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:60 Wednesday 10 July 2024 05:14:03 +0000 (0:00:00.017) 0:02:07.836 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subuid file] ********************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:74 Wednesday 10 July 2024 05:14:03 +0000 (0:00:00.017) 0:02:07.853 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subgid file] ********************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:79 Wednesday 10 July 2024 05:14:03 +0000 (0:00:00.017) 0:02:07.870 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:84 Wednesday 10 July 2024 05:14:03 +0000 (0:00:00.017) 0:02:07.888 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:94 Wednesday 10 July 2024 05:14:03 +0000 (0:00:00.017) 0:02:07.905 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] ***** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:101 Wednesday 10 July 2024 05:14:03 +0000 (0:00:00.016) 0:02:07.922 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62 Wednesday 10 July 2024 05:14:03 +0000 (0:00:00.017) 0:02:07.940 ******** ok: [sut] => { "ansible_facts": { "__podman_activate_systemd_unit": true, "__podman_images_found": [ "quay.io/linux-system-roles/mysql:5.6" ], "__podman_kube_yamls_raw": "", "__podman_service_name": "quadlet-demo-mysql.service", "__podman_systemd_scope": "system", "__podman_user_home_dir": "/root", "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73 Wednesday 10 July 2024 05:14:03 +0000 (0:00:00.033) 0:02:07.973 ******** ok: [sut] => { "ansible_facts": { "__podman_quadlet_path": "/etc/containers/systemd" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get kube yaml contents] *************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77 Wednesday 10 July 2024 05:14:03 +0000 (0:00:00.020) 0:02:07.994 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_kube_yamls_raw | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87 Wednesday 10 July 2024 05:14:03 +0000 (0:00:00.018) 0:02:08.013 ******** ok: [sut] => { "ansible_facts": { "__podman_images": [ "quay.io/linux-system-roles/mysql:5.6" ], "__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo-mysql.container", "__podman_volumes": [ "/tmp/quadlet_demo" ] }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 6] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:103 Wednesday 10 July 2024 05:14:03 +0000 (0:00:00.044) 0:02:08.057 ******** ok: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Cleanup quadlets] ********************* task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:110 Wednesday 10 July 2024 05:14:03 +0000 (0:00:00.051) 0:02:08.109 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_state == \"absent\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create and update quadlets] *********** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:114 Wednesday 10 July 2024 05:14:03 +0000 (0:00:00.017) 0:02:08.127 ******** included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml for sut TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:2 Wednesday 10 July 2024 05:14:03 +0000 (0:00:00.039) 0:02:08.167 ******** included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for sut TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Wednesday 10 July 2024 05:14:03 +0000 (0:00:00.030) 0:02:08.197 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Wednesday 10 July 2024 05:14:03 +0000 (0:00:00.017) 0:02:08.215 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Wednesday 10 July 2024 05:14:03 +0000 (0:00:00.017) 0:02:08.232 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create host directories] ************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:7 Wednesday 10 July 2024 05:14:03 +0000 (0:00:00.016) 0:02:08.249 ******** changed: [sut] => (item=/tmp/quadlet_demo) => { "ansible_loop_var": "item", "changed": true, "gid": 0, "group": "root", "item": "/tmp/quadlet_demo", "mode": "0777", "owner": "root", "path": "/tmp/quadlet_demo", "secontext": "unconfined_u:object_r:user_tmp_t:s0", "size": 40, "state": "directory", "uid": 0 } TASK [fedora.linux_system_roles.podman : Ensure container images are present] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:18 Wednesday 10 July 2024 05:14:03 +0000 (0:00:00.252) 0:02:08.502 ******** changed: [sut] => (item=None) => { "attempts": 1, "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": true } changed: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": true } TASK [fedora.linux_system_roles.podman : Ensure the quadlet directory is present] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:39 Wednesday 10 July 2024 05:14:10 +0000 (0:00:06.857) 0:02:15.359 ******** ok: [sut] => { "changed": false, "gid": 0, "group": "root", "mode": "0755", "owner": "root", "path": "/etc/containers/systemd", "secontext": "system_u:object_r:etc_t:s0", "size": 4096, "state": "directory", "uid": 0 } TASK [fedora.linux_system_roles.podman : Ensure quadlet file is copied] ******** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:48 Wednesday 10 July 2024 05:14:10 +0000 (0:00:00.245) 0:02:15.604 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_quadlet_file_src | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Ensure quadlet file content is present] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:58 Wednesday 10 July 2024 05:14:10 +0000 (0:00:00.019) 0:02:15.624 ******** changed: [sut] => { "changed": true, "checksum": "7bf9312d57274d4dabbd222b01717e2cc0cfa0ed", "dest": "/etc/containers/systemd/quadlet-demo-mysql.container", "gid": 0, "group": "root", "md5sum": "f1cd3875deb45b806389825bb4b5ad4a", "mode": "0644", "owner": "root", "secontext": "system_u:object_r:etc_t:s0", "size": 367, "src": "/root/.ansible/tmp/ansible-tmp-1720588450.9974556-12056-158306006965184/source", "state": "file", "uid": 0 } TASK [fedora.linux_system_roles.podman : Ensure quadlet file is present] ******* task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:70 Wednesday 10 July 2024 05:14:11 +0000 (0:00:00.584) 0:02:16.208 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_copy_content is skipped", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Reload systemctl] ********************* task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:82 Wednesday 10 July 2024 05:14:11 +0000 (0:00:00.018) 0:02:16.227 ******** ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.podman : Start service] ************************ task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:110 Wednesday 10 July 2024 05:14:12 +0000 (0:00:00.812) 0:02:17.039 ******** changed: [sut] => { "changed": true, "name": "quadlet-demo-mysql.service", "state": "started", "status": { "AccessSELinuxContext": "system_u:object_r:systemd_unit_file_t:s0", "ActiveEnterTimestampMonotonic": "0", "ActiveExitTimestampMonotonic": "0", "ActiveState": "inactive", "After": "sysinit.target basic.target quadlet-demo-network.service tmp.mount quadlet-demo-mysql-volume.service -.mount system.slice systemd-journald.socket", "AllowIsolate": "no", "AssertResult": "no", "AssertTimestampMonotonic": "0", "Before": "multi-user.target shutdown.target", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "CPUAccounting": "yes", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "[not set]", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "no", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore", "CleanResult": "success", "CollectMode": "inactive", "ConditionResult": "no", "ConditionTimestampMonotonic": "0", "ConfigurationDirectoryMode": "0755", "Conflicts": "shutdown.target", "ControlGroupId": "0", "ControlPID": "0", "CoredumpFilter": "0x33", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "DefaultStartupMemoryLow": "0", "Delegate": "yes", "DelegateControllers": "cpu cpuset io memory pids", "Description": "quadlet-demo-mysql.service", "DevicePolicy": "auto", "DropInPaths": "/usr/lib/systemd/system/service.d/10-timeout-abort.conf", "DynamicUser": "no", "Environment": "PODMAN_SYSTEMD_UNIT=quadlet-demo-mysql.service", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "0", "ExecMainStartTimestampMonotonic": "0", "ExecMainStatus": "0", "ExecStart": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman run --name=quadlet-demo-mysql --cidfile=/run/quadlet-demo-mysql.cid --replace --rm --cgroups=split --network=systemd-quadlet-demo --sdnotify=conmon -d -v systemd-quadlet-demo-mysql:/var/lib/mysql -v /tmp/quadlet_demo:/var/lib/quadlet_demo:Z --secret mysql-root-password-container,type=env,target=MYSQL_ROOT_PASSWORD --health-cmd /usr/bin/true --health-on-failure kill quay.io/linux-system-roles/mysql:5.6 ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStartEx": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman run --name=quadlet-demo-mysql --cidfile=/run/quadlet-demo-mysql.cid --replace --rm --cgroups=split --network=systemd-quadlet-demo --sdnotify=conmon -d -v systemd-quadlet-demo-mysql:/var/lib/mysql -v /tmp/quadlet_demo:/var/lib/quadlet_demo:Z --secret mysql-root-password-container,type=env,target=MYSQL_ROOT_PASSWORD --health-cmd /usr/bin/true --health-on-failure kill quay.io/linux-system-roles/mysql:5.6 ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStop": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman rm -v -f -i --cidfile=/run/quadlet-demo-mysql.cid ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStopEx": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman rm -v -f -i --cidfile=/run/quadlet-demo-mysql.cid ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStopPost": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman rm -v -f -i --cidfile=/run/quadlet-demo-mysql.cid ; ignore_errors=yes ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStopPostEx": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman rm -v -f -i --cidfile=/run/quadlet-demo-mysql.cid ; flags=ignore-failure ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExitType": "main", "ExtensionImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FileDescriptorStorePreserve": "restart", "FinalKillSignal": "9", "FragmentPath": "/run/systemd/generator/quadlet-demo-mysql.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOReadBytes": "18446744073709551615", "IOReadOperations": "18446744073709551615", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "IOWeight": "[not set]", "IOWriteBytes": "18446744073709551615", "IOWriteOperations": "18446744073709551615", "IPAccounting": "no", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "Id": "quadlet-demo-mysql.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestampMonotonic": "0", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "mixed", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "infinity", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "8388608", "LimitMEMLOCKSoft": "8388608", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "524288", "LimitNOFILESoft": "1024", "LimitNPROC": "14756", "LimitNPROCSoft": "14756", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "14756", "LimitSIGPENDINGSoft": "14756", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "0", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "ManagedOOMSwap": "auto", "MemoryAccounting": "yes", "MemoryAvailable": "infinity", "MemoryCurrent": "[not set]", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryKSM": "no", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemoryPressureThresholdUSec": "200ms", "MemoryPressureWatch": "auto", "MemorySwapMax": "infinity", "MemoryZSwapMax": "infinity", "MountAPIVFS": "no", "MountImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAPolicy": "n/a", "Names": "quadlet-demo-mysql.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "all", "OOMPolicy": "continue", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "OnSuccessJobMode": "fail", "Perpetual": "no", "PrivateDevices": "no", "PrivateIPC": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProcSubset": "all", "ProtectClock": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectHostname": "no", "ProtectKernelLogs": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectProc": "default", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "ReloadResult": "success", "ReloadSignal": "1", "RemainAfterExit": "no", "RemoveIPC": "no", "Requires": "system.slice -.mount tmp.mount sysinit.target quadlet-demo-network.service quadlet-demo-mysql-volume.service", "RequiresMountsFor": "/tmp/quadlet_demo /run/containers", "Restart": "no", "RestartKillSignal": "15", "RestartMaxDelayUSec": "infinity", "RestartMode": "normal", "RestartSteps": "0", "RestartUSec": "100ms", "RestartUSecNext": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RootEphemeral": "no", "RootImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "Slice": "system.slice", "SourcePath": "/etc/containers/systemd/quadlet-demo-mysql.container", "StandardError": "inherit", "StandardInput": "null", "StandardOutput": "journal", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StartupMemoryHigh": "18446744073709551615", "StartupMemoryLow": "0", "StartupMemoryMax": "18446744073709551615", "StartupMemorySwapMax": "18446744073709551615", "StartupMemoryZSwapMax": "18446744073709551615", "StateChangeTimestampMonotonic": "0", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "dead", "SuccessAction": "none", "SyslogFacility": "3", "SyslogIdentifier": "quadlet-demo-mysql", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "2147483646", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "[not set]", "TasksMax": "4426", "TimeoutAbortUSec": "45s", "TimeoutCleanUSec": "infinity", "TimeoutStartFailureMode": "terminate", "TimeoutStartUSec": "45s", "TimeoutStopFailureMode": "abort", "TimeoutStopUSec": "45s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "notify", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "disabled", "UnitFileState": "generated", "UtmpMode": "init", "WantedBy": "multi-user.target", "WatchdogSignal": "6", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "infinity" } } TASK [fedora.linux_system_roles.podman : Restart service] ********************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:125 Wednesday 10 July 2024 05:14:13 +0000 (0:00:00.907) 0:02:17.946 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_service_started is changed", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14 Wednesday 10 July 2024 05:14:13 +0000 (0:00:00.019) 0:02:17.965 ******** ok: [sut] => { "ansible_facts": { "__podman_quadlet_file_src": "envoy-proxy-configmap.yml", "__podman_quadlet_spec": {}, "__podman_quadlet_str": "---\napiVersion: v1\nkind: ConfigMap\nmetadata:\n name: envoy-proxy-config\ndata:\n envoy.yaml: |\n admin:\n address:\n socket_address:\n address: 0.0.0.0\n port_value: 9901\n\n static_resources:\n listeners:\n - name: listener_0\n address:\n socket_address:\n address: 0.0.0.0\n port_value: 8080\n filter_chains:\n - filters:\n - name: envoy.filters.network.http_connection_manager\n typed_config:\n \"@type\": type.googleapis.com/envoy.extensions.filters.network.http_connection_manager.v3.HttpConnectionManager\n stat_prefix: ingress_http\n codec_type: AUTO\n route_config:\n name: local_route\n virtual_hosts:\n - name: local_service\n domains: [\"*\"]\n routes:\n - match:\n prefix: \"/\"\n route:\n cluster: backend\n http_filters:\n - name: envoy.filters.http.router\n typed_config:\n \"@type\": type.googleapis.com/envoy.extensions.filters.http.router.v3.Router\n transport_socket:\n name: envoy.transport_sockets.tls\n typed_config:\n \"@type\": type.googleapis.com/envoy.extensions.transport_sockets.tls.v3.DownstreamTlsContext\n common_tls_context:\n tls_certificates:\n - certificate_chain:\n filename: /etc/envoy-certificates/certificate.pem\n private_key:\n filename: /etc/envoy-certificates/certificate.key\n clusters:\n - name: backend\n connect_timeout: 5s\n type: STATIC\n dns_refresh_rate: 1800s\n lb_policy: ROUND_ROBIN\n load_assignment:\n cluster_name: backend\n endpoints:\n - lb_endpoints:\n - endpoint:\n address:\n socket_address:\n address: 127.0.0.1\n port_value: 80", "__podman_quadlet_template_src": "" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25 Wednesday 10 July 2024 05:14:13 +0000 (0:00:00.025) 0:02:17.991 ******** ok: [sut] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_pull_image": true, "__podman_state": "created", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] ***** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35 Wednesday 10 July 2024 05:14:13 +0000 (0:00:00.052) 0:02:18.044 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_quadlet_file_src", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48 Wednesday 10 July 2024 05:14:13 +0000 (0:00:00.019) 0:02:18.064 ******** ok: [sut] => { "ansible_facts": { "__podman_quadlet_name": "envoy-proxy-configmap", "__podman_quadlet_type": "yml", "__podman_rootless": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57 Wednesday 10 July 2024 05:14:13 +0000 (0:00:00.029) 0:02:18.094 ******** included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for sut TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Wednesday 10 July 2024 05:14:13 +0000 (0:00:00.035) 0:02:18.129 ******** skipping: [sut] => { "changed": false, "false_condition": "'getent_passwd' not in ansible_facts or __podman_user not in ansible_facts['getent_passwd']", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Wednesday 10 July 2024 05:14:13 +0000 (0:00:00.020) 0:02:18.150 ******** skipping: [sut] => { "changed": false, "false_condition": "not ansible_facts[\"getent_passwd\"][__podman_user]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Wednesday 10 July 2024 05:14:13 +0000 (0:00:00.020) 0:02:18.170 ******** ok: [sut] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get group information] **************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:28 Wednesday 10 July 2024 05:14:13 +0000 (0:00:00.027) 0:02:18.197 ******** ok: [sut] => { "ansible_facts": { "getent_group": { "root": [ "x", "0", "" ] } }, "changed": false } TASK [fedora.linux_system_roles.podman : Set group name] *********************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:35 Wednesday 10 July 2024 05:14:13 +0000 (0:00:00.267) 0:02:18.465 ******** ok: [sut] => { "ansible_facts": { "__podman_group_name": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:39 Wednesday 10 July 2024 05:14:13 +0000 (0:00:00.025) 0:02:18.490 ******** ok: [sut] => { "changed": false, "stat": { "atime": 1720588427.5164855, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "72ba28a3be19c467e89ab90f5d21f99b605159a7", "ctime": 1720588377.3007717, "dev": 51714, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 203603, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-pie-executable", "mode": "0755", "mtime": 1696291200.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 16032, "uid": 0, "version": "2092964490", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************ task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:50 Wednesday 10 July 2024 05:14:14 +0000 (0:00:00.235) 0:02:18.725 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check group with getsubids] *********** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:55 Wednesday 10 July 2024 05:14:14 +0000 (0:00:00.017) 0:02:18.743 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:60 Wednesday 10 July 2024 05:14:14 +0000 (0:00:00.017) 0:02:18.761 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subuid file] ********************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:74 Wednesday 10 July 2024 05:14:14 +0000 (0:00:00.017) 0:02:18.778 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subgid file] ********************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:79 Wednesday 10 July 2024 05:14:14 +0000 (0:00:00.017) 0:02:18.795 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:84 Wednesday 10 July 2024 05:14:14 +0000 (0:00:00.017) 0:02:18.813 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:94 Wednesday 10 July 2024 05:14:14 +0000 (0:00:00.016) 0:02:18.829 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] ***** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:101 Wednesday 10 July 2024 05:14:14 +0000 (0:00:00.016) 0:02:18.846 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62 Wednesday 10 July 2024 05:14:14 +0000 (0:00:00.016) 0:02:18.863 ******** ok: [sut] => { "ansible_facts": { "__podman_activate_systemd_unit": true, "__podman_images_found": [], "__podman_kube_yamls_raw": "", "__podman_service_name": "", "__podman_systemd_scope": "system", "__podman_user_home_dir": "/root", "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73 Wednesday 10 July 2024 05:14:14 +0000 (0:00:00.062) 0:02:18.925 ******** ok: [sut] => { "ansible_facts": { "__podman_quadlet_path": "/etc/containers/systemd" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get kube yaml contents] *************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77 Wednesday 10 July 2024 05:14:14 +0000 (0:00:00.020) 0:02:18.946 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_kube_yamls_raw | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87 Wednesday 10 July 2024 05:14:14 +0000 (0:00:00.019) 0:02:18.965 ******** ok: [sut] => { "ansible_facts": { "__podman_images": [], "__podman_quadlet_file": "/etc/containers/systemd/envoy-proxy-configmap.yml", "__podman_volumes": [] }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 6] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:103 Wednesday 10 July 2024 05:14:14 +0000 (0:00:00.045) 0:02:19.011 ******** ok: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Cleanup quadlets] ********************* task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:110 Wednesday 10 July 2024 05:14:14 +0000 (0:00:00.022) 0:02:19.034 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_state == \"absent\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create and update quadlets] *********** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:114 Wednesday 10 July 2024 05:14:14 +0000 (0:00:00.017) 0:02:19.052 ******** included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml for sut TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:2 Wednesday 10 July 2024 05:14:14 +0000 (0:00:00.038) 0:02:19.090 ******** included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for sut TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Wednesday 10 July 2024 05:14:14 +0000 (0:00:00.030) 0:02:19.121 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Wednesday 10 July 2024 05:14:14 +0000 (0:00:00.017) 0:02:19.139 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Wednesday 10 July 2024 05:14:14 +0000 (0:00:00.016) 0:02:19.156 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create host directories] ************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:7 Wednesday 10 July 2024 05:14:14 +0000 (0:00:00.017) 0:02:19.173 ******** skipping: [sut] => { "changed": false, "skipped_reason": "No items in the list" } TASK [fedora.linux_system_roles.podman : Ensure container images are present] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:18 Wednesday 10 July 2024 05:14:14 +0000 (0:00:00.015) 0:02:19.188 ******** skipping: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Ensure the quadlet directory is present] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:39 Wednesday 10 July 2024 05:14:14 +0000 (0:00:00.016) 0:02:19.205 ******** ok: [sut] => { "changed": false, "gid": 0, "group": "root", "mode": "0755", "owner": "root", "path": "/etc/containers/systemd", "secontext": "system_u:object_r:etc_t:s0", "size": 4096, "state": "directory", "uid": 0 } TASK [fedora.linux_system_roles.podman : Ensure quadlet file is copied] ******** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:48 Wednesday 10 July 2024 05:14:14 +0000 (0:00:00.256) 0:02:19.461 ******** changed: [sut] => { "changed": true, "checksum": "d681c7d56f912150d041873e880818b22a90c188", "dest": "/etc/containers/systemd/envoy-proxy-configmap.yml", "gid": 0, "group": "root", "md5sum": "aec75d972c231aac004e1338934544cf", "mode": "0644", "owner": "root", "secontext": "system_u:object_r:etc_t:s0", "size": 2102, "src": "/root/.ansible/tmp/ansible-tmp-1720588454.8339243-12107-16527747628217/source", "state": "file", "uid": 0 } TASK [fedora.linux_system_roles.podman : Ensure quadlet file content is present] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:58 Wednesday 10 July 2024 05:14:15 +0000 (0:00:00.578) 0:02:20.040 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_quadlet_file_src", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Ensure quadlet file is present] ******* task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:70 Wednesday 10 July 2024 05:14:15 +0000 (0:00:00.020) 0:02:20.060 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_copy_file is skipped", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Reload systemctl] ********************* task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:82 Wednesday 10 July 2024 05:14:15 +0000 (0:00:00.017) 0:02:20.078 ******** ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.podman : Start service] ************************ task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:110 Wednesday 10 July 2024 05:14:16 +0000 (0:00:00.910) 0:02:20.988 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_service_name | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Restart service] ********************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:125 Wednesday 10 July 2024 05:14:16 +0000 (0:00:00.022) 0:02:21.010 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_service_name | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14 Wednesday 10 July 2024 05:14:16 +0000 (0:00:00.020) 0:02:21.031 ******** ok: [sut] => { "ansible_facts": { "__podman_quadlet_file_src": "", "__podman_quadlet_spec": {}, "__podman_quadlet_str": "---\napiVersion: v1\nkind: PersistentVolumeClaim\nmetadata:\n name: wp-pv-claim\n labels:\n app: wordpress\nspec:\n accessModes:\n - ReadWriteOnce\n resources:\n requests:\n storage: 20Gi\n---\napiVersion: v1\nkind: Pod\nmetadata:\n name: quadlet-demo\nspec:\n containers:\n - name: wordpress\n image: quay.io/linux-system-roles/wordpress:4.8-apache\n env:\n - name: WORDPRESS_DB_HOST\n value: quadlet-demo-mysql\n - name: WORDPRESS_DB_PASSWORD\n valueFrom:\n secretKeyRef:\n name: mysql-root-password-kube\n key: password\n volumeMounts:\n - name: wordpress-persistent-storage\n mountPath: /var/www/html\n resources:\n requests:\n memory: \"64Mi\"\n cpu: \"250m\"\n limits:\n memory: \"128Mi\"\n cpu: \"500m\"\n - name: envoy\n image: quay.io/linux-system-roles/envoyproxy:v1.25.0\n volumeMounts:\n - name: config-volume\n mountPath: /etc/envoy\n - name: certificates\n mountPath: /etc/envoy-certificates\n env:\n - name: ENVOY_UID\n value: \"0\"\n resources:\n requests:\n memory: \"64Mi\"\n cpu: \"250m\"\n limits:\n memory: \"128Mi\"\n cpu: \"500m\"\n volumes:\n - name: config-volume\n configMap:\n name: envoy-proxy-config\n - name: certificates\n secret:\n secretName: envoy-certificates\n - name: wordpress-persistent-storage\n persistentVolumeClaim:\n claimName: wp-pv-claim\n - name: www # not used - for testing hostpath\n hostPath:\n path: /tmp/httpd3\n - name: create # not used - for testing hostpath\n hostPath:\n path: /tmp/httpd3-create\n", "__podman_quadlet_template_src": "quadlet-demo.yml.j2" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25 Wednesday 10 July 2024 05:14:16 +0000 (0:00:00.057) 0:02:21.088 ******** ok: [sut] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_pull_image": true, "__podman_state": "created", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] ***** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35 Wednesday 10 July 2024 05:14:16 +0000 (0:00:00.024) 0:02:21.112 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_quadlet_str", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48 Wednesday 10 July 2024 05:14:16 +0000 (0:00:00.019) 0:02:21.132 ******** ok: [sut] => { "ansible_facts": { "__podman_quadlet_name": "quadlet-demo", "__podman_quadlet_type": "yml", "__podman_rootless": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57 Wednesday 10 July 2024 05:14:16 +0000 (0:00:00.028) 0:02:21.160 ******** included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for sut TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Wednesday 10 July 2024 05:14:16 +0000 (0:00:00.034) 0:02:21.195 ******** skipping: [sut] => { "changed": false, "false_condition": "'getent_passwd' not in ansible_facts or __podman_user not in ansible_facts['getent_passwd']", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Wednesday 10 July 2024 05:14:16 +0000 (0:00:00.021) 0:02:21.216 ******** skipping: [sut] => { "changed": false, "false_condition": "not ansible_facts[\"getent_passwd\"][__podman_user]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Wednesday 10 July 2024 05:14:16 +0000 (0:00:00.021) 0:02:21.237 ******** ok: [sut] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get group information] **************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:28 Wednesday 10 July 2024 05:14:16 +0000 (0:00:00.026) 0:02:21.264 ******** ok: [sut] => { "ansible_facts": { "getent_group": { "root": [ "x", "0", "" ] } }, "changed": false } TASK [fedora.linux_system_roles.podman : Set group name] *********************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:35 Wednesday 10 July 2024 05:14:16 +0000 (0:00:00.391) 0:02:21.656 ******** ok: [sut] => { "ansible_facts": { "__podman_group_name": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:39 Wednesday 10 July 2024 05:14:17 +0000 (0:00:00.026) 0:02:21.682 ******** ok: [sut] => { "changed": false, "stat": { "atime": 1720588427.5164855, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "72ba28a3be19c467e89ab90f5d21f99b605159a7", "ctime": 1720588377.3007717, "dev": 51714, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 203603, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-pie-executable", "mode": "0755", "mtime": 1696291200.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 16032, "uid": 0, "version": "2092964490", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************ task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:50 Wednesday 10 July 2024 05:14:17 +0000 (0:00:00.290) 0:02:21.973 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check group with getsubids] *********** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:55 Wednesday 10 July 2024 05:14:17 +0000 (0:00:00.018) 0:02:21.991 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:60 Wednesday 10 July 2024 05:14:17 +0000 (0:00:00.017) 0:02:22.008 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subuid file] ********************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:74 Wednesday 10 July 2024 05:14:17 +0000 (0:00:00.017) 0:02:22.026 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subgid file] ********************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:79 Wednesday 10 July 2024 05:14:17 +0000 (0:00:00.017) 0:02:22.044 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:84 Wednesday 10 July 2024 05:14:17 +0000 (0:00:00.046) 0:02:22.090 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:94 Wednesday 10 July 2024 05:14:17 +0000 (0:00:00.018) 0:02:22.109 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] ***** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:101 Wednesday 10 July 2024 05:14:17 +0000 (0:00:00.017) 0:02:22.127 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62 Wednesday 10 July 2024 05:14:17 +0000 (0:00:00.018) 0:02:22.145 ******** ok: [sut] => { "ansible_facts": { "__podman_activate_systemd_unit": true, "__podman_images_found": [], "__podman_kube_yamls_raw": "", "__podman_service_name": "", "__podman_systemd_scope": "system", "__podman_user_home_dir": "/root", "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73 Wednesday 10 July 2024 05:14:17 +0000 (0:00:00.033) 0:02:22.178 ******** ok: [sut] => { "ansible_facts": { "__podman_quadlet_path": "/etc/containers/systemd" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get kube yaml contents] *************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77 Wednesday 10 July 2024 05:14:17 +0000 (0:00:00.019) 0:02:22.198 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_kube_yamls_raw | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87 Wednesday 10 July 2024 05:14:17 +0000 (0:00:00.018) 0:02:22.217 ******** ok: [sut] => { "ansible_facts": { "__podman_images": [], "__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo.yml", "__podman_volumes": [] }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 6] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:103 Wednesday 10 July 2024 05:14:17 +0000 (0:00:00.044) 0:02:22.262 ******** ok: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Cleanup quadlets] ********************* task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:110 Wednesday 10 July 2024 05:14:17 +0000 (0:00:00.023) 0:02:22.285 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_state == \"absent\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create and update quadlets] *********** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:114 Wednesday 10 July 2024 05:14:17 +0000 (0:00:00.016) 0:02:22.302 ******** included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml for sut TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:2 Wednesday 10 July 2024 05:14:17 +0000 (0:00:00.039) 0:02:22.342 ******** included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for sut TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Wednesday 10 July 2024 05:14:17 +0000 (0:00:00.030) 0:02:22.372 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Wednesday 10 July 2024 05:14:17 +0000 (0:00:00.017) 0:02:22.389 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Wednesday 10 July 2024 05:14:17 +0000 (0:00:00.017) 0:02:22.406 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create host directories] ************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:7 Wednesday 10 July 2024 05:14:17 +0000 (0:00:00.017) 0:02:22.424 ******** skipping: [sut] => { "changed": false, "skipped_reason": "No items in the list" } TASK [fedora.linux_system_roles.podman : Ensure container images are present] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:18 Wednesday 10 July 2024 05:14:17 +0000 (0:00:00.014) 0:02:22.439 ******** skipping: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Ensure the quadlet directory is present] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:39 Wednesday 10 July 2024 05:14:17 +0000 (0:00:00.046) 0:02:22.485 ******** ok: [sut] => { "changed": false, "gid": 0, "group": "root", "mode": "0755", "owner": "root", "path": "/etc/containers/systemd", "secontext": "system_u:object_r:etc_t:s0", "size": 4096, "state": "directory", "uid": 0 } TASK [fedora.linux_system_roles.podman : Ensure quadlet file is copied] ******** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:48 Wednesday 10 July 2024 05:14:18 +0000 (0:00:00.343) 0:02:22.828 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_quadlet_file_src | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Ensure quadlet file content is present] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:58 Wednesday 10 July 2024 05:14:18 +0000 (0:00:00.019) 0:02:22.848 ******** changed: [sut] => { "changed": true, "checksum": "998dccde0483b1654327a46ddd89cbaa47650370", "dest": "/etc/containers/systemd/quadlet-demo.yml", "gid": 0, "group": "root", "md5sum": "fd890594adfc24339cb9cdc5e7b19a66", "mode": "0644", "owner": "root", "secontext": "system_u:object_r:etc_t:s0", "size": 1605, "src": "/root/.ansible/tmp/ansible-tmp-1720588458.2299404-12159-197295334066146/source", "state": "file", "uid": 0 } TASK [fedora.linux_system_roles.podman : Ensure quadlet file is present] ******* task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:70 Wednesday 10 July 2024 05:14:19 +0000 (0:00:00.922) 0:02:23.771 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_copy_content is skipped", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Reload systemctl] ********************* task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:82 Wednesday 10 July 2024 05:14:19 +0000 (0:00:00.018) 0:02:23.789 ******** ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.podman : Start service] ************************ task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:110 Wednesday 10 July 2024 05:14:20 +0000 (0:00:01.279) 0:02:25.068 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_service_name | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Restart service] ********************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:125 Wednesday 10 July 2024 05:14:20 +0000 (0:00:00.021) 0:02:25.089 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_service_name | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14 Wednesday 10 July 2024 05:14:20 +0000 (0:00:00.021) 0:02:25.111 ******** ok: [sut] => { "ansible_facts": { "__podman_quadlet_file_src": "quadlet-demo.kube", "__podman_quadlet_spec": {}, "__podman_quadlet_str": "[Install]\nWantedBy=default.target\n\n[Unit]\nRequires=quadlet-demo-mysql.service\nAfter=quadlet-demo-mysql.service\n\n[Kube]\n# Point to the yaml file in the same directory\nYaml=quadlet-demo.yml\n# Use the quadlet-demo network\nNetwork=quadlet-demo.network\n# Publish the envoy proxy data port\nPublishPort=8000:8080\n# Publish the envoy proxy admin port\nPublishPort=9000:9901\n# Use the envoy proxy config map in the same directory\nConfigMap=envoy-proxy-configmap.yml", "__podman_quadlet_template_src": "" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25 Wednesday 10 July 2024 05:14:20 +0000 (0:00:00.026) 0:02:25.137 ******** ok: [sut] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_pull_image": true, "__podman_state": "created", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] ***** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35 Wednesday 10 July 2024 05:14:20 +0000 (0:00:00.024) 0:02:25.162 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_quadlet_file_src", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48 Wednesday 10 July 2024 05:14:20 +0000 (0:00:00.018) 0:02:25.181 ******** ok: [sut] => { "ansible_facts": { "__podman_quadlet_name": "quadlet-demo", "__podman_quadlet_type": "kube", "__podman_rootless": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57 Wednesday 10 July 2024 05:14:20 +0000 (0:00:00.028) 0:02:25.209 ******** included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for sut TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Wednesday 10 July 2024 05:14:20 +0000 (0:00:00.035) 0:02:25.244 ******** skipping: [sut] => { "changed": false, "false_condition": "'getent_passwd' not in ansible_facts or __podman_user not in ansible_facts['getent_passwd']", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Wednesday 10 July 2024 05:14:20 +0000 (0:00:00.020) 0:02:25.265 ******** skipping: [sut] => { "changed": false, "false_condition": "not ansible_facts[\"getent_passwd\"][__podman_user]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Wednesday 10 July 2024 05:14:20 +0000 (0:00:00.020) 0:02:25.285 ******** ok: [sut] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get group information] **************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:28 Wednesday 10 July 2024 05:14:20 +0000 (0:00:00.026) 0:02:25.312 ******** ok: [sut] => { "ansible_facts": { "getent_group": { "root": [ "x", "0", "" ] } }, "changed": false } TASK [fedora.linux_system_roles.podman : Set group name] *********************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:35 Wednesday 10 July 2024 05:14:21 +0000 (0:00:00.362) 0:02:25.674 ******** ok: [sut] => { "ansible_facts": { "__podman_group_name": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:39 Wednesday 10 July 2024 05:14:21 +0000 (0:00:00.026) 0:02:25.700 ******** ok: [sut] => { "changed": false, "stat": { "atime": 1720588427.5164855, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "72ba28a3be19c467e89ab90f5d21f99b605159a7", "ctime": 1720588377.3007717, "dev": 51714, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 203603, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-pie-executable", "mode": "0755", "mtime": 1696291200.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 16032, "uid": 0, "version": "2092964490", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************ task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:50 Wednesday 10 July 2024 05:14:21 +0000 (0:00:00.399) 0:02:26.100 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check group with getsubids] *********** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:55 Wednesday 10 July 2024 05:14:21 +0000 (0:00:00.018) 0:02:26.118 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:60 Wednesday 10 July 2024 05:14:21 +0000 (0:00:00.018) 0:02:26.136 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subuid file] ********************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:74 Wednesday 10 July 2024 05:14:21 +0000 (0:00:00.017) 0:02:26.153 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subgid file] ********************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:79 Wednesday 10 July 2024 05:14:21 +0000 (0:00:00.017) 0:02:26.171 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:84 Wednesday 10 July 2024 05:14:21 +0000 (0:00:00.017) 0:02:26.188 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:94 Wednesday 10 July 2024 05:14:21 +0000 (0:00:00.017) 0:02:26.205 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] ***** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:101 Wednesday 10 July 2024 05:14:21 +0000 (0:00:00.017) 0:02:26.223 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62 Wednesday 10 July 2024 05:14:21 +0000 (0:00:00.016) 0:02:26.240 ******** ok: [sut] => { "ansible_facts": { "__podman_activate_systemd_unit": true, "__podman_images_found": [], "__podman_kube_yamls_raw": [ "quadlet-demo.yml" ], "__podman_service_name": "quadlet-demo.service", "__podman_systemd_scope": "system", "__podman_user_home_dir": "/root", "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73 Wednesday 10 July 2024 05:14:21 +0000 (0:00:00.033) 0:02:26.274 ******** ok: [sut] => { "ansible_facts": { "__podman_quadlet_path": "/etc/containers/systemd" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get kube yaml contents] *************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77 Wednesday 10 July 2024 05:14:21 +0000 (0:00:00.019) 0:02:26.293 ******** ok: [sut] => { "changed": false, "content": "LS0tCmFwaVZlcnNpb246IHYxCmtpbmQ6IFBlcnNpc3RlbnRWb2x1bWVDbGFpbQptZXRhZGF0YToKICBuYW1lOiB3cC1wdi1jbGFpbQogIGxhYmVsczoKICAgIGFwcDogd29yZHByZXNzCnNwZWM6CiAgYWNjZXNzTW9kZXM6CiAgLSBSZWFkV3JpdGVPbmNlCiAgcmVzb3VyY2VzOgogICAgcmVxdWVzdHM6CiAgICAgIHN0b3JhZ2U6IDIwR2kKLS0tCmFwaVZlcnNpb246IHYxCmtpbmQ6IFBvZAptZXRhZGF0YToKICBuYW1lOiBxdWFkbGV0LWRlbW8Kc3BlYzoKICBjb250YWluZXJzOgogIC0gbmFtZTogd29yZHByZXNzCiAgICBpbWFnZTogcXVheS5pby9saW51eC1zeXN0ZW0tcm9sZXMvd29yZHByZXNzOjQuOC1hcGFjaGUKICAgIGVudjoKICAgIC0gbmFtZTogV09SRFBSRVNTX0RCX0hPU1QKICAgICAgdmFsdWU6IHF1YWRsZXQtZGVtby1teXNxbAogICAgLSBuYW1lOiBXT1JEUFJFU1NfREJfUEFTU1dPUkQKICAgICAgdmFsdWVGcm9tOgogICAgICAgIHNlY3JldEtleVJlZjoKICAgICAgICAgIG5hbWU6IG15c3FsLXJvb3QtcGFzc3dvcmQta3ViZQogICAgICAgICAga2V5OiBwYXNzd29yZAogICAgdm9sdW1lTW91bnRzOgogICAgLSBuYW1lOiB3b3JkcHJlc3MtcGVyc2lzdGVudC1zdG9yYWdlCiAgICAgIG1vdW50UGF0aDogL3Zhci93d3cvaHRtbAogICAgcmVzb3VyY2VzOgogICAgICByZXF1ZXN0czoKICAgICAgICBtZW1vcnk6ICI2NE1pIgogICAgICAgIGNwdTogIjI1MG0iCiAgICAgIGxpbWl0czoKICAgICAgICBtZW1vcnk6ICIxMjhNaSIKICAgICAgICBjcHU6ICI1MDBtIgogIC0gbmFtZTogZW52b3kKICAgIGltYWdlOiBxdWF5LmlvL2xpbnV4LXN5c3RlbS1yb2xlcy9lbnZveXByb3h5OnYxLjI1LjAKICAgIHZvbHVtZU1vdW50czoKICAgIC0gbmFtZTogY29uZmlnLXZvbHVtZQogICAgICBtb3VudFBhdGg6IC9ldGMvZW52b3kKICAgIC0gbmFtZTogY2VydGlmaWNhdGVzCiAgICAgIG1vdW50UGF0aDogL2V0Yy9lbnZveS1jZXJ0aWZpY2F0ZXMKICAgIGVudjoKICAgIC0gbmFtZTogRU5WT1lfVUlECiAgICAgIHZhbHVlOiAiMCIKICAgIHJlc291cmNlczoKICAgICAgcmVxdWVzdHM6CiAgICAgICAgbWVtb3J5OiAiNjRNaSIKICAgICAgICBjcHU6ICIyNTBtIgogICAgICBsaW1pdHM6CiAgICAgICAgbWVtb3J5OiAiMTI4TWkiCiAgICAgICAgY3B1OiAiNTAwbSIKICB2b2x1bWVzOgogIC0gbmFtZTogY29uZmlnLXZvbHVtZQogICAgY29uZmlnTWFwOgogICAgICBuYW1lOiBlbnZveS1wcm94eS1jb25maWcKICAtIG5hbWU6IGNlcnRpZmljYXRlcwogICAgc2VjcmV0OgogICAgICBzZWNyZXROYW1lOiBlbnZveS1jZXJ0aWZpY2F0ZXMKICAtIG5hbWU6IHdvcmRwcmVzcy1wZXJzaXN0ZW50LXN0b3JhZ2UKICAgIHBlcnNpc3RlbnRWb2x1bWVDbGFpbToKICAgICAgY2xhaW1OYW1lOiB3cC1wdi1jbGFpbQogIC0gbmFtZTogd3d3ICAjIG5vdCB1c2VkIC0gZm9yIHRlc3RpbmcgaG9zdHBhdGgKICAgIGhvc3RQYXRoOgogICAgICBwYXRoOiAvdG1wL2h0dHBkMwogIC0gbmFtZTogY3JlYXRlICAjIG5vdCB1c2VkIC0gZm9yIHRlc3RpbmcgaG9zdHBhdGgKICAgIGhvc3RQYXRoOgogICAgICBwYXRoOiAvdG1wL2h0dHBkMy1jcmVhdGUK", "encoding": "base64", "source": "/etc/containers/systemd/quadlet-demo.yml" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87 Wednesday 10 July 2024 05:14:21 +0000 (0:00:00.230) 0:02:26.524 ******** ok: [sut] => { "ansible_facts": { "__podman_images": [ "quay.io/linux-system-roles/wordpress:4.8-apache", "quay.io/linux-system-roles/envoyproxy:v1.25.0" ], "__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo.kube", "__podman_volumes": [ "/tmp/httpd3", "/tmp/httpd3-create" ] }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 6] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:103 Wednesday 10 July 2024 05:14:21 +0000 (0:00:00.054) 0:02:26.579 ******** ok: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Cleanup quadlets] ********************* task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:110 Wednesday 10 July 2024 05:14:21 +0000 (0:00:00.022) 0:02:26.601 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_state == \"absent\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create and update quadlets] *********** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:114 Wednesday 10 July 2024 05:14:21 +0000 (0:00:00.018) 0:02:26.620 ******** included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml for sut TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:2 Wednesday 10 July 2024 05:14:21 +0000 (0:00:00.039) 0:02:26.659 ******** included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for sut TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Wednesday 10 July 2024 05:14:22 +0000 (0:00:00.030) 0:02:26.689 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Wednesday 10 July 2024 05:14:22 +0000 (0:00:00.047) 0:02:26.736 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Wednesday 10 July 2024 05:14:22 +0000 (0:00:00.018) 0:02:26.755 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create host directories] ************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:7 Wednesday 10 July 2024 05:14:22 +0000 (0:00:00.017) 0:02:26.772 ******** changed: [sut] => (item=/tmp/httpd3) => { "ansible_loop_var": "item", "changed": true, "gid": 0, "group": "root", "item": "/tmp/httpd3", "mode": "0755", "owner": "root", "path": "/tmp/httpd3", "secontext": "unconfined_u:object_r:user_tmp_t:s0", "size": 40, "state": "directory", "uid": 0 } changed: [sut] => (item=/tmp/httpd3-create) => { "ansible_loop_var": "item", "changed": true, "gid": 0, "group": "root", "item": "/tmp/httpd3-create", "mode": "0755", "owner": "root", "path": "/tmp/httpd3-create", "secontext": "unconfined_u:object_r:user_tmp_t:s0", "size": 40, "state": "directory", "uid": 0 } TASK [fedora.linux_system_roles.podman : Ensure container images are present] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:18 Wednesday 10 July 2024 05:14:22 +0000 (0:00:00.479) 0:02:27.252 ******** changed: [sut] => (item=None) => { "attempts": 1, "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": true } changed: [sut] => (item=None) => { "attempts": 1, "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": true } changed: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": true } TASK [fedora.linux_system_roles.podman : Ensure the quadlet directory is present] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:39 Wednesday 10 July 2024 05:14:40 +0000 (0:00:18.019) 0:02:45.271 ******** ok: [sut] => { "changed": false, "gid": 0, "group": "root", "mode": "0755", "owner": "root", "path": "/etc/containers/systemd", "secontext": "system_u:object_r:etc_t:s0", "size": 4096, "state": "directory", "uid": 0 } TASK [fedora.linux_system_roles.podman : Ensure quadlet file is copied] ******** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:48 Wednesday 10 July 2024 05:14:40 +0000 (0:00:00.247) 0:02:45.519 ******** changed: [sut] => { "changed": true, "checksum": "7a5c73a5d935a42431c87bcdbeb8a04ed0909dc7", "dest": "/etc/containers/systemd/quadlet-demo.kube", "gid": 0, "group": "root", "md5sum": "da53c88f92b68b0487aa209f795b6bb3", "mode": "0644", "owner": "root", "secontext": "system_u:object_r:etc_t:s0", "size": 456, "src": "/root/.ansible/tmp/ansible-tmp-1720588480.89213-12214-115703090519746/source", "state": "file", "uid": 0 } TASK [fedora.linux_system_roles.podman : Ensure quadlet file content is present] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:58 Wednesday 10 July 2024 05:14:41 +0000 (0:00:00.584) 0:02:46.103 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_quadlet_file_src", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Ensure quadlet file is present] ******* task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:70 Wednesday 10 July 2024 05:14:41 +0000 (0:00:00.020) 0:02:46.124 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_copy_file is skipped", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Reload systemctl] ********************* task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:82 Wednesday 10 July 2024 05:14:41 +0000 (0:00:00.017) 0:02:46.142 ******** ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.podman : Start service] ************************ task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:110 Wednesday 10 July 2024 05:14:42 +0000 (0:00:00.812) 0:02:46.954 ******** changed: [sut] => { "changed": true, "name": "quadlet-demo.service", "state": "started", "status": { "AccessSELinuxContext": "system_u:object_r:systemd_unit_file_t:s0", "ActiveEnterTimestampMonotonic": "0", "ActiveExitTimestampMonotonic": "0", "ActiveState": "inactive", "After": "-.mount quadlet-demo-mysql.service basic.target quadlet-demo-network.service sysinit.target systemd-journald.socket system.slice", "AllowIsolate": "no", "AssertResult": "no", "AssertTimestampMonotonic": "0", "Before": "shutdown.target multi-user.target", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "CPUAccounting": "yes", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "[not set]", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "no", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore", "CleanResult": "success", "CollectMode": "inactive", "ConditionResult": "no", "ConditionTimestampMonotonic": "0", "ConfigurationDirectoryMode": "0755", "Conflicts": "shutdown.target", "ControlGroupId": "0", "ControlPID": "0", "CoredumpFilter": "0x33", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "DefaultStartupMemoryLow": "0", "Delegate": "no", "Description": "quadlet-demo.service", "DevicePolicy": "auto", "DropInPaths": "/usr/lib/systemd/system/service.d/10-timeout-abort.conf", "DynamicUser": "no", "Environment": "PODMAN_SYSTEMD_UNIT=quadlet-demo.service", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "0", "ExecMainStartTimestampMonotonic": "0", "ExecMainStatus": "0", "ExecStart": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman kube play --replace --service-container=true --network=systemd-quadlet-demo --configmap /etc/containers/systemd/envoy-proxy-configmap.yml --publish 8000:8080 --publish 9000:9901 /etc/containers/systemd/quadlet-demo.yml ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStartEx": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman kube play --replace --service-container=true --network=systemd-quadlet-demo --configmap /etc/containers/systemd/envoy-proxy-configmap.yml --publish 8000:8080 --publish 9000:9901 /etc/containers/systemd/quadlet-demo.yml ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStopPost": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman kube down /etc/containers/systemd/quadlet-demo.yml ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStopPostEx": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman kube down /etc/containers/systemd/quadlet-demo.yml ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExitType": "main", "ExtensionImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FileDescriptorStorePreserve": "restart", "FinalKillSignal": "9", "FragmentPath": "/run/systemd/generator/quadlet-demo.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOReadBytes": "18446744073709551615", "IOReadOperations": "18446744073709551615", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "IOWeight": "[not set]", "IOWriteBytes": "18446744073709551615", "IOWriteOperations": "18446744073709551615", "IPAccounting": "no", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "Id": "quadlet-demo.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestampMonotonic": "0", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "mixed", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "infinity", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "8388608", "LimitMEMLOCKSoft": "8388608", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "524288", "LimitNOFILESoft": "1024", "LimitNPROC": "14756", "LimitNPROCSoft": "14756", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "14756", "LimitSIGPENDINGSoft": "14756", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "0", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "ManagedOOMSwap": "auto", "MemoryAccounting": "yes", "MemoryAvailable": "infinity", "MemoryCurrent": "[not set]", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryKSM": "no", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemoryPressureThresholdUSec": "200ms", "MemoryPressureWatch": "auto", "MemorySwapMax": "infinity", "MemoryZSwapMax": "infinity", "MountAPIVFS": "no", "MountImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAPolicy": "n/a", "Names": "quadlet-demo.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "all", "OOMPolicy": "stop", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "OnSuccessJobMode": "fail", "Perpetual": "no", "PrivateDevices": "no", "PrivateIPC": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProcSubset": "all", "ProtectClock": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectHostname": "no", "ProtectKernelLogs": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectProc": "default", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "ReloadResult": "success", "ReloadSignal": "1", "RemainAfterExit": "no", "RemoveIPC": "no", "Requires": "sysinit.target system.slice -.mount quadlet-demo-mysql.service quadlet-demo-network.service", "RequiresMountsFor": "/run/containers", "Restart": "no", "RestartKillSignal": "15", "RestartMaxDelayUSec": "infinity", "RestartMode": "normal", "RestartSteps": "0", "RestartUSec": "100ms", "RestartUSecNext": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RootEphemeral": "no", "RootImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "Slice": "system.slice", "SourcePath": "/etc/containers/systemd/quadlet-demo.kube", "StandardError": "inherit", "StandardInput": "null", "StandardOutput": "journal", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StartupMemoryHigh": "18446744073709551615", "StartupMemoryLow": "0", "StartupMemoryMax": "18446744073709551615", "StartupMemorySwapMax": "18446744073709551615", "StartupMemoryZSwapMax": "18446744073709551615", "StateChangeTimestampMonotonic": "0", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "dead", "SuccessAction": "none", "SyslogFacility": "3", "SyslogIdentifier": "quadlet-demo", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "2147483646", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "[not set]", "TasksMax": "4426", "TimeoutAbortUSec": "45s", "TimeoutCleanUSec": "infinity", "TimeoutStartFailureMode": "terminate", "TimeoutStartUSec": "45s", "TimeoutStopFailureMode": "abort", "TimeoutStopUSec": "45s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "notify", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "disabled", "UnitFileState": "generated", "UtmpMode": "init", "WantedBy": "multi-user.target", "WatchdogSignal": "6", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "infinity" } } TASK [fedora.linux_system_roles.podman : Restart service] ********************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:125 Wednesday 10 July 2024 05:14:43 +0000 (0:00:01.335) 0:02:48.290 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_service_started is changed", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Cancel linger] ************************ task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:143 Wednesday 10 July 2024 05:14:43 +0000 (0:00:00.019) 0:02:48.309 ******** skipping: [sut] => { "changed": false, "skipped_reason": "No items in the list" } TASK [fedora.linux_system_roles.podman : Handle credential files - absent] ***** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:149 Wednesday 10 July 2024 05:14:43 +0000 (0:00:00.015) 0:02:48.325 ******** skipping: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Handle certs.d files - absent] ******** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:158 Wednesday 10 July 2024 05:14:43 +0000 (0:00:00.015) 0:02:48.341 ******** skipping: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [Check quadlet files] ***************************************************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/tests/tests_quadlet_demo.yml:84 Wednesday 10 July 2024 05:14:43 +0000 (0:00:00.025) 0:02:48.366 ******** ok: [sut] => { "changed": false, "cmd": [ "ls", "-alrtF", "/etc/containers/systemd" ], "delta": "0:00:00.006798", "end": "2024-07-10 05:14:43.956628", "rc": 0, "start": "2024-07-10 05:14:43.949830" } STDOUT: total 32 drwxr-xr-x. 8 root root 4096 Jul 10 05:13 ../ -rw-r--r--. 1 root root 74 Jul 10 05:13 quadlet-demo.network -rw-r--r--. 1 root root 9 Jul 10 05:14 quadlet-demo-mysql.volume -rw-r--r--. 1 root root 367 Jul 10 05:14 quadlet-demo-mysql.container -rw-r--r--. 1 root root 2102 Jul 10 05:14 envoy-proxy-configmap.yml -rw-r--r--. 1 root root 1605 Jul 10 05:14 quadlet-demo.yml -rw-r--r--. 1 root root 456 Jul 10 05:14 quadlet-demo.kube drwxr-xr-x. 2 root root 4096 Jul 10 05:14 ./ TASK [Check containers] ******************************************************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/tests/tests_quadlet_demo.yml:88 Wednesday 10 July 2024 05:14:43 +0000 (0:00:00.297) 0:02:48.663 ******** ok: [sut] => { "changed": false, "cmd": [ "podman", "ps", "-a" ], "delta": "0:00:00.063239", "end": "2024-07-10 05:14:44.294651", "failed_when_result": false, "rc": 0, "start": "2024-07-10 05:14:44.231412" } STDOUT: CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES 21b1326f5785 quay.io/linux-system-roles/mysql:5.6 mysqld 31 seconds ago Up 31 seconds (starting) quadlet-demo-mysql 9de2222e9853 localhost/podman-pause:4.9.4-1711445992 1 second ago Up 1 second a96f3a51b8d1-service ee4bc05257fc localhost/podman-pause:4.9.4-1711445992 1 second ago Up 1 second 0.0.0.0:8000->8080/tcp, 0.0.0.0:9000->9901/tcp da66443b0d18-infra af5eaff28e9f quay.io/linux-system-roles/wordpress:4.8-apache apache2-foregroun... 1 second ago Up 1 second 0.0.0.0:8000->8080/tcp, 0.0.0.0:9000->9901/tcp quadlet-demo-wordpress dce8c2f6a9a9 quay.io/linux-system-roles/envoyproxy:v1.25.0 envoy -c /etc/env... 1 second ago Up 1 second 0.0.0.0:8000->8080/tcp, 0.0.0.0:9000->9901/tcp quadlet-demo-envoy TASK [Check pods] ************************************************************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/tests/tests_quadlet_demo.yml:93 Wednesday 10 July 2024 05:14:44 +0000 (0:00:00.330) 0:02:48.994 ******** ok: [sut] => { "changed": false, "cmd": [ "podman", "pod", "ps", "--ctr-ids", "--ctr-names", "--ctr-status" ], "delta": "0:00:00.046308", "end": "2024-07-10 05:14:44.616406", "failed_when_result": false, "rc": 0, "start": "2024-07-10 05:14:44.570098" } STDOUT: POD ID NAME STATUS CREATED INFRA ID IDS NAMES STATUS da66443b0d18 quadlet-demo Running 1 second ago ee4bc05257fc ee4bc05257fc,af5eaff28e9f,dce8c2f6a9a9 da66443b0d18-infra,quadlet-demo-wordpress,quadlet-demo-envoy running,running,running TASK [Check systemd] *********************************************************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/tests/tests_quadlet_demo.yml:98 Wednesday 10 July 2024 05:14:44 +0000 (0:00:00.319) 0:02:49.314 ******** ok: [sut] => { "changed": false, "cmd": "set -euo pipefail; systemctl list-units | grep quadlet", "delta": "0:00:00.014096", "end": "2024-07-10 05:14:44.900233", "failed_when_result": false, "rc": 0, "start": "2024-07-10 05:14:44.886137" } STDOUT: quadlet-demo-mysql-volume.service loaded active exited quadlet-demo-mysql-volume.service quadlet-demo-mysql.service loaded active running quadlet-demo-mysql.service quadlet-demo-network.service loaded active exited quadlet-demo-network.service quadlet-demo.service loaded active running quadlet-demo.service TASK [Check web] *************************************************************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/tests/tests_quadlet_demo.yml:104 Wednesday 10 July 2024 05:14:44 +0000 (0:00:00.315) 0:02:49.630 ******** changed: [sut] => { "attempts": 1, "changed": true, "checksum_dest": null, "checksum_src": "d49966cc24f102ead5187c35b5a35440048247ef", "dest": "/run/out", "elapsed": 0, "gid": 0, "group": "root", "md5sum": "a56cfe4431509ef44591163d2a219089", "mode": "0600", "owner": "root", "secontext": "system_u:object_r:var_run_t:s0", "size": 11666, "src": "/root/.ansible/tmp/ansible-moduletmp-1720588485.3553376-550v36b3/tmp_gz0ldug", "state": "file", "status_code": 200, "uid": 0, "url": "https://localhost:8000" } MSG: OK (unknown bytes) TASK [Show web] **************************************************************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/tests/tests_quadlet_demo.yml:115 Wednesday 10 July 2024 05:14:46 +0000 (0:00:01.112) 0:02:50.742 ******** ok: [sut] => { "changed": false, "cmd": [ "cat", "/run/out" ], "delta": "0:00:00.003218", "end": "2024-07-10 05:14:46.277916", "rc": 0, "start": "2024-07-10 05:14:46.274698" } STDOUT: WordPress › Installation

TASK [Error] ******************************************************************* task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/tests/tests_quadlet_demo.yml:120 Wednesday 10 July 2024 05:14:46 +0000 (0:00:00.233) 0:02:50.976 ******** skipping: [sut] => { "changed": false, "false_condition": "__web_status is failed", "skip_reason": "Conditional result was False" } TASK [Check] ******************************************************************* task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/tests/tests_quadlet_demo.yml:131 Wednesday 10 July 2024 05:14:46 +0000 (0:00:00.017) 0:02:50.994 ******** ok: [sut] => { "changed": false, "cmd": [ "podman", "ps", "-a" ], "delta": "0:00:00.039210", "end": "2024-07-10 05:14:46.563898", "rc": 0, "start": "2024-07-10 05:14:46.524688" } STDOUT: CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES 21b1326f5785 quay.io/linux-system-roles/mysql:5.6 mysqld 33 seconds ago Up 33 seconds (starting) quadlet-demo-mysql 9de2222e9853 localhost/podman-pause:4.9.4-1711445992 3 seconds ago Up 3 seconds a96f3a51b8d1-service ee4bc05257fc localhost/podman-pause:4.9.4-1711445992 3 seconds ago Up 3 seconds 0.0.0.0:8000->8080/tcp, 0.0.0.0:9000->9901/tcp da66443b0d18-infra af5eaff28e9f quay.io/linux-system-roles/wordpress:4.8-apache apache2-foregroun... 3 seconds ago Up 3 seconds 0.0.0.0:8000->8080/tcp, 0.0.0.0:9000->9901/tcp quadlet-demo-wordpress dce8c2f6a9a9 quay.io/linux-system-roles/envoyproxy:v1.25.0 envoy -c /etc/env... 3 seconds ago Up 3 seconds 0.0.0.0:8000->8080/tcp, 0.0.0.0:9000->9901/tcp quadlet-demo-envoy TASK [Check pods] ************************************************************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/tests/tests_quadlet_demo.yml:135 Wednesday 10 July 2024 05:14:46 +0000 (0:00:00.267) 0:02:51.261 ******** ok: [sut] => { "changed": false, "cmd": [ "podman", "pod", "ps", "--ctr-ids", "--ctr-names", "--ctr-status" ], "delta": "0:00:00.040209", "end": "2024-07-10 05:14:46.830078", "failed_when_result": false, "rc": 0, "start": "2024-07-10 05:14:46.789869" } STDOUT: POD ID NAME STATUS CREATED INFRA ID IDS NAMES STATUS da66443b0d18 quadlet-demo Running 3 seconds ago ee4bc05257fc ee4bc05257fc,af5eaff28e9f,dce8c2f6a9a9 da66443b0d18-infra,quadlet-demo-wordpress,quadlet-demo-envoy running,running,running TASK [Check systemd] *********************************************************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/tests/tests_quadlet_demo.yml:140 Wednesday 10 July 2024 05:14:46 +0000 (0:00:00.269) 0:02:51.530 ******** ok: [sut] => { "changed": false, "cmd": "set -euo pipefail; systemctl list-units --all | grep quadlet", "delta": "0:00:00.014630", "end": "2024-07-10 05:14:47.077049", "failed_when_result": false, "rc": 0, "start": "2024-07-10 05:14:47.062419" } STDOUT: quadlet-demo-mysql-volume.service loaded active exited quadlet-demo-mysql-volume.service quadlet-demo-mysql.service loaded active running quadlet-demo-mysql.service quadlet-demo-network.service loaded active exited quadlet-demo-network.service quadlet-demo.service loaded active running quadlet-demo.service TASK [LS] ********************************************************************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/tests/tests_quadlet_demo.yml:148 Wednesday 10 July 2024 05:14:47 +0000 (0:00:00.245) 0:02:51.775 ******** ok: [sut] => { "changed": false, "cmd": [ "ls", "-alrtF", "/etc/systemd/system" ], "delta": "0:00:00.004008", "end": "2024-07-10 05:14:47.311045", "failed_when_result": false, "rc": 0, "start": "2024-07-10 05:14:47.307037" } STDOUT: total 56 lrwxrwxrwx. 1 root root 48 Jun 20 10:26 dbus-org.freedesktop.resolve1.service -> /usr/lib/systemd/system/systemd-resolved.service lrwxrwxrwx. 1 root root 44 Jun 20 10:26 dbus-org.freedesktop.oom1.service -> /usr/lib/systemd/system/systemd-oomd.service drwxr-xr-x. 2 root root 4096 Jun 20 10:26 systemd-journald.service.wants/ lrwxrwxrwx. 1 root root 43 Jun 20 10:26 dbus.service -> /usr/lib/systemd/system/dbus-broker.service lrwxrwxrwx. 1 root root 37 Jun 20 10:26 ctrl-alt-del.target -> /usr/lib/systemd/system/reboot.target drwxr-xr-x. 2 root root 4096 Jun 20 10:26 getty.target.wants/ drwxr-xr-x. 5 root root 4096 Jun 20 10:26 ../ lrwxrwxrwx. 1 root root 45 Jun 20 10:26 dbus-org.freedesktop.home1.service -> /usr/lib/systemd/system/systemd-homed.service drwxr-xr-x. 2 root root 4096 Jun 20 10:26 systemd-homed.service.wants/ lrwxrwxrwx. 1 root root 41 Jun 20 10:26 dbus-org.bluez.service -> /usr/lib/systemd/system/bluetooth.service drwxr-xr-x. 2 root root 4096 Jun 20 10:26 bluetooth.target.wants/ drwxr-xr-x. 2 root root 4096 Jun 20 10:26 graphical.target.wants/ drwxr-xr-x. 2 root root 4096 Jun 20 10:27 timers.target.wants/ lrwxrwxrwx. 1 root root 57 Jun 20 10:27 dbus-org.freedesktop.nm-dispatcher.service -> /usr/lib/systemd/system/NetworkManager-dispatcher.service drwxr-xr-x. 2 root root 4096 Jun 20 10:27 network-online.target.wants/ lrwxrwxrwx. 1 root root 41 Jun 20 10:31 default.target -> /usr/lib/systemd/system/multi-user.target drwxr-xr-x. 2 root root 4096 Jun 20 10:45 remote-fs.target.wants/ drwxr-xr-x. 2 root root 4096 Jun 20 10:46 cloud-init.target.wants/ drwxr-xr-x. 2 root root 4096 Jun 20 10:47 sockets.target.wants/ drwxr-xr-x. 2 root root 4096 Jul 10 05:02 sysinit.target.wants/ drwxr-xr-x. 2 root root 4096 Jul 10 05:13 multi-user.target.wants/ lrwxrwxrwx. 1 root root 41 Jul 10 05:13 dbus-org.fedoraproject.FirewallD1.service -> /usr/lib/systemd/system/firewalld.service drwxr-xr-x. 14 root root 4096 Jul 10 05:13 ./ TASK [Cleanup] ***************************************************************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/tests/tests_quadlet_demo.yml:155 Wednesday 10 July 2024 05:14:47 +0000 (0:00:00.235) 0:02:52.010 ******** TASK [fedora.linux_system_roles.podman : Set platform/version specific variables] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:3 Wednesday 10 July 2024 05:14:47 +0000 (0:00:00.042) 0:02:52.053 ******** included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml for sut TASK [fedora.linux_system_roles.podman : Ensure ansible_facts used by role] **** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:3 Wednesday 10 July 2024 05:14:47 +0000 (0:00:00.032) 0:02:52.086 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_required_facts | difference(ansible_facts.keys() | list) | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if system is ostree] ************ task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:11 Wednesday 10 July 2024 05:14:47 +0000 (0:00:00.023) 0:02:52.109 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_is_ostree is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set flag to indicate system is ostree] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:16 Wednesday 10 July 2024 05:14:47 +0000 (0:00:00.018) 0:02:52.128 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_is_ostree is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set platform/version specific variables] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:20 Wednesday 10 July 2024 05:14:47 +0000 (0:00:00.018) 0:02:52.146 ******** ok: [sut] => (item=RedHat.yml) => { "ansible_facts": { "__podman_packages": [ "podman", "shadow-utils-subid" ] }, "ansible_included_var_files": [ "/WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/vars/RedHat.yml" ], "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml" } ok: [sut] => (item=Fedora.yml) => { "ansible_facts": { "__podman_packages": [ "iptables-nft", "podman", "shadow-utils-subid" ] }, "ansible_included_var_files": [ "/WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/vars/Fedora.yml" ], "ansible_loop_var": "item", "changed": false, "item": "Fedora.yml" } skipping: [sut] => (item=Fedora_39.yml) => { "ansible_loop_var": "item", "changed": false, "false_condition": "__vars_file is file", "item": "Fedora_39.yml", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=Fedora_39.yml) => { "ansible_loop_var": "item", "changed": false, "false_condition": "__vars_file is file", "item": "Fedora_39.yml", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Gather the package facts] ************* task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:6 Wednesday 10 July 2024 05:14:47 +0000 (0:00:00.041) 0:02:52.188 ******** ok: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Enable copr if requested] ************* task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:10 Wednesday 10 July 2024 05:14:48 +0000 (0:00:00.841) 0:02:53.029 ******** skipping: [sut] => { "changed": false, "false_condition": "podman_use_copr | d(false)", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Ensure required packages are installed] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:14 Wednesday 10 July 2024 05:14:48 +0000 (0:00:00.020) 0:02:53.049 ******** skipping: [sut] => { "changed": false, "false_condition": "(__podman_packages | difference(ansible_facts.packages))", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get podman version] ******************* task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:22 Wednesday 10 July 2024 05:14:48 +0000 (0:00:00.022) 0:02:53.072 ******** ok: [sut] => { "changed": false, "cmd": [ "podman", "--version" ], "delta": "0:00:00.029344", "end": "2024-07-10 05:14:48.632335", "rc": 0, "start": "2024-07-10 05:14:48.602991" } STDOUT: podman version 4.9.4 TASK [fedora.linux_system_roles.podman : Set podman version] ******************* task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:28 Wednesday 10 July 2024 05:14:48 +0000 (0:00:00.259) 0:02:53.331 ******** ok: [sut] => { "ansible_facts": { "podman_version": "4.9.4" }, "changed": false } TASK [fedora.linux_system_roles.podman : Podman package version must be 4.2 or later] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:32 Wednesday 10 July 2024 05:14:48 +0000 (0:00:00.022) 0:02:53.354 ******** skipping: [sut] => { "changed": false, "false_condition": "podman_version is version(\"4.2\", \"<\")", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Podman package version must be 4.4 or later for quadlet, secrets] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:39 Wednesday 10 July 2024 05:14:48 +0000 (0:00:00.018) 0:02:53.372 ******** skipping: [sut] => { "changed": false, "false_condition": "podman_version is version(\"4.4\", \"<\")", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Podman package version must be 4.4 or later for quadlet, secrets] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:49 Wednesday 10 July 2024 05:14:48 +0000 (0:00:00.023) 0:02:53.395 ******** META: end_host conditional evaluated to False, continuing execution for sut skipping: [sut] => { "skip_reason": "end_host conditional evaluated to False, continuing execution for sut" } MSG: end_host conditional evaluated to false, continuing execution for sut TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:56 Wednesday 10 July 2024 05:14:48 +0000 (0:00:00.025) 0:02:53.421 ******** included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for sut TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Wednesday 10 July 2024 05:14:48 +0000 (0:00:00.040) 0:02:53.461 ******** skipping: [sut] => { "changed": false, "false_condition": "'getent_passwd' not in ansible_facts or __podman_user not in ansible_facts['getent_passwd']", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Wednesday 10 July 2024 05:14:48 +0000 (0:00:00.023) 0:02:53.484 ******** skipping: [sut] => { "changed": false, "false_condition": "not ansible_facts[\"getent_passwd\"][__podman_user]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Wednesday 10 July 2024 05:14:48 +0000 (0:00:00.022) 0:02:53.507 ******** ok: [sut] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get group information] **************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:28 Wednesday 10 July 2024 05:14:48 +0000 (0:00:00.027) 0:02:53.534 ******** ok: [sut] => { "ansible_facts": { "getent_group": { "root": [ "x", "0", "" ] } }, "changed": false } TASK [fedora.linux_system_roles.podman : Set group name] *********************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:35 Wednesday 10 July 2024 05:14:49 +0000 (0:00:00.284) 0:02:53.819 ******** ok: [sut] => { "ansible_facts": { "__podman_group_name": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:39 Wednesday 10 July 2024 05:14:49 +0000 (0:00:00.026) 0:02:53.845 ******** ok: [sut] => { "changed": false, "stat": { "atime": 1720588427.5164855, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "72ba28a3be19c467e89ab90f5d21f99b605159a7", "ctime": 1720588377.3007717, "dev": 51714, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 203603, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-pie-executable", "mode": "0755", "mtime": 1696291200.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 16032, "uid": 0, "version": "2092964490", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************ task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:50 Wednesday 10 July 2024 05:14:49 +0000 (0:00:00.239) 0:02:54.084 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check group with getsubids] *********** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:55 Wednesday 10 July 2024 05:14:49 +0000 (0:00:00.019) 0:02:54.103 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:60 Wednesday 10 July 2024 05:14:49 +0000 (0:00:00.052) 0:02:54.156 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subuid file] ********************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:74 Wednesday 10 July 2024 05:14:49 +0000 (0:00:00.020) 0:02:54.176 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subgid file] ********************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:79 Wednesday 10 July 2024 05:14:49 +0000 (0:00:00.017) 0:02:54.194 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:84 Wednesday 10 July 2024 05:14:49 +0000 (0:00:00.018) 0:02:54.212 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:94 Wednesday 10 July 2024 05:14:49 +0000 (0:00:00.017) 0:02:54.230 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] ***** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:101 Wednesday 10 July 2024 05:14:49 +0000 (0:00:00.018) 0:02:54.248 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set config file paths] **************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:62 Wednesday 10 July 2024 05:14:49 +0000 (0:00:00.018) 0:02:54.266 ******** ok: [sut] => { "ansible_facts": { "__podman_container_conf_file": "/etc/containers/containers.conf.d/50-systemroles.conf", "__podman_policy_json_file": "/etc/containers/policy.json", "__podman_registries_conf_file": "/etc/containers/registries.conf.d/50-systemroles.conf", "__podman_storage_conf_file": "/etc/containers/storage.conf" }, "changed": false } TASK [fedora.linux_system_roles.podman : Handle container.conf.d] ************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:71 Wednesday 10 July 2024 05:14:49 +0000 (0:00:00.024) 0:02:54.291 ******** included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_container_conf_d.yml for sut TASK [fedora.linux_system_roles.podman : Ensure containers.d exists] *********** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_container_conf_d.yml:5 Wednesday 10 July 2024 05:14:49 +0000 (0:00:00.037) 0:02:54.328 ******** skipping: [sut] => { "changed": false, "false_condition": "podman_containers_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Update container config file] ********* task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_container_conf_d.yml:13 Wednesday 10 July 2024 05:14:49 +0000 (0:00:00.018) 0:02:54.347 ******** skipping: [sut] => { "changed": false, "false_condition": "podman_containers_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Handle registries.conf.d] ************* task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:74 Wednesday 10 July 2024 05:14:49 +0000 (0:00:00.019) 0:02:54.367 ******** included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_registries_conf_d.yml for sut TASK [fedora.linux_system_roles.podman : Ensure registries.d exists] *********** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_registries_conf_d.yml:5 Wednesday 10 July 2024 05:14:49 +0000 (0:00:00.037) 0:02:54.404 ******** skipping: [sut] => { "changed": false, "false_condition": "podman_registries_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Update registries config file] ******** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_registries_conf_d.yml:13 Wednesday 10 July 2024 05:14:49 +0000 (0:00:00.019) 0:02:54.423 ******** skipping: [sut] => { "changed": false, "false_condition": "podman_registries_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Handle storage.conf] ****************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:77 Wednesday 10 July 2024 05:14:49 +0000 (0:00:00.018) 0:02:54.442 ******** included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_storage_conf.yml for sut TASK [fedora.linux_system_roles.podman : Ensure storage.conf parent dir exists] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_storage_conf.yml:5 Wednesday 10 July 2024 05:14:49 +0000 (0:00:00.038) 0:02:54.480 ******** skipping: [sut] => { "changed": false, "false_condition": "podman_storage_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Update storage config file] *********** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_storage_conf.yml:13 Wednesday 10 July 2024 05:14:49 +0000 (0:00:00.048) 0:02:54.528 ******** skipping: [sut] => { "changed": false, "false_condition": "podman_storage_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Handle policy.json] ******************* task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:80 Wednesday 10 July 2024 05:14:49 +0000 (0:00:00.020) 0:02:54.549 ******** included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml for sut TASK [fedora.linux_system_roles.podman : Ensure policy.json parent dir exists] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:6 Wednesday 10 July 2024 05:14:49 +0000 (0:00:00.040) 0:02:54.590 ******** skipping: [sut] => { "changed": false, "false_condition": "podman_policy_json | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Stat the policy.json file] ************ task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:14 Wednesday 10 July 2024 05:14:49 +0000 (0:00:00.020) 0:02:54.611 ******** skipping: [sut] => { "changed": false, "false_condition": "podman_policy_json | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get the existing policy.json] ********* task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:19 Wednesday 10 July 2024 05:14:49 +0000 (0:00:00.019) 0:02:54.630 ******** skipping: [sut] => { "changed": false, "false_condition": "podman_policy_json | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Write new policy.json file] *********** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:25 Wednesday 10 July 2024 05:14:49 +0000 (0:00:00.019) 0:02:54.650 ******** skipping: [sut] => { "changed": false, "false_condition": "podman_policy_json | length > 0", "skip_reason": "Conditional result was False" } TASK [Manage firewall for specified ports] ************************************* task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:86 Wednesday 10 July 2024 05:14:50 +0000 (0:00:00.018) 0:02:54.668 ******** TASK [fedora.linux_system_roles.firewall : Setup firewalld] ******************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:2 Wednesday 10 July 2024 05:14:50 +0000 (0:00:00.065) 0:02:54.734 ******** included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml for sut TASK [fedora.linux_system_roles.firewall : Ensure ansible_facts used by role] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:2 Wednesday 10 July 2024 05:14:50 +0000 (0:00:00.032) 0:02:54.766 ******** skipping: [sut] => { "changed": false, "false_condition": "__firewall_required_facts | difference(ansible_facts.keys() | list) | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Check if system is ostree] ********** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:10 Wednesday 10 July 2024 05:14:50 +0000 (0:00:00.025) 0:02:54.791 ******** skipping: [sut] => { "changed": false, "false_condition": "not __firewall_is_ostree is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Set flag to indicate system is ostree] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:15 Wednesday 10 July 2024 05:14:50 +0000 (0:00:00.019) 0:02:54.810 ******** skipping: [sut] => { "changed": false, "false_condition": "not __firewall_is_ostree is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Install firewalld] ****************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:19 Wednesday 10 July 2024 05:14:50 +0000 (0:00:00.018) 0:02:54.829 ******** ok: [sut] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [fedora.linux_system_roles.firewall : Collect service facts] ************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:5 Wednesday 10 July 2024 05:14:52 +0000 (0:00:02.286) 0:02:57.116 ******** skipping: [sut] => { "changed": false, "false_condition": "firewall_disable_conflicting_services | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Attempt to stop and disable conflicting services] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:9 Wednesday 10 July 2024 05:14:52 +0000 (0:00:00.018) 0:02:57.134 ******** skipping: [sut] => (item=nftables) => { "ansible_loop_var": "item", "changed": false, "false_condition": "firewall_disable_conflicting_services | bool", "item": "nftables", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=iptables) => { "ansible_loop_var": "item", "changed": false, "false_condition": "firewall_disable_conflicting_services | bool", "item": "iptables", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=ufw) => { "ansible_loop_var": "item", "changed": false, "false_condition": "firewall_disable_conflicting_services | bool", "item": "ufw", "skip_reason": "Conditional result was False" } skipping: [sut] => { "changed": false } MSG: All items skipped TASK [fedora.linux_system_roles.firewall : Unmask firewalld service] *********** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:22 Wednesday 10 July 2024 05:14:52 +0000 (0:00:00.055) 0:02:57.189 ******** ok: [sut] => { "changed": false, "name": "firewalld", "status": { "AccessSELinuxContext": "system_u:object_r:firewalld_unit_file_t:s0", "ActiveEnterTimestamp": "Wed 2024-07-10 05:13:52 UTC", "ActiveEnterTimestampMonotonic": "952448512", "ActiveExitTimestampMonotonic": "0", "ActiveState": "active", "After": "basic.target polkit.service dbus.socket system.slice dbus-broker.service sysinit.target", "AllowIsolate": "no", "AssertResult": "yes", "AssertTimestamp": "Wed 2024-07-10 05:13:51 UTC", "AssertTimestampMonotonic": "952190941", "Before": "shutdown.target network-pre.target multi-user.target", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "BusName": "org.fedoraproject.FirewallD1", "CPUAccounting": "yes", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "589336000", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "yes", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore", "CleanResult": "success", "CollectMode": "inactive", "ConditionResult": "yes", "ConditionTimestamp": "Wed 2024-07-10 05:13:51 UTC", "ConditionTimestampMonotonic": "952190937", "ConfigurationDirectoryMode": "0755", "Conflicts": "shutdown.target iptables.service nftables.service ebtables.service ip6tables.service ipset.service", "ControlGroup": "/system.slice/firewalld.service", "ControlGroupId": "6620", "ControlPID": "0", "CoredumpFilter": "0x33", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "DefaultStartupMemoryLow": "0", "Delegate": "no", "Description": "firewalld - dynamic firewall daemon", "DevicePolicy": "auto", "Documentation": "\"man:firewalld(1)\"", "DropInPaths": "/usr/lib/systemd/system/service.d/10-timeout-abort.conf", "DynamicUser": "no", "EffectiveCPUs": "0-1", "EffectiveMemoryNodes": "0", "EnvironmentFiles": "/etc/sysconfig/firewalld (ignore_errors=yes)", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "6763", "ExecMainStartTimestamp": "Wed 2024-07-10 05:13:51 UTC", "ExecMainStartTimestampMonotonic": "952204109", "ExecMainStatus": "0", "ExecReload": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecReloadEx": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStart": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStartEx": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExitType": "main", "ExtensionImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FileDescriptorStorePreserve": "restart", "FinalKillSignal": "9", "FragmentPath": "/usr/lib/systemd/system/firewalld.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOReadBytes": "18446744073709551615", "IOReadOperations": "18446744073709551615", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "IOWeight": "[not set]", "IOWriteBytes": "18446744073709551615", "IOWriteOperations": "18446744073709551615", "IPAccounting": "no", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "Id": "firewalld.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestamp": "Wed 2024-07-10 05:13:51 UTC", "InactiveExitTimestampMonotonic": "952204469", "InvocationID": "6eb0bd9cbf764a82a46a19747bb71bc4", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "mixed", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "infinity", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "8388608", "LimitMEMLOCKSoft": "8388608", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "524288", "LimitNOFILESoft": "1024", "LimitNPROC": "14756", "LimitNPROCSoft": "14756", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "14756", "LimitSIGPENDINGSoft": "14756", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "6763", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "ManagedOOMSwap": "auto", "MemoryAccounting": "yes", "MemoryAvailable": "infinity", "MemoryCurrent": "34697216", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryKSM": "no", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemoryPressureThresholdUSec": "200ms", "MemoryPressureWatch": "auto", "MemorySwapMax": "infinity", "MemoryZSwapMax": "infinity", "MountAPIVFS": "no", "MountImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAPolicy": "n/a", "Names": "firewalld.service dbus-org.fedoraproject.FirewallD1.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMPolicy": "stop", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "OnSuccessJobMode": "fail", "Perpetual": "no", "PrivateDevices": "no", "PrivateIPC": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProcSubset": "all", "ProtectClock": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectHostname": "no", "ProtectKernelLogs": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectProc": "default", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "ReloadResult": "success", "ReloadSignal": "1", "RemainAfterExit": "no", "RemoveIPC": "no", "Requires": "sysinit.target system.slice dbus.socket", "Restart": "no", "RestartKillSignal": "15", "RestartMaxDelayUSec": "infinity", "RestartMode": "normal", "RestartSteps": "0", "RestartUSec": "100ms", "RestartUSecNext": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RootEphemeral": "no", "RootImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "Slice": "system.slice", "StandardError": "null", "StandardInput": "null", "StandardOutput": "null", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StartupMemoryHigh": "18446744073709551615", "StartupMemoryLow": "0", "StartupMemoryMax": "18446744073709551615", "StartupMemorySwapMax": "18446744073709551615", "StartupMemoryZSwapMax": "18446744073709551615", "StateChangeTimestamp": "Wed 2024-07-10 05:14:42 UTC", "StateChangeTimestampMonotonic": "1002588722", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "running", "SuccessAction": "none", "SyslogFacility": "3", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "2147483646", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "2", "TasksMax": "4426", "TimeoutAbortUSec": "45s", "TimeoutCleanUSec": "infinity", "TimeoutStartFailureMode": "terminate", "TimeoutStartUSec": "45s", "TimeoutStopFailureMode": "abort", "TimeoutStopUSec": "45s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "dbus", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "enabled", "UnitFileState": "enabled", "UtmpMode": "init", "WantedBy": "multi-user.target", "Wants": "network-pre.target", "WatchdogSignal": "6", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "0" } } TASK [fedora.linux_system_roles.firewall : Enable and start firewalld service] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:28 Wednesday 10 July 2024 05:14:52 +0000 (0:00:00.414) 0:02:57.604 ******** ok: [sut] => { "changed": false, "enabled": true, "name": "firewalld", "state": "started", "status": { "AccessSELinuxContext": "system_u:object_r:firewalld_unit_file_t:s0", "ActiveEnterTimestamp": "Wed 2024-07-10 05:13:52 UTC", "ActiveEnterTimestampMonotonic": "952448512", "ActiveExitTimestampMonotonic": "0", "ActiveState": "active", "After": "basic.target polkit.service dbus.socket system.slice dbus-broker.service sysinit.target", "AllowIsolate": "no", "AssertResult": "yes", "AssertTimestamp": "Wed 2024-07-10 05:13:51 UTC", "AssertTimestampMonotonic": "952190941", "Before": "shutdown.target network-pre.target multi-user.target", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "BusName": "org.fedoraproject.FirewallD1", "CPUAccounting": "yes", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "589336000", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "yes", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore", "CleanResult": "success", "CollectMode": "inactive", "ConditionResult": "yes", "ConditionTimestamp": "Wed 2024-07-10 05:13:51 UTC", "ConditionTimestampMonotonic": "952190937", "ConfigurationDirectoryMode": "0755", "Conflicts": "shutdown.target iptables.service nftables.service ebtables.service ip6tables.service ipset.service", "ControlGroup": "/system.slice/firewalld.service", "ControlGroupId": "6620", "ControlPID": "0", "CoredumpFilter": "0x33", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "DefaultStartupMemoryLow": "0", "Delegate": "no", "Description": "firewalld - dynamic firewall daemon", "DevicePolicy": "auto", "Documentation": "\"man:firewalld(1)\"", "DropInPaths": "/usr/lib/systemd/system/service.d/10-timeout-abort.conf", "DynamicUser": "no", "EffectiveCPUs": "0-1", "EffectiveMemoryNodes": "0", "EnvironmentFiles": "/etc/sysconfig/firewalld (ignore_errors=yes)", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "6763", "ExecMainStartTimestamp": "Wed 2024-07-10 05:13:51 UTC", "ExecMainStartTimestampMonotonic": "952204109", "ExecMainStatus": "0", "ExecReload": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecReloadEx": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStart": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStartEx": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExitType": "main", "ExtensionImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FileDescriptorStorePreserve": "restart", "FinalKillSignal": "9", "FragmentPath": "/usr/lib/systemd/system/firewalld.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOReadBytes": "18446744073709551615", "IOReadOperations": "18446744073709551615", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "IOWeight": "[not set]", "IOWriteBytes": "18446744073709551615", "IOWriteOperations": "18446744073709551615", "IPAccounting": "no", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "Id": "firewalld.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestamp": "Wed 2024-07-10 05:13:51 UTC", "InactiveExitTimestampMonotonic": "952204469", "InvocationID": "6eb0bd9cbf764a82a46a19747bb71bc4", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "mixed", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "infinity", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "8388608", "LimitMEMLOCKSoft": "8388608", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "524288", "LimitNOFILESoft": "1024", "LimitNPROC": "14756", "LimitNPROCSoft": "14756", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "14756", "LimitSIGPENDINGSoft": "14756", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "6763", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "ManagedOOMSwap": "auto", "MemoryAccounting": "yes", "MemoryAvailable": "infinity", "MemoryCurrent": "34697216", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryKSM": "no", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemoryPressureThresholdUSec": "200ms", "MemoryPressureWatch": "auto", "MemorySwapMax": "infinity", "MemoryZSwapMax": "infinity", "MountAPIVFS": "no", "MountImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAPolicy": "n/a", "Names": "firewalld.service dbus-org.fedoraproject.FirewallD1.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMPolicy": "stop", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "OnSuccessJobMode": "fail", "Perpetual": "no", "PrivateDevices": "no", "PrivateIPC": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProcSubset": "all", "ProtectClock": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectHostname": "no", "ProtectKernelLogs": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectProc": "default", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "ReloadResult": "success", "ReloadSignal": "1", "RemainAfterExit": "no", "RemoveIPC": "no", "Requires": "sysinit.target system.slice dbus.socket", "Restart": "no", "RestartKillSignal": "15", "RestartMaxDelayUSec": "infinity", "RestartMode": "normal", "RestartSteps": "0", "RestartUSec": "100ms", "RestartUSecNext": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RootEphemeral": "no", "RootImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "Slice": "system.slice", "StandardError": "null", "StandardInput": "null", "StandardOutput": "null", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StartupMemoryHigh": "18446744073709551615", "StartupMemoryLow": "0", "StartupMemoryMax": "18446744073709551615", "StartupMemorySwapMax": "18446744073709551615", "StartupMemoryZSwapMax": "18446744073709551615", "StateChangeTimestamp": "Wed 2024-07-10 05:14:42 UTC", "StateChangeTimestampMonotonic": "1002588722", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "running", "SuccessAction": "none", "SyslogFacility": "3", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "2147483646", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "2", "TasksMax": "4426", "TimeoutAbortUSec": "45s", "TimeoutCleanUSec": "infinity", "TimeoutStartFailureMode": "terminate", "TimeoutStartUSec": "45s", "TimeoutStopFailureMode": "abort", "TimeoutStopUSec": "45s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "dbus", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "enabled", "UnitFileState": "enabled", "UtmpMode": "init", "WantedBy": "multi-user.target", "Wants": "network-pre.target", "WatchdogSignal": "6", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "0" } } TASK [fedora.linux_system_roles.firewall : Check if previous replaced is defined] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:34 Wednesday 10 July 2024 05:14:53 +0000 (0:00:00.415) 0:02:58.020 ******** ok: [sut] => { "ansible_facts": { "__firewall_previous_replaced": false, "__firewall_python_cmd": "/usr/bin/python3", "__firewall_report_changed": true }, "changed": false } TASK [fedora.linux_system_roles.firewall : Get config files, checksums before and remove] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:43 Wednesday 10 July 2024 05:14:53 +0000 (0:00:00.026) 0:02:58.047 ******** skipping: [sut] => { "changed": false, "false_condition": "__firewall_previous_replaced | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Tell firewall module it is able to report changed] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:55 Wednesday 10 July 2024 05:14:53 +0000 (0:00:00.018) 0:02:58.065 ******** skipping: [sut] => { "changed": false, "false_condition": "__firewall_previous_replaced | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Configure firewall] ***************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:71 Wednesday 10 July 2024 05:14:53 +0000 (0:00:00.018) 0:02:58.084 ******** ok: [sut] => (item={'port': '8000/tcp', 'state': 'enabled'}) => { "__firewall_changed": false, "ansible_loop_var": "item", "changed": false, "item": { "port": "8000/tcp", "state": "enabled" } } ok: [sut] => (item={'port': '9000/tcp', 'state': 'enabled'}) => { "__firewall_changed": false, "ansible_loop_var": "item", "changed": false, "item": { "port": "9000/tcp", "state": "enabled" } } TASK [fedora.linux_system_roles.firewall : Gather firewall config information] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:120 Wednesday 10 July 2024 05:14:54 +0000 (0:00:00.753) 0:02:58.838 ******** skipping: [sut] => (item={'port': '8000/tcp', 'state': 'enabled'}) => { "ansible_loop_var": "item", "changed": false, "false_condition": "firewall | length == 1", "item": { "port": "8000/tcp", "state": "enabled" }, "skip_reason": "Conditional result was False" } skipping: [sut] => (item={'port': '9000/tcp', 'state': 'enabled'}) => { "ansible_loop_var": "item", "changed": false, "false_condition": "firewall | length == 1", "item": { "port": "9000/tcp", "state": "enabled" }, "skip_reason": "Conditional result was False" } skipping: [sut] => { "changed": false } MSG: All items skipped TASK [fedora.linux_system_roles.firewall : Update firewalld_config fact] ******* task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:130 Wednesday 10 July 2024 05:14:54 +0000 (0:00:00.030) 0:02:58.868 ******** skipping: [sut] => { "changed": false, "false_condition": "firewall | length == 1", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Gather firewall config if no arguments] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:139 Wednesday 10 July 2024 05:14:54 +0000 (0:00:00.021) 0:02:58.889 ******** skipping: [sut] => { "changed": false, "false_condition": "firewall == None or firewall | length == 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Update firewalld_config fact] ******* task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:144 Wednesday 10 July 2024 05:14:54 +0000 (0:00:00.020) 0:02:58.909 ******** skipping: [sut] => { "changed": false, "false_condition": "firewall == None or firewall | length == 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Get config files, checksums after] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:153 Wednesday 10 July 2024 05:14:54 +0000 (0:00:00.019) 0:02:58.929 ******** skipping: [sut] => { "changed": false, "false_condition": "__firewall_previous_replaced | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Calculate what has changed] ********* task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:163 Wednesday 10 July 2024 05:14:54 +0000 (0:00:00.017) 0:02:58.947 ******** skipping: [sut] => { "changed": false, "false_condition": "__firewall_previous_replaced | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Show diffs] ************************* task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:169 Wednesday 10 July 2024 05:14:54 +0000 (0:00:00.018) 0:02:58.965 ******** skipping: [sut] => { "false_condition": "__firewall_previous_replaced | bool" } TASK [Manage selinux for specified ports] ************************************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:93 Wednesday 10 July 2024 05:14:54 +0000 (0:00:00.028) 0:02:58.994 ******** skipping: [sut] => { "changed": false, "false_condition": "podman_selinux_ports | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Keep track of users that need to cancel linger] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:100 Wednesday 10 July 2024 05:14:54 +0000 (0:00:00.019) 0:02:59.013 ******** ok: [sut] => { "ansible_facts": { "__podman_cancel_user_linger": [] }, "changed": false } TASK [fedora.linux_system_roles.podman : Handle certs.d files - present] ******* task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:104 Wednesday 10 July 2024 05:14:54 +0000 (0:00:00.018) 0:02:59.031 ******** skipping: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Handle credential files - present] **** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:113 Wednesday 10 July 2024 05:14:54 +0000 (0:00:00.016) 0:02:59.048 ******** skipping: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Handle secrets] *********************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:122 Wednesday 10 July 2024 05:14:54 +0000 (0:00:00.015) 0:02:59.063 ******** included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml for sut => (item=(censored due to no_log)) included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml for sut => (item=(censored due to no_log)) included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml for sut => (item=(censored due to no_log)) TASK [fedora.linux_system_roles.podman : Set variables part 1] ***************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:3 Wednesday 10 July 2024 05:14:54 +0000 (0:00:00.087) 0:02:59.151 ******** ok: [sut] => { "ansible_facts": { "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set variables part 2] ***************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:7 Wednesday 10 July 2024 05:14:54 +0000 (0:00:00.023) 0:02:59.175 ******** ok: [sut] => { "ansible_facts": { "__podman_rootless": false, "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:13 Wednesday 10 July 2024 05:14:54 +0000 (0:00:00.026) 0:02:59.201 ******** included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for sut TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Wednesday 10 July 2024 05:14:54 +0000 (0:00:00.031) 0:02:59.232 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Wednesday 10 July 2024 05:14:54 +0000 (0:00:00.019) 0:02:59.252 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Wednesday 10 July 2024 05:14:54 +0000 (0:00:00.018) 0:02:59.270 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Stat XDG_RUNTIME_DIR] ***************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:18 Wednesday 10 July 2024 05:14:54 +0000 (0:00:00.018) 0:02:59.288 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Manage each secret] ******************* task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:34 Wednesday 10 July 2024 05:14:54 +0000 (0:00:00.018) 0:02:59.307 ******** changed: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": true } TASK [fedora.linux_system_roles.podman : Set variables part 1] ***************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:3 Wednesday 10 July 2024 05:14:54 +0000 (0:00:00.282) 0:02:59.590 ******** ok: [sut] => { "ansible_facts": { "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set variables part 2] ***************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:7 Wednesday 10 July 2024 05:14:54 +0000 (0:00:00.022) 0:02:59.613 ******** ok: [sut] => { "ansible_facts": { "__podman_rootless": false, "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:13 Wednesday 10 July 2024 05:14:54 +0000 (0:00:00.025) 0:02:59.638 ******** included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for sut TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Wednesday 10 July 2024 05:14:55 +0000 (0:00:00.031) 0:02:59.669 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Wednesday 10 July 2024 05:14:55 +0000 (0:00:00.018) 0:02:59.688 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Wednesday 10 July 2024 05:14:55 +0000 (0:00:00.018) 0:02:59.706 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Stat XDG_RUNTIME_DIR] ***************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:18 Wednesday 10 July 2024 05:14:55 +0000 (0:00:00.017) 0:02:59.724 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Manage each secret] ******************* task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:34 Wednesday 10 July 2024 05:14:55 +0000 (0:00:00.049) 0:02:59.773 ******** changed: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": true } TASK [fedora.linux_system_roles.podman : Set variables part 1] ***************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:3 Wednesday 10 July 2024 05:14:55 +0000 (0:00:00.284) 0:03:00.057 ******** ok: [sut] => { "ansible_facts": { "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set variables part 2] ***************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:7 Wednesday 10 July 2024 05:14:55 +0000 (0:00:00.023) 0:03:00.081 ******** ok: [sut] => { "ansible_facts": { "__podman_rootless": false, "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:13 Wednesday 10 July 2024 05:14:55 +0000 (0:00:00.026) 0:03:00.107 ******** included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for sut TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Wednesday 10 July 2024 05:14:55 +0000 (0:00:00.030) 0:03:00.138 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Wednesday 10 July 2024 05:14:55 +0000 (0:00:00.018) 0:03:00.157 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Wednesday 10 July 2024 05:14:55 +0000 (0:00:00.018) 0:03:00.175 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Stat XDG_RUNTIME_DIR] ***************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:18 Wednesday 10 July 2024 05:14:55 +0000 (0:00:00.018) 0:03:00.194 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Manage each secret] ******************* task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:34 Wednesday 10 July 2024 05:14:55 +0000 (0:00:00.017) 0:03:00.211 ******** changed: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": true } TASK [fedora.linux_system_roles.podman : Handle Kubernetes specifications] ***** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:129 Wednesday 10 July 2024 05:14:55 +0000 (0:00:00.282) 0:03:00.494 ******** skipping: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Handle Quadlet specifications] ******** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:136 Wednesday 10 July 2024 05:14:55 +0000 (0:00:00.017) 0:03:00.511 ******** included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for sut => (item=(censored due to no_log)) included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for sut => (item=(censored due to no_log)) included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for sut => (item=(censored due to no_log)) included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for sut => (item=(censored due to no_log)) included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for sut => (item=(censored due to no_log)) included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for sut => (item=(censored due to no_log)) TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14 Wednesday 10 July 2024 05:14:55 +0000 (0:00:00.090) 0:03:00.601 ******** ok: [sut] => { "ansible_facts": { "__podman_quadlet_file_src": "quadlet-demo.kube", "__podman_quadlet_spec": {}, "__podman_quadlet_str": "[Install]\nWantedBy=default.target\n\n[Unit]\nRequires=quadlet-demo-mysql.service\nAfter=quadlet-demo-mysql.service\n\n[Kube]\n# Point to the yaml file in the same directory\nYaml=quadlet-demo.yml\n# Use the quadlet-demo network\nNetwork=quadlet-demo.network\n# Publish the envoy proxy data port\nPublishPort=8000:8080\n# Publish the envoy proxy admin port\nPublishPort=9000:9901\n# Use the envoy proxy config map in the same directory\nConfigMap=envoy-proxy-configmap.yml", "__podman_quadlet_template_src": "" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25 Wednesday 10 July 2024 05:14:55 +0000 (0:00:00.030) 0:03:00.631 ******** ok: [sut] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_pull_image": true, "__podman_state": "absent", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] ***** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35 Wednesday 10 July 2024 05:14:55 +0000 (0:00:00.025) 0:03:00.657 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_quadlet_file_src", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48 Wednesday 10 July 2024 05:14:56 +0000 (0:00:00.019) 0:03:00.676 ******** ok: [sut] => { "ansible_facts": { "__podman_quadlet_name": "quadlet-demo", "__podman_quadlet_type": "kube", "__podman_rootless": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57 Wednesday 10 July 2024 05:14:56 +0000 (0:00:00.065) 0:03:00.741 ******** included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for sut TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Wednesday 10 July 2024 05:14:56 +0000 (0:00:00.037) 0:03:00.779 ******** skipping: [sut] => { "changed": false, "false_condition": "'getent_passwd' not in ansible_facts or __podman_user not in ansible_facts['getent_passwd']", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Wednesday 10 July 2024 05:14:56 +0000 (0:00:00.023) 0:03:00.803 ******** skipping: [sut] => { "changed": false, "false_condition": "not ansible_facts[\"getent_passwd\"][__podman_user]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Wednesday 10 July 2024 05:14:56 +0000 (0:00:00.022) 0:03:00.825 ******** ok: [sut] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get group information] **************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:28 Wednesday 10 July 2024 05:14:56 +0000 (0:00:00.029) 0:03:00.854 ******** ok: [sut] => { "ansible_facts": { "getent_group": { "root": [ "x", "0", "" ] } }, "changed": false } TASK [fedora.linux_system_roles.podman : Set group name] *********************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:35 Wednesday 10 July 2024 05:14:56 +0000 (0:00:00.241) 0:03:01.095 ******** ok: [sut] => { "ansible_facts": { "__podman_group_name": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:39 Wednesday 10 July 2024 05:14:56 +0000 (0:00:00.029) 0:03:01.124 ******** ok: [sut] => { "changed": false, "stat": { "atime": 1720588427.5164855, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "72ba28a3be19c467e89ab90f5d21f99b605159a7", "ctime": 1720588377.3007717, "dev": 51714, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 203603, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-pie-executable", "mode": "0755", "mtime": 1696291200.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 16032, "uid": 0, "version": "2092964490", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************ task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:50 Wednesday 10 July 2024 05:14:56 +0000 (0:00:00.240) 0:03:01.364 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check group with getsubids] *********** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:55 Wednesday 10 July 2024 05:14:56 +0000 (0:00:00.018) 0:03:01.383 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:60 Wednesday 10 July 2024 05:14:56 +0000 (0:00:00.019) 0:03:01.402 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subuid file] ********************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:74 Wednesday 10 July 2024 05:14:56 +0000 (0:00:00.017) 0:03:01.420 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subgid file] ********************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:79 Wednesday 10 July 2024 05:14:56 +0000 (0:00:00.018) 0:03:01.439 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:84 Wednesday 10 July 2024 05:14:56 +0000 (0:00:00.017) 0:03:01.457 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:94 Wednesday 10 July 2024 05:14:56 +0000 (0:00:00.018) 0:03:01.475 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] ***** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:101 Wednesday 10 July 2024 05:14:56 +0000 (0:00:00.018) 0:03:01.493 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62 Wednesday 10 July 2024 05:14:56 +0000 (0:00:00.017) 0:03:01.511 ******** ok: [sut] => { "ansible_facts": { "__podman_activate_systemd_unit": true, "__podman_images_found": [], "__podman_kube_yamls_raw": [ "quadlet-demo.yml" ], "__podman_service_name": "quadlet-demo.service", "__podman_systemd_scope": "system", "__podman_user_home_dir": "/root", "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73 Wednesday 10 July 2024 05:14:56 +0000 (0:00:00.035) 0:03:01.546 ******** ok: [sut] => { "ansible_facts": { "__podman_quadlet_path": "/etc/containers/systemd" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get kube yaml contents] *************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77 Wednesday 10 July 2024 05:14:56 +0000 (0:00:00.020) 0:03:01.566 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_state != \"absent\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87 Wednesday 10 July 2024 05:14:56 +0000 (0:00:00.052) 0:03:01.619 ******** ok: [sut] => { "ansible_facts": { "__podman_images": [], "__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo.kube", "__podman_volumes": [] }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 6] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:103 Wednesday 10 July 2024 05:14:57 +0000 (0:00:00.048) 0:03:01.667 ******** ok: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Cleanup quadlets] ********************* task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:110 Wednesday 10 July 2024 05:14:57 +0000 (0:00:00.024) 0:03:01.692 ******** included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml for sut TASK [fedora.linux_system_roles.podman : Stat XDG_RUNTIME_DIR] ***************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:4 Wednesday 10 July 2024 05:14:57 +0000 (0:00:00.045) 0:03:01.738 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Stop and disable service] ************* task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:12 Wednesday 10 July 2024 05:14:57 +0000 (0:00:00.018) 0:03:01.756 ******** changed: [sut] => { "changed": true, "enabled": false, "failed_when_result": false, "name": "quadlet-demo.service", "state": "stopped", "status": { "AccessSELinuxContext": "system_u:object_r:systemd_unit_file_t:s0", "ActiveEnterTimestamp": "Wed 2024-07-10 05:14:43 UTC", "ActiveEnterTimestampMonotonic": "1003882093", "ActiveExitTimestampMonotonic": "0", "ActiveState": "active", "After": "-.mount quadlet-demo-mysql.service basic.target quadlet-demo-network.service sysinit.target systemd-journald.socket system.slice", "AllowIsolate": "no", "AssertResult": "yes", "AssertTimestamp": "Wed 2024-07-10 05:14:42 UTC", "AssertTimestampMonotonic": "1002988075", "Before": "shutdown.target multi-user.target", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "CPUAccounting": "yes", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "562856000", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "no", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore", "CleanResult": "success", "CollectMode": "inactive", "ConditionResult": "yes", "ConditionTimestamp": "Wed 2024-07-10 05:14:42 UTC", "ConditionTimestampMonotonic": "1002988072", "ConfigurationDirectoryMode": "0755", "Conflicts": "shutdown.target", "ControlGroup": "/system.slice/quadlet-demo.service", "ControlGroupId": "8519", "ControlPID": "0", "CoredumpFilter": "0x33", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "DefaultStartupMemoryLow": "0", "Delegate": "no", "Description": "quadlet-demo.service", "DevicePolicy": "auto", "DropInPaths": "/usr/lib/systemd/system/service.d/10-timeout-abort.conf", "DynamicUser": "no", "EffectiveCPUs": "0-1", "EffectiveMemoryNodes": "0", "Environment": "PODMAN_SYSTEMD_UNIT=quadlet-demo.service", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "9332", "ExecMainStartTimestamp": "Wed 2024-07-10 05:14:43 UTC", "ExecMainStartTimestampMonotonic": "1003881995", "ExecMainStatus": "0", "ExecStart": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman kube play --replace --service-container=true --network=systemd-quadlet-demo --configmap /etc/containers/systemd/envoy-proxy-configmap.yml --publish 8000:8080 --publish 9000:9901 /etc/containers/systemd/quadlet-demo.yml ; ignore_errors=no ; start_time=[Wed 2024-07-10 05:14:42 UTC] ; stop_time=[n/a] ; pid=9293 ; code=(null) ; status=0/0 }", "ExecStartEx": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman kube play --replace --service-container=true --network=systemd-quadlet-demo --configmap /etc/containers/systemd/envoy-proxy-configmap.yml --publish 8000:8080 --publish 9000:9901 /etc/containers/systemd/quadlet-demo.yml ; flags= ; start_time=[Wed 2024-07-10 05:14:42 UTC] ; stop_time=[n/a] ; pid=9293 ; code=(null) ; status=0/0 }", "ExecStopPost": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman kube down /etc/containers/systemd/quadlet-demo.yml ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStopPostEx": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman kube down /etc/containers/systemd/quadlet-demo.yml ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExitType": "main", "ExtensionImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FileDescriptorStorePreserve": "restart", "FinalKillSignal": "9", "FragmentPath": "/run/systemd/generator/quadlet-demo.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOReadBytes": "18446744073709551615", "IOReadOperations": "18446744073709551615", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "IOWeight": "[not set]", "IOWriteBytes": "18446744073709551615", "IOWriteOperations": "18446744073709551615", "IPAccounting": "no", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "Id": "quadlet-demo.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestamp": "Wed 2024-07-10 05:14:42 UTC", "InactiveExitTimestampMonotonic": "1002999439", "InvocationID": "d800043225e649d39861ccd65ac46ecd", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "mixed", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "infinity", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "8388608", "LimitMEMLOCKSoft": "8388608", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "524288", "LimitNOFILESoft": "1024", "LimitNPROC": "14756", "LimitNPROCSoft": "14756", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "14756", "LimitSIGPENDINGSoft": "14756", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "9332", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "ManagedOOMSwap": "auto", "MemoryAccounting": "yes", "MemoryAvailable": "infinity", "MemoryCurrent": "4710400", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryKSM": "no", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemoryPressureThresholdUSec": "200ms", "MemoryPressureWatch": "auto", "MemorySwapMax": "infinity", "MemoryZSwapMax": "infinity", "MountAPIVFS": "no", "MountImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAPolicy": "n/a", "Names": "quadlet-demo.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "all", "OOMPolicy": "stop", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "OnSuccessJobMode": "fail", "Perpetual": "no", "PrivateDevices": "no", "PrivateIPC": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProcSubset": "all", "ProtectClock": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectHostname": "no", "ProtectKernelLogs": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectProc": "default", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "ReloadResult": "success", "ReloadSignal": "1", "RemainAfterExit": "no", "RemoveIPC": "no", "Requires": "sysinit.target system.slice -.mount quadlet-demo-mysql.service quadlet-demo-network.service", "RequiresMountsFor": "/run/containers", "Restart": "no", "RestartKillSignal": "15", "RestartMaxDelayUSec": "infinity", "RestartMode": "normal", "RestartSteps": "0", "RestartUSec": "100ms", "RestartUSecNext": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RootEphemeral": "no", "RootImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "Slice": "system.slice", "SourcePath": "/etc/containers/systemd/quadlet-demo.kube", "StandardError": "inherit", "StandardInput": "null", "StandardOutput": "journal", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StartupMemoryHigh": "18446744073709551615", "StartupMemoryLow": "0", "StartupMemoryMax": "18446744073709551615", "StartupMemorySwapMax": "18446744073709551615", "StartupMemoryZSwapMax": "18446744073709551615", "StateChangeTimestamp": "Wed 2024-07-10 05:14:43 UTC", "StateChangeTimestampMonotonic": "1003882093", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "running", "SuccessAction": "none", "SyslogFacility": "3", "SyslogIdentifier": "quadlet-demo", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "2147483646", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "4", "TasksMax": "4426", "TimeoutAbortUSec": "45s", "TimeoutCleanUSec": "infinity", "TimeoutStartFailureMode": "terminate", "TimeoutStartUSec": "45s", "TimeoutStopFailureMode": "abort", "TimeoutStopUSec": "45s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "notify", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "disabled", "UnitFileState": "generated", "UtmpMode": "init", "WantedBy": "multi-user.target", "WatchdogSignal": "6", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "0" } } TASK [fedora.linux_system_roles.podman : Remove quadlet file] ****************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:33 Wednesday 10 July 2024 05:14:58 +0000 (0:00:01.416) 0:03:03.173 ******** changed: [sut] => { "changed": true, "path": "/etc/containers/systemd/quadlet-demo.kube", "state": "absent" } TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:39 Wednesday 10 July 2024 05:14:58 +0000 (0:00:00.243) 0:03:03.416 ******** included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for sut TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Wednesday 10 July 2024 05:14:58 +0000 (0:00:00.031) 0:03:03.448 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Wednesday 10 July 2024 05:14:58 +0000 (0:00:00.018) 0:03:03.467 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Wednesday 10 July 2024 05:14:58 +0000 (0:00:00.018) 0:03:03.486 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Reload systemctl] ********************* task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:47 Wednesday 10 July 2024 05:14:58 +0000 (0:00:00.018) 0:03:03.504 ******** ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.podman : Prune images no longer in use] ******** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:56 Wednesday 10 July 2024 05:14:59 +0000 (0:00:00.833) 0:03:04.338 ******** changed: [sut] => { "changed": true, "cmd": [ "podman", "image", "prune", "--all", "-f" ], "delta": "0:00:00.544309", "end": "2024-07-10 05:15:00.419831", "rc": 0, "start": "2024-07-10 05:14:59.875522" } STDOUT: fcf3e41b8864a14d75a6d0627d3d02154e28a153aa57e8baa392cd744ffa0d0b 5af2585e22ed1562885d9407efab74010090427be79048c2cd6a226517cc1e1d 6c68feb81ac7fb684f7b4d483de9bffaff110fa95445c248359b77bb5d3b2295 TASK [fedora.linux_system_roles.podman : Remove networks no longer in use] ***** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:64 Wednesday 10 July 2024 05:15:00 +0000 (0:00:00.781) 0:03:05.119 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_quadlet_type == \"network\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Remove volumes no longer in use] ****** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:73 Wednesday 10 July 2024 05:15:00 +0000 (0:00:00.020) 0:03:05.139 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_quadlet_type == \"volume\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : For testing and debugging - images] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:82 Wednesday 10 July 2024 05:15:00 +0000 (0:00:00.020) 0:03:05.160 ******** ok: [sut] => { "changed": false, "cmd": [ "podman", "images", "-n" ], "delta": "0:00:00.035798", "end": "2024-07-10 05:15:00.733039", "rc": 0, "start": "2024-07-10 05:15:00.697241" } STDOUT: quay.io/linux-system-roles/mysql 5.6 dd3b2a5dcb48 2 years ago 308 MB TASK [fedora.linux_system_roles.podman : For testing and debugging - volumes] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:92 Wednesday 10 July 2024 05:15:00 +0000 (0:00:00.271) 0:03:05.432 ******** ok: [sut] => { "changed": false, "cmd": [ "podman", "volume", "ls", "-n" ], "delta": "0:00:00.032788", "end": "2024-07-10 05:15:01.002321", "rc": 0, "start": "2024-07-10 05:15:00.969533" } STDOUT: local systemd-quadlet-demo-mysql local wp-pv-claim local envoy-proxy-config local envoy-certificates TASK [fedora.linux_system_roles.podman : For testing and debugging - containers] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:102 Wednesday 10 July 2024 05:15:01 +0000 (0:00:00.305) 0:03:05.738 ******** ok: [sut] => { "changed": false, "cmd": [ "podman", "ps", "--noheading" ], "delta": "0:00:00.039909", "end": "2024-07-10 05:15:01.317214", "rc": 0, "start": "2024-07-10 05:15:01.277305" } STDOUT: 21b1326f5785 quay.io/linux-system-roles/mysql:5.6 mysqld 48 seconds ago Up 48 seconds (starting) quadlet-demo-mysql TASK [fedora.linux_system_roles.podman : For testing and debugging - networks] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:112 Wednesday 10 July 2024 05:15:01 +0000 (0:00:00.278) 0:03:06.016 ******** ok: [sut] => { "changed": false, "cmd": [ "podman", "network", "ls", "-n", "-q" ], "delta": "0:00:00.031987", "end": "2024-07-10 05:15:01.587427", "rc": 0, "start": "2024-07-10 05:15:01.555440" } STDOUT: podman systemd-quadlet-demo TASK [fedora.linux_system_roles.podman : For testing and debugging - secrets] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:122 Wednesday 10 July 2024 05:15:01 +0000 (0:00:00.270) 0:03:06.287 ******** ok: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : For testing and debugging - services] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:133 Wednesday 10 July 2024 05:15:01 +0000 (0:00:00.273) 0:03:06.561 ******** ok: [sut] => { "ansible_facts": { "services": { "21b1326f57855b147efcb07cbd10c5f7b7edbabb6693b7752105182196a3f742.service": { "name": "21b1326f57855b147efcb07cbd10c5f7b7edbabb6693b7752105182196a3f742.service", "source": "systemd", "state": "stopped", "status": "failed" }, "NetworkManager-dispatcher.service": { "name": "NetworkManager-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "NetworkManager-wait-online.service": { "name": "NetworkManager-wait-online.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "NetworkManager.service": { "name": "NetworkManager.service", "source": "systemd", "state": "running", "status": "enabled" }, "arp-ethers.service": { "name": "arp-ethers.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "auditd.service": { "name": "auditd.service", "source": "systemd", "state": "running", "status": "enabled" }, "auth-rpcgss-module.service": { "name": "auth-rpcgss-module.service", "source": "systemd", "state": "stopped", "status": "static" }, "autovt@.service": { "name": "autovt@.service", "source": "systemd", "state": "unknown", "status": "alias" }, "blk-availability.service": { "name": "blk-availability.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "bluetooth.service": { "name": "bluetooth.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "certmonger.service": { "name": "certmonger.service", "source": "systemd", "state": "running", "status": "enabled" }, "chrony-wait.service": { "name": "chrony-wait.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd-restricted.service": { "name": "chronyd-restricted.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd.service": { "name": "chronyd.service", "source": "systemd", "state": "running", "status": "enabled" }, "cloud-config.service": { "name": "cloud-config.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-final.service": { "name": "cloud-final.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init-hotplugd.service": { "name": "cloud-init-hotplugd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cloud-init-local.service": { "name": "cloud-init-local.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init.service": { "name": "cloud-init.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "console-getty.service": { "name": "console-getty.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "container-getty@.service": { "name": "container-getty@.service", "source": "systemd", "state": "unknown", "status": "static" }, "dbus-broker.service": { "name": "dbus-broker.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-org.bluez.service": { "name": "dbus-org.bluez.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.fedoraproject.FirewallD1.service": { "name": "dbus-org.fedoraproject.FirewallD1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.home1.service": { "name": "dbus-org.freedesktop.home1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.hostname1.service": { "name": "dbus-org.freedesktop.hostname1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.locale1.service": { "name": "dbus-org.freedesktop.locale1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.login1.service": { "name": "dbus-org.freedesktop.login1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.nm-dispatcher.service": { "name": "dbus-org.freedesktop.nm-dispatcher.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.oom1.service": { "name": "dbus-org.freedesktop.oom1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.portable1.service": { "name": "dbus-org.freedesktop.portable1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.resolve1.service": { "name": "dbus-org.freedesktop.resolve1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.timedate1.service": { "name": "dbus-org.freedesktop.timedate1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus.service": { "name": "dbus.service", "source": "systemd", "state": "active", "status": "alias" }, "debug-shell.service": { "name": "debug-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "display-manager.service": { "name": "display-manager.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "dm-event.service": { "name": "dm-event.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-makecache.service": { "name": "dnf-makecache.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-system-upgrade-cleanup.service": { "name": "dnf-system-upgrade-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "dnf-system-upgrade.service": { "name": "dnf-system-upgrade.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dracut-cmdline.service": { "name": "dracut-cmdline.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-initqueue.service": { "name": "dracut-initqueue.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-mount.service": { "name": "dracut-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-mount.service": { "name": "dracut-pre-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-pivot.service": { "name": "dracut-pre-pivot.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-trigger.service": { "name": "dracut-pre-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-udev.service": { "name": "dracut-pre-udev.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown-onfailure.service": { "name": "dracut-shutdown-onfailure.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown.service": { "name": "dracut-shutdown.service", "source": "systemd", "state": "stopped", "status": "static" }, "ebtables.service": { "name": "ebtables.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "emergency.service": { "name": "emergency.service", "source": "systemd", "state": "stopped", "status": "static" }, "firewalld.service": { "name": "firewalld.service", "source": "systemd", "state": "running", "status": "enabled" }, "fsidd.service": { "name": "fsidd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "fstrim.service": { "name": "fstrim.service", "source": "systemd", "state": "stopped", "status": "static" }, "fwupd-offline-update.service": { "name": "fwupd-offline-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd-refresh.service": { "name": "fwupd-refresh.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd.service": { "name": "fwupd.service", "source": "systemd", "state": "inactive", "status": "static" }, "getty@.service": { "name": "getty@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "getty@tty1.service": { "name": "getty@tty1.service", "source": "systemd", "state": "running", "status": "active" }, "grub-boot-indeterminate.service": { "name": "grub-boot-indeterminate.service", "source": "systemd", "state": "inactive", "status": "static" }, "grub2-systemd-integration.service": { "name": "grub2-systemd-integration.service", "source": "systemd", "state": "inactive", "status": "static" }, "gssproxy.service": { "name": "gssproxy.service", "source": "systemd", "state": "running", "status": "disabled" }, "hv_kvp_daemon.service": { "name": "hv_kvp_daemon.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "import-state.service": { "name": "import-state.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "initrd-cleanup.service": { "name": "initrd-cleanup.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-parse-etc.service": { "name": "initrd-parse-etc.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-switch-root.service": { "name": "initrd-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-udevadm-cleanup-db.service": { "name": "initrd-udevadm-cleanup-db.service", "source": "systemd", "state": "stopped", "status": "static" }, "ip6tables.service": { "name": "ip6tables.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ipset.service": { "name": "ipset.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "iptables.service": { "name": "iptables.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "kmod-static-nodes.service": { "name": "kmod-static-nodes.service", "source": "systemd", "state": "stopped", "status": "static" }, "ldconfig.service": { "name": "ldconfig.service", "source": "systemd", "state": "stopped", "status": "static" }, "loadmodules.service": { "name": "loadmodules.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "lvm2-lvmpolld.service": { "name": "lvm2-lvmpolld.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-monitor.service": { "name": "lvm2-monitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "man-db-cache-update.service": { "name": "man-db-cache-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "man-db-restart-cache-update.service": { "name": "man-db-restart-cache-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "mdadm-grow-continue@.service": { "name": "mdadm-grow-continue@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdadm-last-resort@.service": { "name": "mdadm-last-resort@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdcheck_continue.service": { "name": "mdcheck_continue.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdcheck_start.service": { "name": "mdcheck_start.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmon@.service": { "name": "mdmon@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdmonitor-oneshot.service": { "name": "mdmonitor-oneshot.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmonitor.service": { "name": "mdmonitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "modprobe@.service": { "name": "modprobe@.service", "source": "systemd", "state": "unknown", "status": "static" }, "modprobe@configfs.service": { "name": "modprobe@configfs.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "modprobe@dm_mod.service": { "name": "modprobe@dm_mod.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "modprobe@drm.service": { "name": "modprobe@drm.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "modprobe@efi_pstore.service": { "name": "modprobe@efi_pstore.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "modprobe@fuse.service": { "name": "modprobe@fuse.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "modprobe@loop.service": { "name": "modprobe@loop.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "netavark-dhcp-proxy.service": { "name": "netavark-dhcp-proxy.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "netavark-firewalld-reload.service": { "name": "netavark-firewalld-reload.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "network.service": { "name": "network.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "nfs-blkmap.service": { "name": "nfs-blkmap.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-idmapd.service": { "name": "nfs-idmapd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-mountd.service": { "name": "nfs-mountd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-server.service": { "name": "nfs-server.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-utils.service": { "name": "nfs-utils.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfsdcld.service": { "name": "nfsdcld.service", "source": "systemd", "state": "stopped", "status": "static" }, "nftables.service": { "name": "nftables.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nis-domainname.service": { "name": "nis-domainname.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nm-priv-helper.service": { "name": "nm-priv-helper.service", "source": "systemd", "state": "inactive", "status": "static" }, "ntpd.service": { "name": "ntpd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ntpdate.service": { "name": "ntpdate.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "pam_namespace.service": { "name": "pam_namespace.service", "source": "systemd", "state": "inactive", "status": "static" }, "pcscd.service": { "name": "pcscd.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "plymouth-halt.service": { "name": "plymouth-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-kexec.service": { "name": "plymouth-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-poweroff.service": { "name": "plymouth-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-quit-wait.service": { "name": "plymouth-quit-wait.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-quit.service": { "name": "plymouth-quit.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-read-write.service": { "name": "plymouth-read-write.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-reboot.service": { "name": "plymouth-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-start.service": { "name": "plymouth-start.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-switch-root-initramfs.service": { "name": "plymouth-switch-root-initramfs.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-switch-root.service": { "name": "plymouth-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "podman-auto-update.service": { "name": "podman-auto-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "podman-clean-transient.service": { "name": "podman-clean-transient.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "podman-kube@.service": { "name": "podman-kube@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "podman-restart.service": { "name": "podman-restart.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "podman.service": { "name": "podman.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "polkit.service": { "name": "polkit.service", "source": "systemd", "state": "running", "status": "static" }, "quadlet-demo-mysql-volume.service": { "name": "quadlet-demo-mysql-volume.service", "source": "systemd", "state": "stopped", "status": "generated" }, "quadlet-demo-mysql.service": { "name": "quadlet-demo-mysql.service", "source": "systemd", "state": "running", "status": "generated" }, "quadlet-demo-network.service": { "name": "quadlet-demo-network.service", "source": "systemd", "state": "stopped", "status": "generated" }, "quotaon.service": { "name": "quotaon.service", "source": "systemd", "state": "inactive", "status": "static" }, "raid-check.service": { "name": "raid-check.service", "source": "systemd", "state": "stopped", "status": "static" }, "rc-local.service": { "name": "rc-local.service", "source": "systemd", "state": "stopped", "status": "static" }, "rescue.service": { "name": "rescue.service", "source": "systemd", "state": "stopped", "status": "static" }, "restraintd.service": { "name": "restraintd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rngd.service": { "name": "rngd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpc-gssd.service": { "name": "rpc-gssd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd-notify.service": { "name": "rpc-statd-notify.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd.service": { "name": "rpc-statd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-svcgssd.service": { "name": "rpc-svcgssd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "rpcbind.service": { "name": "rpcbind.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "rpmdb-migrate.service": { "name": "rpmdb-migrate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rpmdb-rebuild.service": { "name": "rpmdb-rebuild.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "selinux-autorelabel-mark.service": { "name": "selinux-autorelabel-mark.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "selinux-autorelabel.service": { "name": "selinux-autorelabel.service", "source": "systemd", "state": "inactive", "status": "static" }, "selinux-check-proper-disable.service": { "name": "selinux-check-proper-disable.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "serial-getty@.service": { "name": "serial-getty@.service", "source": "systemd", "state": "unknown", "status": "indirect" }, "serial-getty@ttyS0.service": { "name": "serial-getty@ttyS0.service", "source": "systemd", "state": "running", "status": "active" }, "snapd.seeded.service": { "name": "snapd.seeded.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sntp.service": { "name": "sntp.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ssh-host-keys-migration.service": { "name": "ssh-host-keys-migration.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "sshd-keygen.service": { "name": "sshd-keygen.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sshd-keygen@.service": { "name": "sshd-keygen@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@ecdsa.service": { "name": "sshd-keygen@ecdsa.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd-keygen@ed25519.service": { "name": "sshd-keygen@ed25519.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd-keygen@rsa.service": { "name": "sshd-keygen@rsa.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd.service": { "name": "sshd.service", "source": "systemd", "state": "running", "status": "enabled" }, "sshd@.service": { "name": "sshd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "sssd-autofs.service": { "name": "sssd-autofs.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-kcm.service": { "name": "sssd-kcm.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "sssd-nss.service": { "name": "sssd-nss.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pac.service": { "name": "sssd-pac.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pam.service": { "name": "sssd-pam.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-ssh.service": { "name": "sssd-ssh.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-sudo.service": { "name": "sssd-sudo.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd.service": { "name": "sssd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "syslog.service": { "name": "syslog.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "system-update-cleanup.service": { "name": "system-update-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-ask-password-console.service": { "name": "systemd-ask-password-console.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-plymouth.service": { "name": "systemd-ask-password-plymouth.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-wall.service": { "name": "systemd-ask-password-wall.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-backlight@.service": { "name": "systemd-backlight@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-battery-check.service": { "name": "systemd-battery-check.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-binfmt.service": { "name": "systemd-binfmt.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-bless-boot.service": { "name": "systemd-bless-boot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-boot-check-no-failures.service": { "name": "systemd-boot-check-no-failures.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-boot-random-seed.service": { "name": "systemd-boot-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-boot-update.service": { "name": "systemd-boot-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-confext.service": { "name": "systemd-confext.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-coredump@.service": { "name": "systemd-coredump@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-exit.service": { "name": "systemd-exit.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-firstboot.service": { "name": "systemd-firstboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck-root.service": { "name": "systemd-fsck-root.service", "source": "systemd", "state": "stopped", "status": "enabled-runtime" }, "systemd-fsck@.service": { "name": "systemd-fsck@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-growfs-root.service": { "name": "systemd-growfs-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-growfs@.service": { "name": "systemd-growfs@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-halt.service": { "name": "systemd-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hibernate.service": { "name": "systemd-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-homed-activate.service": { "name": "systemd-homed-activate.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-homed.service": { "name": "systemd-homed.service", "source": "systemd", "state": "running", "status": "enabled" }, "systemd-hostnamed.service": { "name": "systemd-hostnamed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hwdb-update.service": { "name": "systemd-hwdb-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-hybrid-sleep.service": { "name": "systemd-hybrid-sleep.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-initctl.service": { "name": "systemd-initctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-catalog-update.service": { "name": "systemd-journal-catalog-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-flush.service": { "name": "systemd-journal-flush.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journald.service": { "name": "systemd-journald.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-journald@.service": { "name": "systemd-journald@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-kexec.service": { "name": "systemd-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-localed.service": { "name": "systemd-localed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-logind.service": { "name": "systemd-logind.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-machine-id-commit.service": { "name": "systemd-machine-id-commit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-modules-load.service": { "name": "systemd-modules-load.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-network-generator.service": { "name": "systemd-network-generator.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-networkd-wait-online.service": { "name": "systemd-networkd-wait-online.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "systemd-networkd-wait-online@.service": { "name": "systemd-networkd-wait-online@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "systemd-networkd.service": { "name": "systemd-networkd.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "systemd-oomd.service": { "name": "systemd-oomd.service", "source": "systemd", "state": "running", "status": "enabled" }, "systemd-pcrfs-root.service": { "name": "systemd-pcrfs-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pcrfs@.service": { "name": "systemd-pcrfs@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-pcrmachine.service": { "name": "systemd-pcrmachine.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-pcrphase-initrd.service": { "name": "systemd-pcrphase-initrd.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-pcrphase-sysinit.service": { "name": "systemd-pcrphase-sysinit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-pcrphase.service": { "name": "systemd-pcrphase.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-portabled.service": { "name": "systemd-portabled.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-poweroff.service": { "name": "systemd-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pstore.service": { "name": "systemd-pstore.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-quotacheck.service": { "name": "systemd-quotacheck.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-random-seed.service": { "name": "systemd-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-reboot.service": { "name": "systemd-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-remount-fs.service": { "name": "systemd-remount-fs.service", "source": "systemd", "state": "stopped", "status": "enabled-runtime" }, "systemd-repart.service": { "name": "systemd-repart.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-resolved.service": { "name": "systemd-resolved.service", "source": "systemd", "state": "running", "status": "enabled" }, "systemd-rfkill.service": { "name": "systemd-rfkill.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-soft-reboot.service": { "name": "systemd-soft-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend-then-hibernate.service": { "name": "systemd-suspend-then-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend.service": { "name": "systemd-suspend.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-sysctl.service": { "name": "systemd-sysctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-sysext.service": { "name": "systemd-sysext.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "systemd-sysupdate-reboot.service": { "name": "systemd-sysupdate-reboot.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "systemd-sysupdate.service": { "name": "systemd-sysupdate.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "systemd-sysusers.service": { "name": "systemd-sysusers.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-time-wait-sync.service": { "name": "systemd-time-wait-sync.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-timedated.service": { "name": "systemd-timedated.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-timesyncd.service": { "name": "systemd-timesyncd.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "systemd-tmpfiles-clean.service": { "name": "systemd-tmpfiles-clean.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev-early.service": { "name": "systemd-tmpfiles-setup-dev-early.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev.service": { "name": "systemd-tmpfiles-setup-dev.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup.service": { "name": "systemd-tmpfiles-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-settle.service": { "name": "systemd-udev-settle.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-trigger.service": { "name": "systemd-udev-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udevd.service": { "name": "systemd-udevd.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-update-done.service": { "name": "systemd-update-done.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp-runlevel.service": { "name": "systemd-update-utmp-runlevel.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp.service": { "name": "systemd-update-utmp.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-user-sessions.service": { "name": "systemd-user-sessions.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-userdbd.service": { "name": "systemd-userdbd.service", "source": "systemd", "state": "running", "status": "indirect" }, "systemd-vconsole-setup.service": { "name": "systemd-vconsole-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-volatile-root.service": { "name": "systemd-volatile-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-zram-setup@.service": { "name": "systemd-zram-setup@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-zram-setup@zram0.service": { "name": "systemd-zram-setup@zram0.service", "source": "systemd", "state": "stopped", "status": "active" }, "udisks2.service": { "name": "udisks2.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "unbound-anchor.service": { "name": "unbound-anchor.service", "source": "systemd", "state": "stopped", "status": "static" }, "user-runtime-dir@.service": { "name": "user-runtime-dir@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user-runtime-dir@0.service": { "name": "user-runtime-dir@0.service", "source": "systemd", "state": "stopped", "status": "active" }, "user@.service": { "name": "user@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user@0.service": { "name": "user@0.service", "source": "systemd", "state": "running", "status": "active" } } }, "changed": false } TASK [fedora.linux_system_roles.podman : Create and update quadlets] *********** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:114 Wednesday 10 July 2024 05:15:04 +0000 (0:00:02.304) 0:03:08.866 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_state != \"absent\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14 Wednesday 10 July 2024 05:15:04 +0000 (0:00:00.021) 0:03:08.887 ******** ok: [sut] => { "ansible_facts": { "__podman_quadlet_file_src": "", "__podman_quadlet_spec": {}, "__podman_quadlet_str": "---\napiVersion: v1\nkind: PersistentVolumeClaim\nmetadata:\n name: wp-pv-claim\n labels:\n app: wordpress\nspec:\n accessModes:\n - ReadWriteOnce\n resources:\n requests:\n storage: 20Gi\n---\napiVersion: v1\nkind: Pod\nmetadata:\n name: quadlet-demo\nspec:\n containers:\n - name: wordpress\n image: quay.io/linux-system-roles/wordpress:4.8-apache\n env:\n - name: WORDPRESS_DB_HOST\n value: quadlet-demo-mysql\n - name: WORDPRESS_DB_PASSWORD\n valueFrom:\n secretKeyRef:\n name: mysql-root-password-kube\n key: password\n volumeMounts:\n - name: wordpress-persistent-storage\n mountPath: /var/www/html\n resources:\n requests:\n memory: \"64Mi\"\n cpu: \"250m\"\n limits:\n memory: \"128Mi\"\n cpu: \"500m\"\n - name: envoy\n image: quay.io/linux-system-roles/envoyproxy:v1.25.0\n volumeMounts:\n - name: config-volume\n mountPath: /etc/envoy\n - name: certificates\n mountPath: /etc/envoy-certificates\n env:\n - name: ENVOY_UID\n value: \"0\"\n resources:\n requests:\n memory: \"64Mi\"\n cpu: \"250m\"\n limits:\n memory: \"128Mi\"\n cpu: \"500m\"\n volumes:\n - name: config-volume\n configMap:\n name: envoy-proxy-config\n - name: certificates\n secret:\n secretName: envoy-certificates\n - name: wordpress-persistent-storage\n persistentVolumeClaim:\n claimName: wp-pv-claim\n - name: www # not used - for testing hostpath\n hostPath:\n path: /tmp/httpd3\n - name: create # not used - for testing hostpath\n hostPath:\n path: /tmp/httpd3-create\n", "__podman_quadlet_template_src": "quadlet-demo.yml.j2" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25 Wednesday 10 July 2024 05:15:04 +0000 (0:00:00.073) 0:03:08.961 ******** ok: [sut] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_pull_image": true, "__podman_state": "absent", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] ***** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35 Wednesday 10 July 2024 05:15:04 +0000 (0:00:00.028) 0:03:08.989 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_quadlet_str", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48 Wednesday 10 July 2024 05:15:04 +0000 (0:00:00.023) 0:03:09.012 ******** ok: [sut] => { "ansible_facts": { "__podman_quadlet_name": "quadlet-demo", "__podman_quadlet_type": "yml", "__podman_rootless": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57 Wednesday 10 July 2024 05:15:04 +0000 (0:00:00.032) 0:03:09.045 ******** included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for sut TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Wednesday 10 July 2024 05:15:04 +0000 (0:00:00.040) 0:03:09.085 ******** skipping: [sut] => { "changed": false, "false_condition": "'getent_passwd' not in ansible_facts or __podman_user not in ansible_facts['getent_passwd']", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Wednesday 10 July 2024 05:15:04 +0000 (0:00:00.025) 0:03:09.111 ******** skipping: [sut] => { "changed": false, "false_condition": "not ansible_facts[\"getent_passwd\"][__podman_user]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Wednesday 10 July 2024 05:15:04 +0000 (0:00:00.025) 0:03:09.136 ******** ok: [sut] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get group information] **************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:28 Wednesday 10 July 2024 05:15:04 +0000 (0:00:00.030) 0:03:09.166 ******** ok: [sut] => { "ansible_facts": { "getent_group": { "root": [ "x", "0", "" ] } }, "changed": false } TASK [fedora.linux_system_roles.podman : Set group name] *********************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:35 Wednesday 10 July 2024 05:15:04 +0000 (0:00:00.323) 0:03:09.490 ******** ok: [sut] => { "ansible_facts": { "__podman_group_name": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:39 Wednesday 10 July 2024 05:15:04 +0000 (0:00:00.031) 0:03:09.522 ******** ok: [sut] => { "changed": false, "stat": { "atime": 1720588427.5164855, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "72ba28a3be19c467e89ab90f5d21f99b605159a7", "ctime": 1720588377.3007717, "dev": 51714, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 203603, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-pie-executable", "mode": "0755", "mtime": 1696291200.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 16032, "uid": 0, "version": "2092964490", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************ task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:50 Wednesday 10 July 2024 05:15:05 +0000 (0:00:00.240) 0:03:09.763 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check group with getsubids] *********** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:55 Wednesday 10 July 2024 05:15:05 +0000 (0:00:00.020) 0:03:09.783 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:60 Wednesday 10 July 2024 05:15:05 +0000 (0:00:00.020) 0:03:09.804 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subuid file] ********************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:74 Wednesday 10 July 2024 05:15:05 +0000 (0:00:00.020) 0:03:09.825 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subgid file] ********************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:79 Wednesday 10 July 2024 05:15:05 +0000 (0:00:00.020) 0:03:09.846 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:84 Wednesday 10 July 2024 05:15:05 +0000 (0:00:00.019) 0:03:09.865 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:94 Wednesday 10 July 2024 05:15:05 +0000 (0:00:00.020) 0:03:09.886 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] ***** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:101 Wednesday 10 July 2024 05:15:05 +0000 (0:00:00.020) 0:03:09.906 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62 Wednesday 10 July 2024 05:15:05 +0000 (0:00:00.020) 0:03:09.927 ******** ok: [sut] => { "ansible_facts": { "__podman_activate_systemd_unit": true, "__podman_images_found": [], "__podman_kube_yamls_raw": "", "__podman_service_name": "", "__podman_systemd_scope": "system", "__podman_user_home_dir": "/root", "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73 Wednesday 10 July 2024 05:15:05 +0000 (0:00:00.037) 0:03:09.964 ******** ok: [sut] => { "ansible_facts": { "__podman_quadlet_path": "/etc/containers/systemd" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get kube yaml contents] *************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77 Wednesday 10 July 2024 05:15:05 +0000 (0:00:00.022) 0:03:09.987 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_state != \"absent\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87 Wednesday 10 July 2024 05:15:05 +0000 (0:00:00.020) 0:03:10.007 ******** ok: [sut] => { "ansible_facts": { "__podman_images": [], "__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo.yml", "__podman_volumes": [] }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 6] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:103 Wednesday 10 July 2024 05:15:05 +0000 (0:00:00.087) 0:03:10.094 ******** ok: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Cleanup quadlets] ********************* task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:110 Wednesday 10 July 2024 05:15:05 +0000 (0:00:00.028) 0:03:10.123 ******** included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml for sut TASK [fedora.linux_system_roles.podman : Stat XDG_RUNTIME_DIR] ***************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:4 Wednesday 10 July 2024 05:15:05 +0000 (0:00:00.048) 0:03:10.171 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Stop and disable service] ************* task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:12 Wednesday 10 July 2024 05:15:05 +0000 (0:00:00.021) 0:03:10.193 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_service_name | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Remove quadlet file] ****************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:33 Wednesday 10 July 2024 05:15:05 +0000 (0:00:00.023) 0:03:10.216 ******** changed: [sut] => { "changed": true, "path": "/etc/containers/systemd/quadlet-demo.yml", "state": "absent" } TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:39 Wednesday 10 July 2024 05:15:05 +0000 (0:00:00.240) 0:03:10.456 ******** included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for sut TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Wednesday 10 July 2024 05:15:05 +0000 (0:00:00.035) 0:03:10.492 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Wednesday 10 July 2024 05:15:05 +0000 (0:00:00.021) 0:03:10.513 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Wednesday 10 July 2024 05:15:05 +0000 (0:00:00.020) 0:03:10.533 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Reload systemctl] ********************* task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:47 Wednesday 10 July 2024 05:15:05 +0000 (0:00:00.020) 0:03:10.553 ******** ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.podman : Prune images no longer in use] ******** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:56 Wednesday 10 July 2024 05:15:06 +0000 (0:00:00.825) 0:03:11.379 ******** changed: [sut] => { "changed": true, "cmd": [ "podman", "image", "prune", "--all", "-f" ], "delta": "0:00:00.032062", "end": "2024-07-10 05:15:06.947090", "rc": 0, "start": "2024-07-10 05:15:06.915028" } TASK [fedora.linux_system_roles.podman : Remove networks no longer in use] ***** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:64 Wednesday 10 July 2024 05:15:06 +0000 (0:00:00.270) 0:03:11.649 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_quadlet_type == \"network\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Remove volumes no longer in use] ****** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:73 Wednesday 10 July 2024 05:15:07 +0000 (0:00:00.022) 0:03:11.671 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_quadlet_type == \"volume\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : For testing and debugging - images] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:82 Wednesday 10 July 2024 05:15:07 +0000 (0:00:00.060) 0:03:11.731 ******** ok: [sut] => { "changed": false, "cmd": [ "podman", "images", "-n" ], "delta": "0:00:00.034605", "end": "2024-07-10 05:15:07.306336", "rc": 0, "start": "2024-07-10 05:15:07.271731" } STDOUT: quay.io/linux-system-roles/mysql 5.6 dd3b2a5dcb48 2 years ago 308 MB TASK [fedora.linux_system_roles.podman : For testing and debugging - volumes] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:92 Wednesday 10 July 2024 05:15:07 +0000 (0:00:00.278) 0:03:12.010 ******** ok: [sut] => { "changed": false, "cmd": [ "podman", "volume", "ls", "-n" ], "delta": "0:00:00.035355", "end": "2024-07-10 05:15:07.582134", "rc": 0, "start": "2024-07-10 05:15:07.546779" } STDOUT: local systemd-quadlet-demo-mysql local wp-pv-claim local envoy-proxy-config local envoy-certificates TASK [fedora.linux_system_roles.podman : For testing and debugging - containers] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:102 Wednesday 10 July 2024 05:15:07 +0000 (0:00:00.273) 0:03:12.284 ******** ok: [sut] => { "changed": false, "cmd": [ "podman", "ps", "--noheading" ], "delta": "0:00:00.039611", "end": "2024-07-10 05:15:07.861356", "rc": 0, "start": "2024-07-10 05:15:07.821745" } STDOUT: 21b1326f5785 quay.io/linux-system-roles/mysql:5.6 mysqld 55 seconds ago Up 54 seconds (starting) quadlet-demo-mysql TASK [fedora.linux_system_roles.podman : For testing and debugging - networks] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:112 Wednesday 10 July 2024 05:15:07 +0000 (0:00:00.279) 0:03:12.563 ******** ok: [sut] => { "changed": false, "cmd": [ "podman", "network", "ls", "-n", "-q" ], "delta": "0:00:00.032467", "end": "2024-07-10 05:15:08.134329", "rc": 0, "start": "2024-07-10 05:15:08.101862" } STDOUT: podman systemd-quadlet-demo TASK [fedora.linux_system_roles.podman : For testing and debugging - secrets] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:122 Wednesday 10 July 2024 05:15:08 +0000 (0:00:00.274) 0:03:12.838 ******** ok: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : For testing and debugging - services] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:133 Wednesday 10 July 2024 05:15:08 +0000 (0:00:00.271) 0:03:13.110 ******** ok: [sut] => { "ansible_facts": { "services": { "21b1326f57855b147efcb07cbd10c5f7b7edbabb6693b7752105182196a3f742.service": { "name": "21b1326f57855b147efcb07cbd10c5f7b7edbabb6693b7752105182196a3f742.service", "source": "systemd", "state": "stopped", "status": "failed" }, "NetworkManager-dispatcher.service": { "name": "NetworkManager-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "NetworkManager-wait-online.service": { "name": "NetworkManager-wait-online.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "NetworkManager.service": { "name": "NetworkManager.service", "source": "systemd", "state": "running", "status": "enabled" }, "arp-ethers.service": { "name": "arp-ethers.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "auditd.service": { "name": "auditd.service", "source": "systemd", "state": "running", "status": "enabled" }, "auth-rpcgss-module.service": { "name": "auth-rpcgss-module.service", "source": "systemd", "state": "stopped", "status": "static" }, "autovt@.service": { "name": "autovt@.service", "source": "systemd", "state": "unknown", "status": "alias" }, "blk-availability.service": { "name": "blk-availability.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "bluetooth.service": { "name": "bluetooth.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "certmonger.service": { "name": "certmonger.service", "source": "systemd", "state": "running", "status": "enabled" }, "chrony-wait.service": { "name": "chrony-wait.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd-restricted.service": { "name": "chronyd-restricted.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd.service": { "name": "chronyd.service", "source": "systemd", "state": "running", "status": "enabled" }, "cloud-config.service": { "name": "cloud-config.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-final.service": { "name": "cloud-final.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init-hotplugd.service": { "name": "cloud-init-hotplugd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cloud-init-local.service": { "name": "cloud-init-local.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init.service": { "name": "cloud-init.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "console-getty.service": { "name": "console-getty.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "container-getty@.service": { "name": "container-getty@.service", "source": "systemd", "state": "unknown", "status": "static" }, "dbus-broker.service": { "name": "dbus-broker.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-org.bluez.service": { "name": "dbus-org.bluez.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.fedoraproject.FirewallD1.service": { "name": "dbus-org.fedoraproject.FirewallD1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.home1.service": { "name": "dbus-org.freedesktop.home1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.hostname1.service": { "name": "dbus-org.freedesktop.hostname1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.locale1.service": { "name": "dbus-org.freedesktop.locale1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.login1.service": { "name": "dbus-org.freedesktop.login1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.nm-dispatcher.service": { "name": "dbus-org.freedesktop.nm-dispatcher.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.oom1.service": { "name": "dbus-org.freedesktop.oom1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.portable1.service": { "name": "dbus-org.freedesktop.portable1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.resolve1.service": { "name": "dbus-org.freedesktop.resolve1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.timedate1.service": { "name": "dbus-org.freedesktop.timedate1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus.service": { "name": "dbus.service", "source": "systemd", "state": "active", "status": "alias" }, "debug-shell.service": { "name": "debug-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "display-manager.service": { "name": "display-manager.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "dm-event.service": { "name": "dm-event.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-makecache.service": { "name": "dnf-makecache.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-system-upgrade-cleanup.service": { "name": "dnf-system-upgrade-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "dnf-system-upgrade.service": { "name": "dnf-system-upgrade.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dracut-cmdline.service": { "name": "dracut-cmdline.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-initqueue.service": { "name": "dracut-initqueue.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-mount.service": { "name": "dracut-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-mount.service": { "name": "dracut-pre-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-pivot.service": { "name": "dracut-pre-pivot.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-trigger.service": { "name": "dracut-pre-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-udev.service": { "name": "dracut-pre-udev.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown-onfailure.service": { "name": "dracut-shutdown-onfailure.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown.service": { "name": "dracut-shutdown.service", "source": "systemd", "state": "stopped", "status": "static" }, "ebtables.service": { "name": "ebtables.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "emergency.service": { "name": "emergency.service", "source": "systemd", "state": "stopped", "status": "static" }, "firewalld.service": { "name": "firewalld.service", "source": "systemd", "state": "running", "status": "enabled" }, "fsidd.service": { "name": "fsidd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "fstrim.service": { "name": "fstrim.service", "source": "systemd", "state": "stopped", "status": "static" }, "fwupd-offline-update.service": { "name": "fwupd-offline-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd-refresh.service": { "name": "fwupd-refresh.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd.service": { "name": "fwupd.service", "source": "systemd", "state": "inactive", "status": "static" }, "getty@.service": { "name": "getty@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "getty@tty1.service": { "name": "getty@tty1.service", "source": "systemd", "state": "running", "status": "active" }, "grub-boot-indeterminate.service": { "name": "grub-boot-indeterminate.service", "source": "systemd", "state": "inactive", "status": "static" }, "grub2-systemd-integration.service": { "name": "grub2-systemd-integration.service", "source": "systemd", "state": "inactive", "status": "static" }, "gssproxy.service": { "name": "gssproxy.service", "source": "systemd", "state": "running", "status": "disabled" }, "hv_kvp_daemon.service": { "name": "hv_kvp_daemon.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "import-state.service": { "name": "import-state.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "initrd-cleanup.service": { "name": "initrd-cleanup.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-parse-etc.service": { "name": "initrd-parse-etc.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-switch-root.service": { "name": "initrd-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-udevadm-cleanup-db.service": { "name": "initrd-udevadm-cleanup-db.service", "source": "systemd", "state": "stopped", "status": "static" }, "ip6tables.service": { "name": "ip6tables.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ipset.service": { "name": "ipset.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "iptables.service": { "name": "iptables.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "kmod-static-nodes.service": { "name": "kmod-static-nodes.service", "source": "systemd", "state": "stopped", "status": "static" }, "ldconfig.service": { "name": "ldconfig.service", "source": "systemd", "state": "stopped", "status": "static" }, "loadmodules.service": { "name": "loadmodules.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "lvm2-lvmpolld.service": { "name": "lvm2-lvmpolld.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-monitor.service": { "name": "lvm2-monitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "man-db-cache-update.service": { "name": "man-db-cache-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "man-db-restart-cache-update.service": { "name": "man-db-restart-cache-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "mdadm-grow-continue@.service": { "name": "mdadm-grow-continue@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdadm-last-resort@.service": { "name": "mdadm-last-resort@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdcheck_continue.service": { "name": "mdcheck_continue.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdcheck_start.service": { "name": "mdcheck_start.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmon@.service": { "name": "mdmon@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdmonitor-oneshot.service": { "name": "mdmonitor-oneshot.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmonitor.service": { "name": "mdmonitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "modprobe@.service": { "name": "modprobe@.service", "source": "systemd", "state": "unknown", "status": "static" }, "modprobe@configfs.service": { "name": "modprobe@configfs.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "modprobe@dm_mod.service": { "name": "modprobe@dm_mod.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "modprobe@drm.service": { "name": "modprobe@drm.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "modprobe@efi_pstore.service": { "name": "modprobe@efi_pstore.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "modprobe@fuse.service": { "name": "modprobe@fuse.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "modprobe@loop.service": { "name": "modprobe@loop.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "netavark-dhcp-proxy.service": { "name": "netavark-dhcp-proxy.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "netavark-firewalld-reload.service": { "name": "netavark-firewalld-reload.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "network.service": { "name": "network.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "nfs-blkmap.service": { "name": "nfs-blkmap.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-idmapd.service": { "name": "nfs-idmapd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-mountd.service": { "name": "nfs-mountd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-server.service": { "name": "nfs-server.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-utils.service": { "name": "nfs-utils.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfsdcld.service": { "name": "nfsdcld.service", "source": "systemd", "state": "stopped", "status": "static" }, "nftables.service": { "name": "nftables.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nis-domainname.service": { "name": "nis-domainname.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nm-priv-helper.service": { "name": "nm-priv-helper.service", "source": "systemd", "state": "inactive", "status": "static" }, "ntpd.service": { "name": "ntpd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ntpdate.service": { "name": "ntpdate.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "pam_namespace.service": { "name": "pam_namespace.service", "source": "systemd", "state": "inactive", "status": "static" }, "pcscd.service": { "name": "pcscd.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "plymouth-halt.service": { "name": "plymouth-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-kexec.service": { "name": "plymouth-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-poweroff.service": { "name": "plymouth-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-quit-wait.service": { "name": "plymouth-quit-wait.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-quit.service": { "name": "plymouth-quit.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-read-write.service": { "name": "plymouth-read-write.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-reboot.service": { "name": "plymouth-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-start.service": { "name": "plymouth-start.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-switch-root-initramfs.service": { "name": "plymouth-switch-root-initramfs.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-switch-root.service": { "name": "plymouth-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "podman-auto-update.service": { "name": "podman-auto-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "podman-clean-transient.service": { "name": "podman-clean-transient.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "podman-kube@.service": { "name": "podman-kube@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "podman-restart.service": { "name": "podman-restart.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "podman.service": { "name": "podman.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "polkit.service": { "name": "polkit.service", "source": "systemd", "state": "running", "status": "static" }, "quadlet-demo-mysql-volume.service": { "name": "quadlet-demo-mysql-volume.service", "source": "systemd", "state": "stopped", "status": "generated" }, "quadlet-demo-mysql.service": { "name": "quadlet-demo-mysql.service", "source": "systemd", "state": "running", "status": "generated" }, "quadlet-demo-network.service": { "name": "quadlet-demo-network.service", "source": "systemd", "state": "stopped", "status": "generated" }, "quotaon.service": { "name": "quotaon.service", "source": "systemd", "state": "inactive", "status": "static" }, "raid-check.service": { "name": "raid-check.service", "source": "systemd", "state": "stopped", "status": "static" }, "rc-local.service": { "name": "rc-local.service", "source": "systemd", "state": "stopped", "status": "static" }, "rescue.service": { "name": "rescue.service", "source": "systemd", "state": "stopped", "status": "static" }, "restraintd.service": { "name": "restraintd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rngd.service": { "name": "rngd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpc-gssd.service": { "name": "rpc-gssd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd-notify.service": { "name": "rpc-statd-notify.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd.service": { "name": "rpc-statd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-svcgssd.service": { "name": "rpc-svcgssd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "rpcbind.service": { "name": "rpcbind.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "rpmdb-migrate.service": { "name": "rpmdb-migrate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rpmdb-rebuild.service": { "name": "rpmdb-rebuild.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "selinux-autorelabel-mark.service": { "name": "selinux-autorelabel-mark.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "selinux-autorelabel.service": { "name": "selinux-autorelabel.service", "source": "systemd", "state": "inactive", "status": "static" }, "selinux-check-proper-disable.service": { "name": "selinux-check-proper-disable.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "serial-getty@.service": { "name": "serial-getty@.service", "source": "systemd", "state": "unknown", "status": "indirect" }, "serial-getty@ttyS0.service": { "name": "serial-getty@ttyS0.service", "source": "systemd", "state": "running", "status": "active" }, "snapd.seeded.service": { "name": "snapd.seeded.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sntp.service": { "name": "sntp.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ssh-host-keys-migration.service": { "name": "ssh-host-keys-migration.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "sshd-keygen.service": { "name": "sshd-keygen.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sshd-keygen@.service": { "name": "sshd-keygen@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@ecdsa.service": { "name": "sshd-keygen@ecdsa.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd-keygen@ed25519.service": { "name": "sshd-keygen@ed25519.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd-keygen@rsa.service": { "name": "sshd-keygen@rsa.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd.service": { "name": "sshd.service", "source": "systemd", "state": "running", "status": "enabled" }, "sshd@.service": { "name": "sshd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "sssd-autofs.service": { "name": "sssd-autofs.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-kcm.service": { "name": "sssd-kcm.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "sssd-nss.service": { "name": "sssd-nss.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pac.service": { "name": "sssd-pac.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pam.service": { "name": "sssd-pam.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-ssh.service": { "name": "sssd-ssh.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-sudo.service": { "name": "sssd-sudo.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd.service": { "name": "sssd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "syslog.service": { "name": "syslog.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "system-update-cleanup.service": { "name": "system-update-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-ask-password-console.service": { "name": "systemd-ask-password-console.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-plymouth.service": { "name": "systemd-ask-password-plymouth.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-wall.service": { "name": "systemd-ask-password-wall.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-backlight@.service": { "name": "systemd-backlight@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-battery-check.service": { "name": "systemd-battery-check.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-binfmt.service": { "name": "systemd-binfmt.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-bless-boot.service": { "name": "systemd-bless-boot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-boot-check-no-failures.service": { "name": "systemd-boot-check-no-failures.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-boot-random-seed.service": { "name": "systemd-boot-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-boot-update.service": { "name": "systemd-boot-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-confext.service": { "name": "systemd-confext.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-coredump@.service": { "name": "systemd-coredump@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-exit.service": { "name": "systemd-exit.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-firstboot.service": { "name": "systemd-firstboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck-root.service": { "name": "systemd-fsck-root.service", "source": "systemd", "state": "stopped", "status": "enabled-runtime" }, "systemd-fsck@.service": { "name": "systemd-fsck@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-growfs-root.service": { "name": "systemd-growfs-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-growfs@.service": { "name": "systemd-growfs@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-halt.service": { "name": "systemd-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hibernate.service": { "name": "systemd-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-homed-activate.service": { "name": "systemd-homed-activate.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-homed.service": { "name": "systemd-homed.service", "source": "systemd", "state": "running", "status": "enabled" }, "systemd-hostnamed.service": { "name": "systemd-hostnamed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hwdb-update.service": { "name": "systemd-hwdb-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-hybrid-sleep.service": { "name": "systemd-hybrid-sleep.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-initctl.service": { "name": "systemd-initctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-catalog-update.service": { "name": "systemd-journal-catalog-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-flush.service": { "name": "systemd-journal-flush.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journald.service": { "name": "systemd-journald.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-journald@.service": { "name": "systemd-journald@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-kexec.service": { "name": "systemd-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-localed.service": { "name": "systemd-localed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-logind.service": { "name": "systemd-logind.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-machine-id-commit.service": { "name": "systemd-machine-id-commit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-modules-load.service": { "name": "systemd-modules-load.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-network-generator.service": { "name": "systemd-network-generator.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-networkd-wait-online.service": { "name": "systemd-networkd-wait-online.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "systemd-networkd-wait-online@.service": { "name": "systemd-networkd-wait-online@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "systemd-networkd.service": { "name": "systemd-networkd.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "systemd-oomd.service": { "name": "systemd-oomd.service", "source": "systemd", "state": "running", "status": "enabled" }, "systemd-pcrfs-root.service": { "name": "systemd-pcrfs-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pcrfs@.service": { "name": "systemd-pcrfs@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-pcrmachine.service": { "name": "systemd-pcrmachine.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-pcrphase-initrd.service": { "name": "systemd-pcrphase-initrd.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-pcrphase-sysinit.service": { "name": "systemd-pcrphase-sysinit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-pcrphase.service": { "name": "systemd-pcrphase.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-portabled.service": { "name": "systemd-portabled.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-poweroff.service": { "name": "systemd-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pstore.service": { "name": "systemd-pstore.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-quotacheck.service": { "name": "systemd-quotacheck.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-random-seed.service": { "name": "systemd-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-reboot.service": { "name": "systemd-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-remount-fs.service": { "name": "systemd-remount-fs.service", "source": "systemd", "state": "stopped", "status": "enabled-runtime" }, "systemd-repart.service": { "name": "systemd-repart.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-resolved.service": { "name": "systemd-resolved.service", "source": "systemd", "state": "running", "status": "enabled" }, "systemd-rfkill.service": { "name": "systemd-rfkill.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-soft-reboot.service": { "name": "systemd-soft-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend-then-hibernate.service": { "name": "systemd-suspend-then-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend.service": { "name": "systemd-suspend.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-sysctl.service": { "name": "systemd-sysctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-sysext.service": { "name": "systemd-sysext.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "systemd-sysupdate-reboot.service": { "name": "systemd-sysupdate-reboot.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "systemd-sysupdate.service": { "name": "systemd-sysupdate.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "systemd-sysusers.service": { "name": "systemd-sysusers.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-time-wait-sync.service": { "name": "systemd-time-wait-sync.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-timedated.service": { "name": "systemd-timedated.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-timesyncd.service": { "name": "systemd-timesyncd.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "systemd-tmpfiles-clean.service": { "name": "systemd-tmpfiles-clean.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev-early.service": { "name": "systemd-tmpfiles-setup-dev-early.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev.service": { "name": "systemd-tmpfiles-setup-dev.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup.service": { "name": "systemd-tmpfiles-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-settle.service": { "name": "systemd-udev-settle.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-trigger.service": { "name": "systemd-udev-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udevd.service": { "name": "systemd-udevd.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-update-done.service": { "name": "systemd-update-done.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp-runlevel.service": { "name": "systemd-update-utmp-runlevel.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp.service": { "name": "systemd-update-utmp.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-user-sessions.service": { "name": "systemd-user-sessions.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-userdbd.service": { "name": "systemd-userdbd.service", "source": "systemd", "state": "running", "status": "indirect" }, "systemd-vconsole-setup.service": { "name": "systemd-vconsole-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-volatile-root.service": { "name": "systemd-volatile-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-zram-setup@.service": { "name": "systemd-zram-setup@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-zram-setup@zram0.service": { "name": "systemd-zram-setup@zram0.service", "source": "systemd", "state": "stopped", "status": "active" }, "udisks2.service": { "name": "udisks2.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "unbound-anchor.service": { "name": "unbound-anchor.service", "source": "systemd", "state": "stopped", "status": "static" }, "user-runtime-dir@.service": { "name": "user-runtime-dir@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user-runtime-dir@0.service": { "name": "user-runtime-dir@0.service", "source": "systemd", "state": "stopped", "status": "active" }, "user@.service": { "name": "user@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user@0.service": { "name": "user@0.service", "source": "systemd", "state": "running", "status": "active" } } }, "changed": false } TASK [fedora.linux_system_roles.podman : Create and update quadlets] *********** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:114 Wednesday 10 July 2024 05:15:10 +0000 (0:00:02.210) 0:03:15.320 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_state != \"absent\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14 Wednesday 10 July 2024 05:15:10 +0000 (0:00:00.021) 0:03:15.342 ******** ok: [sut] => { "ansible_facts": { "__podman_quadlet_file_src": "envoy-proxy-configmap.yml", "__podman_quadlet_spec": {}, "__podman_quadlet_str": "---\napiVersion: v1\nkind: ConfigMap\nmetadata:\n name: envoy-proxy-config\ndata:\n envoy.yaml: |\n admin:\n address:\n socket_address:\n address: 0.0.0.0\n port_value: 9901\n\n static_resources:\n listeners:\n - name: listener_0\n address:\n socket_address:\n address: 0.0.0.0\n port_value: 8080\n filter_chains:\n - filters:\n - name: envoy.filters.network.http_connection_manager\n typed_config:\n \"@type\": type.googleapis.com/envoy.extensions.filters.network.http_connection_manager.v3.HttpConnectionManager\n stat_prefix: ingress_http\n codec_type: AUTO\n route_config:\n name: local_route\n virtual_hosts:\n - name: local_service\n domains: [\"*\"]\n routes:\n - match:\n prefix: \"/\"\n route:\n cluster: backend\n http_filters:\n - name: envoy.filters.http.router\n typed_config:\n \"@type\": type.googleapis.com/envoy.extensions.filters.http.router.v3.Router\n transport_socket:\n name: envoy.transport_sockets.tls\n typed_config:\n \"@type\": type.googleapis.com/envoy.extensions.transport_sockets.tls.v3.DownstreamTlsContext\n common_tls_context:\n tls_certificates:\n - certificate_chain:\n filename: /etc/envoy-certificates/certificate.pem\n private_key:\n filename: /etc/envoy-certificates/certificate.key\n clusters:\n - name: backend\n connect_timeout: 5s\n type: STATIC\n dns_refresh_rate: 1800s\n lb_policy: ROUND_ROBIN\n load_assignment:\n cluster_name: backend\n endpoints:\n - lb_endpoints:\n - endpoint:\n address:\n socket_address:\n address: 127.0.0.1\n port_value: 80", "__podman_quadlet_template_src": "" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25 Wednesday 10 July 2024 05:15:10 +0000 (0:00:00.030) 0:03:15.373 ******** ok: [sut] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_pull_image": true, "__podman_state": "absent", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] ***** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35 Wednesday 10 July 2024 05:15:10 +0000 (0:00:00.026) 0:03:15.400 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_quadlet_file_src", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48 Wednesday 10 July 2024 05:15:10 +0000 (0:00:00.022) 0:03:15.422 ******** ok: [sut] => { "ansible_facts": { "__podman_quadlet_name": "envoy-proxy-configmap", "__podman_quadlet_type": "yml", "__podman_rootless": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57 Wednesday 10 July 2024 05:15:10 +0000 (0:00:00.032) 0:03:15.454 ******** included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for sut TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Wednesday 10 July 2024 05:15:10 +0000 (0:00:00.040) 0:03:15.495 ******** skipping: [sut] => { "changed": false, "false_condition": "'getent_passwd' not in ansible_facts or __podman_user not in ansible_facts['getent_passwd']", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Wednesday 10 July 2024 05:15:10 +0000 (0:00:00.063) 0:03:15.558 ******** skipping: [sut] => { "changed": false, "false_condition": "not ansible_facts[\"getent_passwd\"][__podman_user]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Wednesday 10 July 2024 05:15:10 +0000 (0:00:00.026) 0:03:15.585 ******** ok: [sut] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get group information] **************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:28 Wednesday 10 July 2024 05:15:10 +0000 (0:00:00.032) 0:03:15.617 ******** ok: [sut] => { "ansible_facts": { "getent_group": { "root": [ "x", "0", "" ] } }, "changed": false } TASK [fedora.linux_system_roles.podman : Set group name] *********************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:35 Wednesday 10 July 2024 05:15:11 +0000 (0:00:00.242) 0:03:15.859 ******** ok: [sut] => { "ansible_facts": { "__podman_group_name": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:39 Wednesday 10 July 2024 05:15:11 +0000 (0:00:00.030) 0:03:15.889 ******** ok: [sut] => { "changed": false, "stat": { "atime": 1720588427.5164855, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "72ba28a3be19c467e89ab90f5d21f99b605159a7", "ctime": 1720588377.3007717, "dev": 51714, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 203603, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-pie-executable", "mode": "0755", "mtime": 1696291200.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 16032, "uid": 0, "version": "2092964490", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************ task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:50 Wednesday 10 July 2024 05:15:11 +0000 (0:00:00.241) 0:03:16.131 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check group with getsubids] *********** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:55 Wednesday 10 July 2024 05:15:11 +0000 (0:00:00.021) 0:03:16.152 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:60 Wednesday 10 July 2024 05:15:11 +0000 (0:00:00.020) 0:03:16.173 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subuid file] ********************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:74 Wednesday 10 July 2024 05:15:11 +0000 (0:00:00.020) 0:03:16.193 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subgid file] ********************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:79 Wednesday 10 July 2024 05:15:11 +0000 (0:00:00.020) 0:03:16.214 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:84 Wednesday 10 July 2024 05:15:11 +0000 (0:00:00.020) 0:03:16.234 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:94 Wednesday 10 July 2024 05:15:11 +0000 (0:00:00.019) 0:03:16.253 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] ***** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:101 Wednesday 10 July 2024 05:15:11 +0000 (0:00:00.020) 0:03:16.274 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62 Wednesday 10 July 2024 05:15:11 +0000 (0:00:00.020) 0:03:16.294 ******** ok: [sut] => { "ansible_facts": { "__podman_activate_systemd_unit": true, "__podman_images_found": [], "__podman_kube_yamls_raw": "", "__podman_service_name": "", "__podman_systemd_scope": "system", "__podman_user_home_dir": "/root", "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73 Wednesday 10 July 2024 05:15:11 +0000 (0:00:00.037) 0:03:16.331 ******** ok: [sut] => { "ansible_facts": { "__podman_quadlet_path": "/etc/containers/systemd" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get kube yaml contents] *************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77 Wednesday 10 July 2024 05:15:11 +0000 (0:00:00.061) 0:03:16.393 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_state != \"absent\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87 Wednesday 10 July 2024 05:15:11 +0000 (0:00:00.022) 0:03:16.415 ******** ok: [sut] => { "ansible_facts": { "__podman_images": [], "__podman_quadlet_file": "/etc/containers/systemd/envoy-proxy-configmap.yml", "__podman_volumes": [] }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 6] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:103 Wednesday 10 July 2024 05:15:11 +0000 (0:00:00.049) 0:03:16.464 ******** ok: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Cleanup quadlets] ********************* task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:110 Wednesday 10 July 2024 05:15:11 +0000 (0:00:00.025) 0:03:16.490 ******** included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml for sut TASK [fedora.linux_system_roles.podman : Stat XDG_RUNTIME_DIR] ***************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:4 Wednesday 10 July 2024 05:15:11 +0000 (0:00:00.047) 0:03:16.538 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Stop and disable service] ************* task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:12 Wednesday 10 July 2024 05:15:11 +0000 (0:00:00.020) 0:03:16.559 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_service_name | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Remove quadlet file] ****************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:33 Wednesday 10 July 2024 05:15:11 +0000 (0:00:00.022) 0:03:16.581 ******** changed: [sut] => { "changed": true, "path": "/etc/containers/systemd/envoy-proxy-configmap.yml", "state": "absent" } TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:39 Wednesday 10 July 2024 05:15:12 +0000 (0:00:00.240) 0:03:16.821 ******** included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for sut TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Wednesday 10 July 2024 05:15:12 +0000 (0:00:00.035) 0:03:16.856 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Wednesday 10 July 2024 05:15:12 +0000 (0:00:00.020) 0:03:16.877 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Wednesday 10 July 2024 05:15:12 +0000 (0:00:00.019) 0:03:16.897 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Reload systemctl] ********************* task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:47 Wednesday 10 July 2024 05:15:12 +0000 (0:00:00.020) 0:03:16.918 ******** ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.podman : Prune images no longer in use] ******** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:56 Wednesday 10 July 2024 05:15:13 +0000 (0:00:00.843) 0:03:17.761 ******** changed: [sut] => { "changed": true, "cmd": [ "podman", "image", "prune", "--all", "-f" ], "delta": "0:00:00.031594", "end": "2024-07-10 05:15:13.330891", "rc": 0, "start": "2024-07-10 05:15:13.299297" } TASK [fedora.linux_system_roles.podman : Remove networks no longer in use] ***** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:64 Wednesday 10 July 2024 05:15:13 +0000 (0:00:00.272) 0:03:18.034 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_quadlet_type == \"network\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Remove volumes no longer in use] ****** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:73 Wednesday 10 July 2024 05:15:13 +0000 (0:00:00.022) 0:03:18.057 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_quadlet_type == \"volume\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : For testing and debugging - images] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:82 Wednesday 10 July 2024 05:15:13 +0000 (0:00:00.022) 0:03:18.079 ******** ok: [sut] => { "changed": false, "cmd": [ "podman", "images", "-n" ], "delta": "0:00:00.037608", "end": "2024-07-10 05:15:13.652381", "rc": 0, "start": "2024-07-10 05:15:13.614773" } STDOUT: quay.io/linux-system-roles/mysql 5.6 dd3b2a5dcb48 2 years ago 308 MB TASK [fedora.linux_system_roles.podman : For testing and debugging - volumes] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:92 Wednesday 10 July 2024 05:15:13 +0000 (0:00:00.273) 0:03:18.353 ******** ok: [sut] => { "changed": false, "cmd": [ "podman", "volume", "ls", "-n" ], "delta": "0:00:00.061854", "end": "2024-07-10 05:15:13.953104", "rc": 0, "start": "2024-07-10 05:15:13.891250" } STDOUT: local systemd-quadlet-demo-mysql local wp-pv-claim local envoy-proxy-config local envoy-certificates TASK [fedora.linux_system_roles.podman : For testing and debugging - containers] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:102 Wednesday 10 July 2024 05:15:13 +0000 (0:00:00.307) 0:03:18.660 ******** ok: [sut] => { "changed": false, "cmd": [ "podman", "ps", "--noheading" ], "delta": "0:00:00.104094", "end": "2024-07-10 05:15:14.399793", "rc": 0, "start": "2024-07-10 05:15:14.295699" } TASK [fedora.linux_system_roles.podman : For testing and debugging - networks] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:112 Wednesday 10 July 2024 05:15:14 +0000 (0:00:00.443) 0:03:19.103 ******** ok: [sut] => { "changed": false, "cmd": [ "podman", "network", "ls", "-n", "-q" ], "delta": "0:00:00.032065", "end": "2024-07-10 05:15:14.674350", "rc": 0, "start": "2024-07-10 05:15:14.642285" } STDOUT: podman systemd-quadlet-demo TASK [fedora.linux_system_roles.podman : For testing and debugging - secrets] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:122 Wednesday 10 July 2024 05:15:14 +0000 (0:00:00.273) 0:03:19.377 ******** ok: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : For testing and debugging - services] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:133 Wednesday 10 July 2024 05:15:14 +0000 (0:00:00.264) 0:03:19.641 ******** ok: [sut] => { "ansible_facts": { "services": { "NetworkManager-dispatcher.service": { "name": "NetworkManager-dispatcher.service", "source": "systemd", "state": "running", "status": "enabled" }, "NetworkManager-wait-online.service": { "name": "NetworkManager-wait-online.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "NetworkManager.service": { "name": "NetworkManager.service", "source": "systemd", "state": "running", "status": "enabled" }, "arp-ethers.service": { "name": "arp-ethers.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "auditd.service": { "name": "auditd.service", "source": "systemd", "state": "running", "status": "enabled" }, "auth-rpcgss-module.service": { "name": "auth-rpcgss-module.service", "source": "systemd", "state": "stopped", "status": "static" }, "autovt@.service": { "name": "autovt@.service", "source": "systemd", "state": "unknown", "status": "alias" }, "blk-availability.service": { "name": "blk-availability.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "bluetooth.service": { "name": "bluetooth.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "certmonger.service": { "name": "certmonger.service", "source": "systemd", "state": "running", "status": "enabled" }, "chrony-wait.service": { "name": "chrony-wait.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd-restricted.service": { "name": "chronyd-restricted.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd.service": { "name": "chronyd.service", "source": "systemd", "state": "running", "status": "enabled" }, "cloud-config.service": { "name": "cloud-config.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-final.service": { "name": "cloud-final.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init-hotplugd.service": { "name": "cloud-init-hotplugd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cloud-init-local.service": { "name": "cloud-init-local.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init.service": { "name": "cloud-init.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "console-getty.service": { "name": "console-getty.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "container-getty@.service": { "name": "container-getty@.service", "source": "systemd", "state": "unknown", "status": "static" }, "dbus-broker.service": { "name": "dbus-broker.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-org.bluez.service": { "name": "dbus-org.bluez.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.fedoraproject.FirewallD1.service": { "name": "dbus-org.fedoraproject.FirewallD1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.home1.service": { "name": "dbus-org.freedesktop.home1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.hostname1.service": { "name": "dbus-org.freedesktop.hostname1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.locale1.service": { "name": "dbus-org.freedesktop.locale1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.login1.service": { "name": "dbus-org.freedesktop.login1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.nm-dispatcher.service": { "name": "dbus-org.freedesktop.nm-dispatcher.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.oom1.service": { "name": "dbus-org.freedesktop.oom1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.portable1.service": { "name": "dbus-org.freedesktop.portable1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.resolve1.service": { "name": "dbus-org.freedesktop.resolve1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.timedate1.service": { "name": "dbus-org.freedesktop.timedate1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus.service": { "name": "dbus.service", "source": "systemd", "state": "active", "status": "alias" }, "debug-shell.service": { "name": "debug-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "display-manager.service": { "name": "display-manager.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "dm-event.service": { "name": "dm-event.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-makecache.service": { "name": "dnf-makecache.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-system-upgrade-cleanup.service": { "name": "dnf-system-upgrade-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "dnf-system-upgrade.service": { "name": "dnf-system-upgrade.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dracut-cmdline.service": { "name": "dracut-cmdline.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-initqueue.service": { "name": "dracut-initqueue.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-mount.service": { "name": "dracut-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-mount.service": { "name": "dracut-pre-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-pivot.service": { "name": "dracut-pre-pivot.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-trigger.service": { "name": "dracut-pre-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-udev.service": { "name": "dracut-pre-udev.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown-onfailure.service": { "name": "dracut-shutdown-onfailure.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown.service": { "name": "dracut-shutdown.service", "source": "systemd", "state": "stopped", "status": "static" }, "ebtables.service": { "name": "ebtables.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "emergency.service": { "name": "emergency.service", "source": "systemd", "state": "stopped", "status": "static" }, "firewalld.service": { "name": "firewalld.service", "source": "systemd", "state": "running", "status": "enabled" }, "fsidd.service": { "name": "fsidd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "fstrim.service": { "name": "fstrim.service", "source": "systemd", "state": "stopped", "status": "static" }, "fwupd-offline-update.service": { "name": "fwupd-offline-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd-refresh.service": { "name": "fwupd-refresh.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd.service": { "name": "fwupd.service", "source": "systemd", "state": "inactive", "status": "static" }, "getty@.service": { "name": "getty@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "getty@tty1.service": { "name": "getty@tty1.service", "source": "systemd", "state": "running", "status": "active" }, "grub-boot-indeterminate.service": { "name": "grub-boot-indeterminate.service", "source": "systemd", "state": "inactive", "status": "static" }, "grub2-systemd-integration.service": { "name": "grub2-systemd-integration.service", "source": "systemd", "state": "inactive", "status": "static" }, "gssproxy.service": { "name": "gssproxy.service", "source": "systemd", "state": "running", "status": "disabled" }, "hv_kvp_daemon.service": { "name": "hv_kvp_daemon.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "import-state.service": { "name": "import-state.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "initrd-cleanup.service": { "name": "initrd-cleanup.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-parse-etc.service": { "name": "initrd-parse-etc.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-switch-root.service": { "name": "initrd-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-udevadm-cleanup-db.service": { "name": "initrd-udevadm-cleanup-db.service", "source": "systemd", "state": "stopped", "status": "static" }, "ip6tables.service": { "name": "ip6tables.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ipset.service": { "name": "ipset.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "iptables.service": { "name": "iptables.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "kmod-static-nodes.service": { "name": "kmod-static-nodes.service", "source": "systemd", "state": "stopped", "status": "static" }, "ldconfig.service": { "name": "ldconfig.service", "source": "systemd", "state": "stopped", "status": "static" }, "loadmodules.service": { "name": "loadmodules.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "lvm2-lvmpolld.service": { "name": "lvm2-lvmpolld.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-monitor.service": { "name": "lvm2-monitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "man-db-cache-update.service": { "name": "man-db-cache-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "man-db-restart-cache-update.service": { "name": "man-db-restart-cache-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "mdadm-grow-continue@.service": { "name": "mdadm-grow-continue@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdadm-last-resort@.service": { "name": "mdadm-last-resort@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdcheck_continue.service": { "name": "mdcheck_continue.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdcheck_start.service": { "name": "mdcheck_start.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmon@.service": { "name": "mdmon@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdmonitor-oneshot.service": { "name": "mdmonitor-oneshot.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmonitor.service": { "name": "mdmonitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "modprobe@.service": { "name": "modprobe@.service", "source": "systemd", "state": "unknown", "status": "static" }, "modprobe@configfs.service": { "name": "modprobe@configfs.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "modprobe@dm_mod.service": { "name": "modprobe@dm_mod.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "modprobe@drm.service": { "name": "modprobe@drm.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "modprobe@efi_pstore.service": { "name": "modprobe@efi_pstore.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "modprobe@fuse.service": { "name": "modprobe@fuse.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "modprobe@loop.service": { "name": "modprobe@loop.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "netavark-dhcp-proxy.service": { "name": "netavark-dhcp-proxy.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "netavark-firewalld-reload.service": { "name": "netavark-firewalld-reload.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "network.service": { "name": "network.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "nfs-blkmap.service": { "name": "nfs-blkmap.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-idmapd.service": { "name": "nfs-idmapd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-mountd.service": { "name": "nfs-mountd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-server.service": { "name": "nfs-server.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-utils.service": { "name": "nfs-utils.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfsdcld.service": { "name": "nfsdcld.service", "source": "systemd", "state": "stopped", "status": "static" }, "nftables.service": { "name": "nftables.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nis-domainname.service": { "name": "nis-domainname.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nm-priv-helper.service": { "name": "nm-priv-helper.service", "source": "systemd", "state": "inactive", "status": "static" }, "ntpd.service": { "name": "ntpd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ntpdate.service": { "name": "ntpdate.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "pam_namespace.service": { "name": "pam_namespace.service", "source": "systemd", "state": "inactive", "status": "static" }, "pcscd.service": { "name": "pcscd.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "plymouth-halt.service": { "name": "plymouth-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-kexec.service": { "name": "plymouth-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-poweroff.service": { "name": "plymouth-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-quit-wait.service": { "name": "plymouth-quit-wait.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-quit.service": { "name": "plymouth-quit.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-read-write.service": { "name": "plymouth-read-write.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-reboot.service": { "name": "plymouth-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-start.service": { "name": "plymouth-start.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-switch-root-initramfs.service": { "name": "plymouth-switch-root-initramfs.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-switch-root.service": { "name": "plymouth-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "podman-auto-update.service": { "name": "podman-auto-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "podman-clean-transient.service": { "name": "podman-clean-transient.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "podman-kube@.service": { "name": "podman-kube@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "podman-restart.service": { "name": "podman-restart.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "podman.service": { "name": "podman.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "polkit.service": { "name": "polkit.service", "source": "systemd", "state": "running", "status": "static" }, "quadlet-demo-mysql-volume.service": { "name": "quadlet-demo-mysql-volume.service", "source": "systemd", "state": "stopped", "status": "generated" }, "quadlet-demo-mysql.service": { "name": "quadlet-demo-mysql.service", "source": "systemd", "state": "stopped", "status": "failed" }, "quadlet-demo-network.service": { "name": "quadlet-demo-network.service", "source": "systemd", "state": "stopped", "status": "generated" }, "quotaon.service": { "name": "quotaon.service", "source": "systemd", "state": "inactive", "status": "static" }, "raid-check.service": { "name": "raid-check.service", "source": "systemd", "state": "stopped", "status": "static" }, "rc-local.service": { "name": "rc-local.service", "source": "systemd", "state": "stopped", "status": "static" }, "rescue.service": { "name": "rescue.service", "source": "systemd", "state": "stopped", "status": "static" }, "restraintd.service": { "name": "restraintd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rngd.service": { "name": "rngd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpc-gssd.service": { "name": "rpc-gssd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd-notify.service": { "name": "rpc-statd-notify.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd.service": { "name": "rpc-statd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-svcgssd.service": { "name": "rpc-svcgssd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "rpcbind.service": { "name": "rpcbind.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "rpmdb-migrate.service": { "name": "rpmdb-migrate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rpmdb-rebuild.service": { "name": "rpmdb-rebuild.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "selinux-autorelabel-mark.service": { "name": "selinux-autorelabel-mark.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "selinux-autorelabel.service": { "name": "selinux-autorelabel.service", "source": "systemd", "state": "inactive", "status": "static" }, "selinux-check-proper-disable.service": { "name": "selinux-check-proper-disable.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "serial-getty@.service": { "name": "serial-getty@.service", "source": "systemd", "state": "unknown", "status": "indirect" }, "serial-getty@ttyS0.service": { "name": "serial-getty@ttyS0.service", "source": "systemd", "state": "running", "status": "active" }, "snapd.seeded.service": { "name": "snapd.seeded.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sntp.service": { "name": "sntp.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ssh-host-keys-migration.service": { "name": "ssh-host-keys-migration.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "sshd-keygen.service": { "name": "sshd-keygen.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sshd-keygen@.service": { "name": "sshd-keygen@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@ecdsa.service": { "name": "sshd-keygen@ecdsa.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd-keygen@ed25519.service": { "name": "sshd-keygen@ed25519.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd-keygen@rsa.service": { "name": "sshd-keygen@rsa.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd.service": { "name": "sshd.service", "source": "systemd", "state": "running", "status": "enabled" }, "sshd@.service": { "name": "sshd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "sssd-autofs.service": { "name": "sssd-autofs.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-kcm.service": { "name": "sssd-kcm.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "sssd-nss.service": { "name": "sssd-nss.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pac.service": { "name": "sssd-pac.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pam.service": { "name": "sssd-pam.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-ssh.service": { "name": "sssd-ssh.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-sudo.service": { "name": "sssd-sudo.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd.service": { "name": "sssd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "syslog.service": { "name": "syslog.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "system-update-cleanup.service": { "name": "system-update-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-ask-password-console.service": { "name": "systemd-ask-password-console.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-plymouth.service": { "name": "systemd-ask-password-plymouth.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-wall.service": { "name": "systemd-ask-password-wall.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-backlight@.service": { "name": "systemd-backlight@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-battery-check.service": { "name": "systemd-battery-check.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-binfmt.service": { "name": "systemd-binfmt.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-bless-boot.service": { "name": "systemd-bless-boot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-boot-check-no-failures.service": { "name": "systemd-boot-check-no-failures.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-boot-random-seed.service": { "name": "systemd-boot-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-boot-update.service": { "name": "systemd-boot-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-confext.service": { "name": "systemd-confext.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-coredump@.service": { "name": "systemd-coredump@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-exit.service": { "name": "systemd-exit.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-firstboot.service": { "name": "systemd-firstboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck-root.service": { "name": "systemd-fsck-root.service", "source": "systemd", "state": "stopped", "status": "enabled-runtime" }, "systemd-fsck@.service": { "name": "systemd-fsck@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-growfs-root.service": { "name": "systemd-growfs-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-growfs@.service": { "name": "systemd-growfs@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-halt.service": { "name": "systemd-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hibernate.service": { "name": "systemd-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-homed-activate.service": { "name": "systemd-homed-activate.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-homed.service": { "name": "systemd-homed.service", "source": "systemd", "state": "running", "status": "enabled" }, "systemd-hostnamed.service": { "name": "systemd-hostnamed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hwdb-update.service": { "name": "systemd-hwdb-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-hybrid-sleep.service": { "name": "systemd-hybrid-sleep.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-initctl.service": { "name": "systemd-initctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-catalog-update.service": { "name": "systemd-journal-catalog-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-flush.service": { "name": "systemd-journal-flush.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journald.service": { "name": "systemd-journald.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-journald@.service": { "name": "systemd-journald@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-kexec.service": { "name": "systemd-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-localed.service": { "name": "systemd-localed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-logind.service": { "name": "systemd-logind.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-machine-id-commit.service": { "name": "systemd-machine-id-commit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-modules-load.service": { "name": "systemd-modules-load.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-network-generator.service": { "name": "systemd-network-generator.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-networkd-wait-online.service": { "name": "systemd-networkd-wait-online.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "systemd-networkd-wait-online@.service": { "name": "systemd-networkd-wait-online@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "systemd-networkd.service": { "name": "systemd-networkd.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "systemd-oomd.service": { "name": "systemd-oomd.service", "source": "systemd", "state": "running", "status": "enabled" }, "systemd-pcrfs-root.service": { "name": "systemd-pcrfs-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pcrfs@.service": { "name": "systemd-pcrfs@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-pcrmachine.service": { "name": "systemd-pcrmachine.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-pcrphase-initrd.service": { "name": "systemd-pcrphase-initrd.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-pcrphase-sysinit.service": { "name": "systemd-pcrphase-sysinit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-pcrphase.service": { "name": "systemd-pcrphase.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-portabled.service": { "name": "systemd-portabled.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-poweroff.service": { "name": "systemd-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pstore.service": { "name": "systemd-pstore.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-quotacheck.service": { "name": "systemd-quotacheck.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-random-seed.service": { "name": "systemd-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-reboot.service": { "name": "systemd-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-remount-fs.service": { "name": "systemd-remount-fs.service", "source": "systemd", "state": "stopped", "status": "enabled-runtime" }, "systemd-repart.service": { "name": "systemd-repart.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-resolved.service": { "name": "systemd-resolved.service", "source": "systemd", "state": "running", "status": "enabled" }, "systemd-rfkill.service": { "name": "systemd-rfkill.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-soft-reboot.service": { "name": "systemd-soft-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend-then-hibernate.service": { "name": "systemd-suspend-then-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend.service": { "name": "systemd-suspend.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-sysctl.service": { "name": "systemd-sysctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-sysext.service": { "name": "systemd-sysext.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "systemd-sysupdate-reboot.service": { "name": "systemd-sysupdate-reboot.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "systemd-sysupdate.service": { "name": "systemd-sysupdate.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "systemd-sysusers.service": { "name": "systemd-sysusers.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-time-wait-sync.service": { "name": "systemd-time-wait-sync.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-timedated.service": { "name": "systemd-timedated.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-timesyncd.service": { "name": "systemd-timesyncd.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "systemd-tmpfiles-clean.service": { "name": "systemd-tmpfiles-clean.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev-early.service": { "name": "systemd-tmpfiles-setup-dev-early.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev.service": { "name": "systemd-tmpfiles-setup-dev.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup.service": { "name": "systemd-tmpfiles-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-settle.service": { "name": "systemd-udev-settle.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-trigger.service": { "name": "systemd-udev-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udevd.service": { "name": "systemd-udevd.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-update-done.service": { "name": "systemd-update-done.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp-runlevel.service": { "name": "systemd-update-utmp-runlevel.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp.service": { "name": "systemd-update-utmp.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-user-sessions.service": { "name": "systemd-user-sessions.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-userdbd.service": { "name": "systemd-userdbd.service", "source": "systemd", "state": "running", "status": "indirect" }, "systemd-vconsole-setup.service": { "name": "systemd-vconsole-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-volatile-root.service": { "name": "systemd-volatile-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-zram-setup@.service": { "name": "systemd-zram-setup@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-zram-setup@zram0.service": { "name": "systemd-zram-setup@zram0.service", "source": "systemd", "state": "stopped", "status": "active" }, "udisks2.service": { "name": "udisks2.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "unbound-anchor.service": { "name": "unbound-anchor.service", "source": "systemd", "state": "stopped", "status": "static" }, "user-runtime-dir@.service": { "name": "user-runtime-dir@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user-runtime-dir@0.service": { "name": "user-runtime-dir@0.service", "source": "systemd", "state": "stopped", "status": "active" }, "user@.service": { "name": "user@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user@0.service": { "name": "user@0.service", "source": "systemd", "state": "running", "status": "active" } } }, "changed": false } TASK [fedora.linux_system_roles.podman : Create and update quadlets] *********** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:114 Wednesday 10 July 2024 05:15:17 +0000 (0:00:02.182) 0:03:21.823 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_state != \"absent\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14 Wednesday 10 July 2024 05:15:17 +0000 (0:00:00.020) 0:03:21.844 ******** ok: [sut] => { "ansible_facts": { "__podman_quadlet_file_src": "", "__podman_quadlet_spec": {}, "__podman_quadlet_str": "[Install]\nWantedBy=default.target\n\n[Container]\nImage=quay.io/linux-system-roles/mysql:5.6\nContainerName=quadlet-demo-mysql\nVolume=quadlet-demo-mysql.volume:/var/lib/mysql\nVolume=/tmp/quadlet_demo:/var/lib/quadlet_demo:Z\nNetwork=quadlet-demo.network\nSecret=mysql-root-password-container,type=env,target=MYSQL_ROOT_PASSWORD\nHealthCmd=/usr/bin/true\nHealthOnFailure=kill\n", "__podman_quadlet_template_src": "quadlet-demo-mysql.container.j2" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25 Wednesday 10 July 2024 05:15:17 +0000 (0:00:00.070) 0:03:21.915 ******** ok: [sut] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_pull_image": true, "__podman_state": "absent", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] ***** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35 Wednesday 10 July 2024 05:15:17 +0000 (0:00:00.026) 0:03:21.942 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_quadlet_str", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48 Wednesday 10 July 2024 05:15:17 +0000 (0:00:00.022) 0:03:21.964 ******** ok: [sut] => { "ansible_facts": { "__podman_quadlet_name": "quadlet-demo-mysql", "__podman_quadlet_type": "container", "__podman_rootless": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57 Wednesday 10 July 2024 05:15:17 +0000 (0:00:00.073) 0:03:22.037 ******** included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for sut TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Wednesday 10 July 2024 05:15:17 +0000 (0:00:00.042) 0:03:22.080 ******** skipping: [sut] => { "changed": false, "false_condition": "'getent_passwd' not in ansible_facts or __podman_user not in ansible_facts['getent_passwd']", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Wednesday 10 July 2024 05:15:17 +0000 (0:00:00.026) 0:03:22.106 ******** skipping: [sut] => { "changed": false, "false_condition": "not ansible_facts[\"getent_passwd\"][__podman_user]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Wednesday 10 July 2024 05:15:17 +0000 (0:00:00.025) 0:03:22.132 ******** ok: [sut] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get group information] **************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:28 Wednesday 10 July 2024 05:15:17 +0000 (0:00:00.032) 0:03:22.165 ******** ok: [sut] => { "ansible_facts": { "getent_group": { "root": [ "x", "0", "" ] } }, "changed": false } TASK [fedora.linux_system_roles.podman : Set group name] *********************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:35 Wednesday 10 July 2024 05:15:17 +0000 (0:00:00.243) 0:03:22.408 ******** ok: [sut] => { "ansible_facts": { "__podman_group_name": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:39 Wednesday 10 July 2024 05:15:17 +0000 (0:00:00.032) 0:03:22.441 ******** ok: [sut] => { "changed": false, "stat": { "atime": 1720588427.5164855, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "72ba28a3be19c467e89ab90f5d21f99b605159a7", "ctime": 1720588377.3007717, "dev": 51714, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 203603, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-pie-executable", "mode": "0755", "mtime": 1696291200.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 16032, "uid": 0, "version": "2092964490", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************ task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:50 Wednesday 10 July 2024 05:15:18 +0000 (0:00:00.241) 0:03:22.682 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check group with getsubids] *********** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:55 Wednesday 10 July 2024 05:15:18 +0000 (0:00:00.021) 0:03:22.703 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:60 Wednesday 10 July 2024 05:15:18 +0000 (0:00:00.020) 0:03:22.724 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subuid file] ********************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:74 Wednesday 10 July 2024 05:15:18 +0000 (0:00:00.020) 0:03:22.744 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subgid file] ********************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:79 Wednesday 10 July 2024 05:15:18 +0000 (0:00:00.020) 0:03:22.764 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:84 Wednesday 10 July 2024 05:15:18 +0000 (0:00:00.019) 0:03:22.784 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:94 Wednesday 10 July 2024 05:15:18 +0000 (0:00:00.019) 0:03:22.804 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] ***** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:101 Wednesday 10 July 2024 05:15:18 +0000 (0:00:00.060) 0:03:22.865 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62 Wednesday 10 July 2024 05:15:18 +0000 (0:00:00.022) 0:03:22.888 ******** ok: [sut] => { "ansible_facts": { "__podman_activate_systemd_unit": true, "__podman_images_found": [ "quay.io/linux-system-roles/mysql:5.6" ], "__podman_kube_yamls_raw": "", "__podman_service_name": "quadlet-demo-mysql.service", "__podman_systemd_scope": "system", "__podman_user_home_dir": "/root", "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73 Wednesday 10 July 2024 05:15:18 +0000 (0:00:00.039) 0:03:22.927 ******** ok: [sut] => { "ansible_facts": { "__podman_quadlet_path": "/etc/containers/systemd" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get kube yaml contents] *************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77 Wednesday 10 July 2024 05:15:18 +0000 (0:00:00.023) 0:03:22.950 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_state != \"absent\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87 Wednesday 10 July 2024 05:15:18 +0000 (0:00:00.020) 0:03:22.971 ******** ok: [sut] => { "ansible_facts": { "__podman_images": [ "quay.io/linux-system-roles/mysql:5.6" ], "__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo-mysql.container", "__podman_volumes": [ "/tmp/quadlet_demo" ] }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 6] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:103 Wednesday 10 July 2024 05:15:18 +0000 (0:00:00.049) 0:03:23.020 ******** ok: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Cleanup quadlets] ********************* task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:110 Wednesday 10 July 2024 05:15:18 +0000 (0:00:00.026) 0:03:23.047 ******** included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml for sut TASK [fedora.linux_system_roles.podman : Stat XDG_RUNTIME_DIR] ***************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:4 Wednesday 10 July 2024 05:15:18 +0000 (0:00:00.046) 0:03:23.093 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Stop and disable service] ************* task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:12 Wednesday 10 July 2024 05:15:18 +0000 (0:00:00.020) 0:03:23.114 ******** changed: [sut] => { "changed": true, "enabled": false, "failed_when_result": false, "name": "quadlet-demo-mysql.service", "state": "stopped", "status": { "AccessSELinuxContext": "system_u:object_r:systemd_unit_file_t:s0", "ActiveEnterTimestamp": "Wed 2024-07-10 05:14:13 UTC", "ActiveEnterTimestampMonotonic": "973531941", "ActiveExitTimestamp": "Wed 2024-07-10 05:15:14 UTC", "ActiveExitTimestampMonotonic": "1034753169", "ActiveState": "failed", "After": "basic.target system.slice systemd-journald.socket quadlet-demo-mysql-volume.service quadlet-demo-network.service sysinit.target -.mount tmp.mount", "AllowIsolate": "no", "AssertResult": "yes", "AssertTimestamp": "Wed 2024-07-10 05:14:12 UTC", "AssertTimestampMonotonic": "973072307", "Before": "shutdown.target multi-user.target", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "CPUAccounting": "yes", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "3750354000", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "no", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore", "CleanResult": "success", "CollectMode": "inactive", "ConditionResult": "yes", "ConditionTimestamp": "Wed 2024-07-10 05:14:12 UTC", "ConditionTimestampMonotonic": "973072303", "ConfigurationDirectoryMode": "0755", "Conflicts": "shutdown.target", "ControlGroupId": "7462", "ControlPID": "0", "CoredumpFilter": "0x33", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "DefaultStartupMemoryLow": "0", "Delegate": "yes", "DelegateControllers": "cpu cpuset io memory pids", "Description": "quadlet-demo-mysql.service", "DevicePolicy": "auto", "DropInPaths": "/usr/lib/systemd/system/service.d/10-timeout-abort.conf", "DynamicUser": "no", "Environment": "PODMAN_SYSTEMD_UNIT=quadlet-demo-mysql.service", "ExecMainCode": "1", "ExecMainExitTimestamp": "Wed 2024-07-10 05:15:14 UTC", "ExecMainExitTimestampMonotonic": "1034735500", "ExecMainPID": "8040", "ExecMainStartTimestamp": "Wed 2024-07-10 05:14:13 UTC", "ExecMainStartTimestampMonotonic": "973531872", "ExecMainStatus": "137", "ExecStart": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman run --name=quadlet-demo-mysql --cidfile=/run/quadlet-demo-mysql.cid --replace --rm --cgroups=split --network=systemd-quadlet-demo --sdnotify=conmon -d -v systemd-quadlet-demo-mysql:/var/lib/mysql -v /tmp/quadlet_demo:/var/lib/quadlet_demo:Z --secret mysql-root-password-container,type=env,target=MYSQL_ROOT_PASSWORD --health-cmd /usr/bin/true --health-on-failure kill quay.io/linux-system-roles/mysql:5.6 ; ignore_errors=no ; start_time=[Wed 2024-07-10 05:14:13 UTC] ; stop_time=[Wed 2024-07-10 05:15:14 UTC] ; pid=8040 ; code=exited ; status=137 }", "ExecStartEx": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman run --name=quadlet-demo-mysql --cidfile=/run/quadlet-demo-mysql.cid --replace --rm --cgroups=split --network=systemd-quadlet-demo --sdnotify=conmon -d -v systemd-quadlet-demo-mysql:/var/lib/mysql -v /tmp/quadlet_demo:/var/lib/quadlet_demo:Z --secret mysql-root-password-container,type=env,target=MYSQL_ROOT_PASSWORD --health-cmd /usr/bin/true --health-on-failure kill quay.io/linux-system-roles/mysql:5.6 ; flags= ; start_time=[Wed 2024-07-10 05:14:13 UTC] ; stop_time=[Wed 2024-07-10 05:15:14 UTC] ; pid=8040 ; code=exited ; status=137 }", "ExecStop": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman rm -v -f -i --cidfile=/run/quadlet-demo-mysql.cid ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStopEx": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman rm -v -f -i --cidfile=/run/quadlet-demo-mysql.cid ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStopPost": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman rm -v -f -i --cidfile=/run/quadlet-demo-mysql.cid ; ignore_errors=yes ; start_time=[Wed 2024-07-10 05:15:14 UTC] ; stop_time=[Wed 2024-07-10 05:15:14 UTC] ; pid=11298 ; code=exited ; status=0 }", "ExecStopPostEx": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman rm -v -f -i --cidfile=/run/quadlet-demo-mysql.cid ; flags=ignore-failure ; start_time=[Wed 2024-07-10 05:15:14 UTC] ; stop_time=[Wed 2024-07-10 05:15:14 UTC] ; pid=11298 ; code=exited ; status=0 }", "ExitType": "main", "ExtensionImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FileDescriptorStorePreserve": "restart", "FinalKillSignal": "9", "FragmentPath": "/run/systemd/generator/quadlet-demo-mysql.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOReadBytes": "18446744073709551615", "IOReadOperations": "18446744073709551615", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "IOWeight": "[not set]", "IOWriteBytes": "18446744073709551615", "IOWriteOperations": "18446744073709551615", "IPAccounting": "no", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "Id": "quadlet-demo-mysql.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestamp": "Wed 2024-07-10 05:15:14 UTC", "InactiveEnterTimestampMonotonic": "1034788061", "InactiveExitTimestamp": "Wed 2024-07-10 05:14:12 UTC", "InactiveExitTimestampMonotonic": "973080495", "InvocationID": "16ff4258ca4845f99e8a96e4512592e3", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "mixed", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "infinity", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "8388608", "LimitMEMLOCKSoft": "8388608", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "524288", "LimitNOFILESoft": "1024", "LimitNPROC": "14756", "LimitNPROCSoft": "14756", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "14756", "LimitSIGPENDINGSoft": "14756", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "0", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "ManagedOOMSwap": "auto", "MemoryAccounting": "yes", "MemoryAvailable": "infinity", "MemoryCurrent": "[not set]", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryKSM": "no", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemoryPressureThresholdUSec": "200ms", "MemoryPressureWatch": "auto", "MemorySwapMax": "infinity", "MemoryZSwapMax": "infinity", "MountAPIVFS": "no", "MountImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAPolicy": "n/a", "Names": "quadlet-demo-mysql.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "all", "OOMPolicy": "continue", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "OnSuccessJobMode": "fail", "Perpetual": "no", "PrivateDevices": "no", "PrivateIPC": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProcSubset": "all", "ProtectClock": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectHostname": "no", "ProtectKernelLogs": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectProc": "default", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "ReloadResult": "success", "ReloadSignal": "1", "RemainAfterExit": "no", "RemoveIPC": "no", "Requires": "quadlet-demo-network.service -.mount sysinit.target system.slice quadlet-demo-mysql-volume.service tmp.mount", "RequiresMountsFor": "/tmp/quadlet_demo /run/containers", "Restart": "no", "RestartKillSignal": "15", "RestartMaxDelayUSec": "infinity", "RestartMode": "normal", "RestartSteps": "0", "RestartUSec": "100ms", "RestartUSecNext": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "exit-code", "RootDirectoryStartOnly": "no", "RootEphemeral": "no", "RootImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "Slice": "system.slice", "SourcePath": "/etc/containers/systemd/quadlet-demo-mysql.container", "StandardError": "inherit", "StandardInput": "null", "StandardOutput": "journal", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StartupMemoryHigh": "18446744073709551615", "StartupMemoryLow": "0", "StartupMemoryMax": "18446744073709551615", "StartupMemorySwapMax": "18446744073709551615", "StartupMemoryZSwapMax": "18446744073709551615", "StateChangeTimestamp": "Wed 2024-07-10 05:15:14 UTC", "StateChangeTimestampMonotonic": "1034788061", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "failed", "SuccessAction": "none", "SyslogFacility": "3", "SyslogIdentifier": "quadlet-demo-mysql", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "2147483646", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "[not set]", "TasksMax": "4426", "TimeoutAbortUSec": "45s", "TimeoutCleanUSec": "infinity", "TimeoutStartFailureMode": "terminate", "TimeoutStartUSec": "45s", "TimeoutStopFailureMode": "abort", "TimeoutStopUSec": "45s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "notify", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "disabled", "UnitFileState": "generated", "UtmpMode": "init", "WantedBy": "multi-user.target", "WatchdogSignal": "6", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "0" } } TASK [fedora.linux_system_roles.podman : Remove quadlet file] ****************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:33 Wednesday 10 July 2024 05:15:19 +0000 (0:00:00.837) 0:03:23.952 ******** changed: [sut] => { "changed": true, "path": "/etc/containers/systemd/quadlet-demo-mysql.container", "state": "absent" } TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:39 Wednesday 10 July 2024 05:15:19 +0000 (0:00:00.240) 0:03:24.192 ******** included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for sut TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Wednesday 10 July 2024 05:15:19 +0000 (0:00:00.034) 0:03:24.226 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Wednesday 10 July 2024 05:15:19 +0000 (0:00:00.062) 0:03:24.289 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Wednesday 10 July 2024 05:15:19 +0000 (0:00:00.022) 0:03:24.311 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Reload systemctl] ********************* task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:47 Wednesday 10 July 2024 05:15:19 +0000 (0:00:00.020) 0:03:24.332 ******** ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.podman : Prune images no longer in use] ******** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:56 Wednesday 10 July 2024 05:15:20 +0000 (0:00:00.878) 0:03:25.210 ******** changed: [sut] => { "changed": true, "cmd": [ "podman", "image", "prune", "--all", "-f" ], "delta": "0:00:00.229999", "end": "2024-07-10 05:15:20.974256", "rc": 0, "start": "2024-07-10 05:15:20.744257" } STDOUT: dd3b2a5dcb48ff61113592ed5ddd762581be4387c7bc552375a2159422aa6bf5 TASK [fedora.linux_system_roles.podman : Remove networks no longer in use] ***** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:64 Wednesday 10 July 2024 05:15:21 +0000 (0:00:00.466) 0:03:25.677 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_quadlet_type == \"network\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Remove volumes no longer in use] ****** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:73 Wednesday 10 July 2024 05:15:21 +0000 (0:00:00.022) 0:03:25.699 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_quadlet_type == \"volume\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : For testing and debugging - images] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:82 Wednesday 10 July 2024 05:15:21 +0000 (0:00:00.022) 0:03:25.722 ******** ok: [sut] => { "changed": false, "cmd": [ "podman", "images", "-n" ], "delta": "0:00:00.036569", "end": "2024-07-10 05:15:21.294266", "rc": 0, "start": "2024-07-10 05:15:21.257697" } TASK [fedora.linux_system_roles.podman : For testing and debugging - volumes] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:92 Wednesday 10 July 2024 05:15:21 +0000 (0:00:00.275) 0:03:25.998 ******** ok: [sut] => { "changed": false, "cmd": [ "podman", "volume", "ls", "-n" ], "delta": "0:00:00.036158", "end": "2024-07-10 05:15:21.569057", "rc": 0, "start": "2024-07-10 05:15:21.532899" } STDOUT: local systemd-quadlet-demo-mysql local wp-pv-claim local envoy-proxy-config local envoy-certificates TASK [fedora.linux_system_roles.podman : For testing and debugging - containers] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:102 Wednesday 10 July 2024 05:15:21 +0000 (0:00:00.274) 0:03:26.272 ******** ok: [sut] => { "changed": false, "cmd": [ "podman", "ps", "--noheading" ], "delta": "0:00:00.036024", "end": "2024-07-10 05:15:21.845993", "rc": 0, "start": "2024-07-10 05:15:21.809969" } TASK [fedora.linux_system_roles.podman : For testing and debugging - networks] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:112 Wednesday 10 July 2024 05:15:21 +0000 (0:00:00.274) 0:03:26.547 ******** ok: [sut] => { "changed": false, "cmd": [ "podman", "network", "ls", "-n", "-q" ], "delta": "0:00:00.035355", "end": "2024-07-10 05:15:22.120706", "rc": 0, "start": "2024-07-10 05:15:22.085351" } STDOUT: podman systemd-quadlet-demo TASK [fedora.linux_system_roles.podman : For testing and debugging - secrets] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:122 Wednesday 10 July 2024 05:15:22 +0000 (0:00:00.276) 0:03:26.823 ******** ok: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : For testing and debugging - services] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:133 Wednesday 10 July 2024 05:15:22 +0000 (0:00:00.267) 0:03:27.090 ******** ok: [sut] => { "ansible_facts": { "services": { "NetworkManager-dispatcher.service": { "name": "NetworkManager-dispatcher.service", "source": "systemd", "state": "running", "status": "enabled" }, "NetworkManager-wait-online.service": { "name": "NetworkManager-wait-online.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "NetworkManager.service": { "name": "NetworkManager.service", "source": "systemd", "state": "running", "status": "enabled" }, "arp-ethers.service": { "name": "arp-ethers.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "auditd.service": { "name": "auditd.service", "source": "systemd", "state": "running", "status": "enabled" }, "auth-rpcgss-module.service": { "name": "auth-rpcgss-module.service", "source": "systemd", "state": "stopped", "status": "static" }, "autovt@.service": { "name": "autovt@.service", "source": "systemd", "state": "unknown", "status": "alias" }, "blk-availability.service": { "name": "blk-availability.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "bluetooth.service": { "name": "bluetooth.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "certmonger.service": { "name": "certmonger.service", "source": "systemd", "state": "running", "status": "enabled" }, "chrony-wait.service": { "name": "chrony-wait.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd-restricted.service": { "name": "chronyd-restricted.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd.service": { "name": "chronyd.service", "source": "systemd", "state": "running", "status": "enabled" }, "cloud-config.service": { "name": "cloud-config.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-final.service": { "name": "cloud-final.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init-hotplugd.service": { "name": "cloud-init-hotplugd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cloud-init-local.service": { "name": "cloud-init-local.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init.service": { "name": "cloud-init.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "console-getty.service": { "name": "console-getty.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "container-getty@.service": { "name": "container-getty@.service", "source": "systemd", "state": "unknown", "status": "static" }, "dbus-broker.service": { "name": "dbus-broker.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-org.bluez.service": { "name": "dbus-org.bluez.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.fedoraproject.FirewallD1.service": { "name": "dbus-org.fedoraproject.FirewallD1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.home1.service": { "name": "dbus-org.freedesktop.home1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.hostname1.service": { "name": "dbus-org.freedesktop.hostname1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.locale1.service": { "name": "dbus-org.freedesktop.locale1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.login1.service": { "name": "dbus-org.freedesktop.login1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.nm-dispatcher.service": { "name": "dbus-org.freedesktop.nm-dispatcher.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.oom1.service": { "name": "dbus-org.freedesktop.oom1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.portable1.service": { "name": "dbus-org.freedesktop.portable1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.resolve1.service": { "name": "dbus-org.freedesktop.resolve1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.timedate1.service": { "name": "dbus-org.freedesktop.timedate1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus.service": { "name": "dbus.service", "source": "systemd", "state": "active", "status": "alias" }, "debug-shell.service": { "name": "debug-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "display-manager.service": { "name": "display-manager.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "dm-event.service": { "name": "dm-event.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-makecache.service": { "name": "dnf-makecache.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-system-upgrade-cleanup.service": { "name": "dnf-system-upgrade-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "dnf-system-upgrade.service": { "name": "dnf-system-upgrade.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dracut-cmdline.service": { "name": "dracut-cmdline.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-initqueue.service": { "name": "dracut-initqueue.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-mount.service": { "name": "dracut-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-mount.service": { "name": "dracut-pre-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-pivot.service": { "name": "dracut-pre-pivot.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-trigger.service": { "name": "dracut-pre-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-udev.service": { "name": "dracut-pre-udev.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown-onfailure.service": { "name": "dracut-shutdown-onfailure.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown.service": { "name": "dracut-shutdown.service", "source": "systemd", "state": "stopped", "status": "static" }, "ebtables.service": { "name": "ebtables.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "emergency.service": { "name": "emergency.service", "source": "systemd", "state": "stopped", "status": "static" }, "firewalld.service": { "name": "firewalld.service", "source": "systemd", "state": "running", "status": "enabled" }, "fsidd.service": { "name": "fsidd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "fstrim.service": { "name": "fstrim.service", "source": "systemd", "state": "stopped", "status": "static" }, "fwupd-offline-update.service": { "name": "fwupd-offline-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd-refresh.service": { "name": "fwupd-refresh.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd.service": { "name": "fwupd.service", "source": "systemd", "state": "inactive", "status": "static" }, "getty@.service": { "name": "getty@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "getty@tty1.service": { "name": "getty@tty1.service", "source": "systemd", "state": "running", "status": "active" }, "grub-boot-indeterminate.service": { "name": "grub-boot-indeterminate.service", "source": "systemd", "state": "inactive", "status": "static" }, "grub2-systemd-integration.service": { "name": "grub2-systemd-integration.service", "source": "systemd", "state": "inactive", "status": "static" }, "gssproxy.service": { "name": "gssproxy.service", "source": "systemd", "state": "running", "status": "disabled" }, "hv_kvp_daemon.service": { "name": "hv_kvp_daemon.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "import-state.service": { "name": "import-state.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "initrd-cleanup.service": { "name": "initrd-cleanup.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-parse-etc.service": { "name": "initrd-parse-etc.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-switch-root.service": { "name": "initrd-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-udevadm-cleanup-db.service": { "name": "initrd-udevadm-cleanup-db.service", "source": "systemd", "state": "stopped", "status": "static" }, "ip6tables.service": { "name": "ip6tables.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ipset.service": { "name": "ipset.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "iptables.service": { "name": "iptables.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "kmod-static-nodes.service": { "name": "kmod-static-nodes.service", "source": "systemd", "state": "stopped", "status": "static" }, "ldconfig.service": { "name": "ldconfig.service", "source": "systemd", "state": "stopped", "status": "static" }, "loadmodules.service": { "name": "loadmodules.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "lvm2-lvmpolld.service": { "name": "lvm2-lvmpolld.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-monitor.service": { "name": "lvm2-monitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "man-db-cache-update.service": { "name": "man-db-cache-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "man-db-restart-cache-update.service": { "name": "man-db-restart-cache-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "mdadm-grow-continue@.service": { "name": "mdadm-grow-continue@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdadm-last-resort@.service": { "name": "mdadm-last-resort@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdcheck_continue.service": { "name": "mdcheck_continue.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdcheck_start.service": { "name": "mdcheck_start.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmon@.service": { "name": "mdmon@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdmonitor-oneshot.service": { "name": "mdmonitor-oneshot.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmonitor.service": { "name": "mdmonitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "modprobe@.service": { "name": "modprobe@.service", "source": "systemd", "state": "unknown", "status": "static" }, "modprobe@configfs.service": { "name": "modprobe@configfs.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "modprobe@dm_mod.service": { "name": "modprobe@dm_mod.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "modprobe@drm.service": { "name": "modprobe@drm.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "modprobe@efi_pstore.service": { "name": "modprobe@efi_pstore.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "modprobe@fuse.service": { "name": "modprobe@fuse.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "modprobe@loop.service": { "name": "modprobe@loop.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "netavark-dhcp-proxy.service": { "name": "netavark-dhcp-proxy.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "netavark-firewalld-reload.service": { "name": "netavark-firewalld-reload.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "network.service": { "name": "network.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "nfs-blkmap.service": { "name": "nfs-blkmap.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-idmapd.service": { "name": "nfs-idmapd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-mountd.service": { "name": "nfs-mountd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-server.service": { "name": "nfs-server.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-utils.service": { "name": "nfs-utils.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfsdcld.service": { "name": "nfsdcld.service", "source": "systemd", "state": "stopped", "status": "static" }, "nftables.service": { "name": "nftables.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nis-domainname.service": { "name": "nis-domainname.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nm-priv-helper.service": { "name": "nm-priv-helper.service", "source": "systemd", "state": "inactive", "status": "static" }, "ntpd.service": { "name": "ntpd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ntpdate.service": { "name": "ntpdate.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "pam_namespace.service": { "name": "pam_namespace.service", "source": "systemd", "state": "inactive", "status": "static" }, "pcscd.service": { "name": "pcscd.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "plymouth-halt.service": { "name": "plymouth-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-kexec.service": { "name": "plymouth-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-poweroff.service": { "name": "plymouth-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-quit-wait.service": { "name": "plymouth-quit-wait.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-quit.service": { "name": "plymouth-quit.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-read-write.service": { "name": "plymouth-read-write.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-reboot.service": { "name": "plymouth-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-start.service": { "name": "plymouth-start.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-switch-root-initramfs.service": { "name": "plymouth-switch-root-initramfs.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-switch-root.service": { "name": "plymouth-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "podman-auto-update.service": { "name": "podman-auto-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "podman-clean-transient.service": { "name": "podman-clean-transient.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "podman-kube@.service": { "name": "podman-kube@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "podman-restart.service": { "name": "podman-restart.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "podman.service": { "name": "podman.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "polkit.service": { "name": "polkit.service", "source": "systemd", "state": "running", "status": "static" }, "quadlet-demo-mysql-volume.service": { "name": "quadlet-demo-mysql-volume.service", "source": "systemd", "state": "stopped", "status": "generated" }, "quadlet-demo-mysql.service": { "name": "quadlet-demo-mysql.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "quadlet-demo-network.service": { "name": "quadlet-demo-network.service", "source": "systemd", "state": "stopped", "status": "generated" }, "quotaon.service": { "name": "quotaon.service", "source": "systemd", "state": "inactive", "status": "static" }, "raid-check.service": { "name": "raid-check.service", "source": "systemd", "state": "stopped", "status": "static" }, "rc-local.service": { "name": "rc-local.service", "source": "systemd", "state": "stopped", "status": "static" }, "rescue.service": { "name": "rescue.service", "source": "systemd", "state": "stopped", "status": "static" }, "restraintd.service": { "name": "restraintd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rngd.service": { "name": "rngd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpc-gssd.service": { "name": "rpc-gssd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd-notify.service": { "name": "rpc-statd-notify.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd.service": { "name": "rpc-statd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-svcgssd.service": { "name": "rpc-svcgssd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "rpcbind.service": { "name": "rpcbind.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "rpmdb-migrate.service": { "name": "rpmdb-migrate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rpmdb-rebuild.service": { "name": "rpmdb-rebuild.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "selinux-autorelabel-mark.service": { "name": "selinux-autorelabel-mark.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "selinux-autorelabel.service": { "name": "selinux-autorelabel.service", "source": "systemd", "state": "inactive", "status": "static" }, "selinux-check-proper-disable.service": { "name": "selinux-check-proper-disable.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "serial-getty@.service": { "name": "serial-getty@.service", "source": "systemd", "state": "unknown", "status": "indirect" }, "serial-getty@ttyS0.service": { "name": "serial-getty@ttyS0.service", "source": "systemd", "state": "running", "status": "active" }, "snapd.seeded.service": { "name": "snapd.seeded.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sntp.service": { "name": "sntp.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ssh-host-keys-migration.service": { "name": "ssh-host-keys-migration.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "sshd-keygen.service": { "name": "sshd-keygen.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sshd-keygen@.service": { "name": "sshd-keygen@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@ecdsa.service": { "name": "sshd-keygen@ecdsa.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd-keygen@ed25519.service": { "name": "sshd-keygen@ed25519.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd-keygen@rsa.service": { "name": "sshd-keygen@rsa.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd.service": { "name": "sshd.service", "source": "systemd", "state": "running", "status": "enabled" }, "sshd@.service": { "name": "sshd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "sssd-autofs.service": { "name": "sssd-autofs.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-kcm.service": { "name": "sssd-kcm.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "sssd-nss.service": { "name": "sssd-nss.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pac.service": { "name": "sssd-pac.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pam.service": { "name": "sssd-pam.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-ssh.service": { "name": "sssd-ssh.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-sudo.service": { "name": "sssd-sudo.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd.service": { "name": "sssd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "syslog.service": { "name": "syslog.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "system-update-cleanup.service": { "name": "system-update-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-ask-password-console.service": { "name": "systemd-ask-password-console.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-plymouth.service": { "name": "systemd-ask-password-plymouth.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-wall.service": { "name": "systemd-ask-password-wall.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-backlight@.service": { "name": "systemd-backlight@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-battery-check.service": { "name": "systemd-battery-check.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-binfmt.service": { "name": "systemd-binfmt.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-bless-boot.service": { "name": "systemd-bless-boot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-boot-check-no-failures.service": { "name": "systemd-boot-check-no-failures.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-boot-random-seed.service": { "name": "systemd-boot-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-boot-update.service": { "name": "systemd-boot-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-confext.service": { "name": "systemd-confext.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-coredump@.service": { "name": "systemd-coredump@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-exit.service": { "name": "systemd-exit.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-firstboot.service": { "name": "systemd-firstboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck-root.service": { "name": "systemd-fsck-root.service", "source": "systemd", "state": "stopped", "status": "enabled-runtime" }, "systemd-fsck@.service": { "name": "systemd-fsck@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-growfs-root.service": { "name": "systemd-growfs-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-growfs@.service": { "name": "systemd-growfs@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-halt.service": { "name": "systemd-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hibernate.service": { "name": "systemd-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-homed-activate.service": { "name": "systemd-homed-activate.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-homed.service": { "name": "systemd-homed.service", "source": "systemd", "state": "running", "status": "enabled" }, "systemd-hostnamed.service": { "name": "systemd-hostnamed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hwdb-update.service": { "name": "systemd-hwdb-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-hybrid-sleep.service": { "name": "systemd-hybrid-sleep.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-initctl.service": { "name": "systemd-initctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-catalog-update.service": { "name": "systemd-journal-catalog-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-flush.service": { "name": "systemd-journal-flush.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journald.service": { "name": "systemd-journald.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-journald@.service": { "name": "systemd-journald@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-kexec.service": { "name": "systemd-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-localed.service": { "name": "systemd-localed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-logind.service": { "name": "systemd-logind.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-machine-id-commit.service": { "name": "systemd-machine-id-commit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-modules-load.service": { "name": "systemd-modules-load.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-network-generator.service": { "name": "systemd-network-generator.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-networkd-wait-online.service": { "name": "systemd-networkd-wait-online.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "systemd-networkd-wait-online@.service": { "name": "systemd-networkd-wait-online@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "systemd-networkd.service": { "name": "systemd-networkd.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "systemd-oomd.service": { "name": "systemd-oomd.service", "source": "systemd", "state": "running", "status": "enabled" }, "systemd-pcrfs-root.service": { "name": "systemd-pcrfs-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pcrfs@.service": { "name": "systemd-pcrfs@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-pcrmachine.service": { "name": "systemd-pcrmachine.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-pcrphase-initrd.service": { "name": "systemd-pcrphase-initrd.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-pcrphase-sysinit.service": { "name": "systemd-pcrphase-sysinit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-pcrphase.service": { "name": "systemd-pcrphase.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-portabled.service": { "name": "systemd-portabled.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-poweroff.service": { "name": "systemd-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pstore.service": { "name": "systemd-pstore.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-quotacheck.service": { "name": "systemd-quotacheck.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-random-seed.service": { "name": "systemd-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-reboot.service": { "name": "systemd-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-remount-fs.service": { "name": "systemd-remount-fs.service", "source": "systemd", "state": "stopped", "status": "enabled-runtime" }, "systemd-repart.service": { "name": "systemd-repart.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-resolved.service": { "name": "systemd-resolved.service", "source": "systemd", "state": "running", "status": "enabled" }, "systemd-rfkill.service": { "name": "systemd-rfkill.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-soft-reboot.service": { "name": "systemd-soft-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend-then-hibernate.service": { "name": "systemd-suspend-then-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend.service": { "name": "systemd-suspend.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-sysctl.service": { "name": "systemd-sysctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-sysext.service": { "name": "systemd-sysext.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "systemd-sysupdate-reboot.service": { "name": "systemd-sysupdate-reboot.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "systemd-sysupdate.service": { "name": "systemd-sysupdate.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "systemd-sysusers.service": { "name": "systemd-sysusers.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-time-wait-sync.service": { "name": "systemd-time-wait-sync.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-timedated.service": { "name": "systemd-timedated.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-timesyncd.service": { "name": "systemd-timesyncd.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "systemd-tmpfiles-clean.service": { "name": "systemd-tmpfiles-clean.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev-early.service": { "name": "systemd-tmpfiles-setup-dev-early.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev.service": { "name": "systemd-tmpfiles-setup-dev.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup.service": { "name": "systemd-tmpfiles-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-settle.service": { "name": "systemd-udev-settle.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-trigger.service": { "name": "systemd-udev-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udevd.service": { "name": "systemd-udevd.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-update-done.service": { "name": "systemd-update-done.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp-runlevel.service": { "name": "systemd-update-utmp-runlevel.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp.service": { "name": "systemd-update-utmp.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-user-sessions.service": { "name": "systemd-user-sessions.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-userdbd.service": { "name": "systemd-userdbd.service", "source": "systemd", "state": "running", "status": "indirect" }, "systemd-vconsole-setup.service": { "name": "systemd-vconsole-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-volatile-root.service": { "name": "systemd-volatile-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-zram-setup@.service": { "name": "systemd-zram-setup@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-zram-setup@zram0.service": { "name": "systemd-zram-setup@zram0.service", "source": "systemd", "state": "stopped", "status": "active" }, "udisks2.service": { "name": "udisks2.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "unbound-anchor.service": { "name": "unbound-anchor.service", "source": "systemd", "state": "stopped", "status": "static" }, "user-runtime-dir@.service": { "name": "user-runtime-dir@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user-runtime-dir@0.service": { "name": "user-runtime-dir@0.service", "source": "systemd", "state": "stopped", "status": "active" }, "user@.service": { "name": "user@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user@0.service": { "name": "user@0.service", "source": "systemd", "state": "running", "status": "active" } } }, "changed": false } TASK [fedora.linux_system_roles.podman : Create and update quadlets] *********** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:114 Wednesday 10 July 2024 05:15:24 +0000 (0:00:02.186) 0:03:29.277 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_state != \"absent\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14 Wednesday 10 July 2024 05:15:24 +0000 (0:00:00.020) 0:03:29.297 ******** ok: [sut] => { "ansible_facts": { "__podman_quadlet_file_src": "quadlet-demo-mysql.volume", "__podman_quadlet_spec": {}, "__podman_quadlet_str": "[Volume]", "__podman_quadlet_template_src": "" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25 Wednesday 10 July 2024 05:15:24 +0000 (0:00:00.073) 0:03:29.371 ******** ok: [sut] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_pull_image": true, "__podman_state": "absent", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] ***** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35 Wednesday 10 July 2024 05:15:24 +0000 (0:00:00.028) 0:03:29.399 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_quadlet_file_src", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48 Wednesday 10 July 2024 05:15:24 +0000 (0:00:00.023) 0:03:29.422 ******** ok: [sut] => { "ansible_facts": { "__podman_quadlet_name": "quadlet-demo-mysql", "__podman_quadlet_type": "volume", "__podman_rootless": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57 Wednesday 10 July 2024 05:15:24 +0000 (0:00:00.032) 0:03:29.455 ******** included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for sut TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Wednesday 10 July 2024 05:15:24 +0000 (0:00:00.041) 0:03:29.497 ******** skipping: [sut] => { "changed": false, "false_condition": "'getent_passwd' not in ansible_facts or __podman_user not in ansible_facts['getent_passwd']", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Wednesday 10 July 2024 05:15:24 +0000 (0:00:00.024) 0:03:29.521 ******** skipping: [sut] => { "changed": false, "false_condition": "not ansible_facts[\"getent_passwd\"][__podman_user]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Wednesday 10 July 2024 05:15:24 +0000 (0:00:00.024) 0:03:29.546 ******** ok: [sut] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get group information] **************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:28 Wednesday 10 July 2024 05:15:24 +0000 (0:00:00.031) 0:03:29.578 ******** ok: [sut] => { "ansible_facts": { "getent_group": { "root": [ "x", "0", "" ] } }, "changed": false } TASK [fedora.linux_system_roles.podman : Set group name] *********************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:35 Wednesday 10 July 2024 05:15:25 +0000 (0:00:00.239) 0:03:29.818 ******** ok: [sut] => { "ansible_facts": { "__podman_group_name": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:39 Wednesday 10 July 2024 05:15:25 +0000 (0:00:00.029) 0:03:29.847 ******** ok: [sut] => { "changed": false, "stat": { "atime": 1720588427.5164855, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "72ba28a3be19c467e89ab90f5d21f99b605159a7", "ctime": 1720588377.3007717, "dev": 51714, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 203603, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-pie-executable", "mode": "0755", "mtime": 1696291200.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 16032, "uid": 0, "version": "2092964490", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************ task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:50 Wednesday 10 July 2024 05:15:25 +0000 (0:00:00.235) 0:03:30.082 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check group with getsubids] *********** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:55 Wednesday 10 July 2024 05:15:25 +0000 (0:00:00.020) 0:03:30.103 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:60 Wednesday 10 July 2024 05:15:25 +0000 (0:00:00.020) 0:03:30.123 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subuid file] ********************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:74 Wednesday 10 July 2024 05:15:25 +0000 (0:00:00.019) 0:03:30.143 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subgid file] ********************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:79 Wednesday 10 July 2024 05:15:25 +0000 (0:00:00.058) 0:03:30.201 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:84 Wednesday 10 July 2024 05:15:25 +0000 (0:00:00.022) 0:03:30.224 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:94 Wednesday 10 July 2024 05:15:25 +0000 (0:00:00.020) 0:03:30.244 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] ***** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:101 Wednesday 10 July 2024 05:15:25 +0000 (0:00:00.020) 0:03:30.265 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62 Wednesday 10 July 2024 05:15:25 +0000 (0:00:00.020) 0:03:30.286 ******** ok: [sut] => { "ansible_facts": { "__podman_activate_systemd_unit": true, "__podman_images_found": [], "__podman_kube_yamls_raw": "", "__podman_service_name": "quadlet-demo-mysql-volume.service", "__podman_systemd_scope": "system", "__podman_user_home_dir": "/root", "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73 Wednesday 10 July 2024 05:15:25 +0000 (0:00:00.038) 0:03:30.325 ******** ok: [sut] => { "ansible_facts": { "__podman_quadlet_path": "/etc/containers/systemd" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get kube yaml contents] *************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77 Wednesday 10 July 2024 05:15:25 +0000 (0:00:00.022) 0:03:30.348 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_state != \"absent\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87 Wednesday 10 July 2024 05:15:25 +0000 (0:00:00.020) 0:03:30.369 ******** ok: [sut] => { "ansible_facts": { "__podman_images": [], "__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo-mysql.volume", "__podman_volumes": [] }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 6] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:103 Wednesday 10 July 2024 05:15:25 +0000 (0:00:00.049) 0:03:30.418 ******** ok: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Cleanup quadlets] ********************* task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:110 Wednesday 10 July 2024 05:15:25 +0000 (0:00:00.026) 0:03:30.445 ******** included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml for sut TASK [fedora.linux_system_roles.podman : Stat XDG_RUNTIME_DIR] ***************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:4 Wednesday 10 July 2024 05:15:25 +0000 (0:00:00.046) 0:03:30.492 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Stop and disable service] ************* task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:12 Wednesday 10 July 2024 05:15:25 +0000 (0:00:00.020) 0:03:30.512 ******** changed: [sut] => { "changed": true, "enabled": false, "failed_when_result": false, "name": "quadlet-demo-mysql-volume.service", "state": "stopped", "status": { "AccessSELinuxContext": "system_u:object_r:systemd_unit_file_t:s0", "ActiveEnterTimestamp": "Wed 2024-07-10 05:14:02 UTC", "ActiveEnterTimestampMonotonic": "962640855", "ActiveExitTimestampMonotonic": "0", "ActiveState": "active", "After": "-.mount system.slice sysinit.target systemd-journald.socket basic.target", "AllowIsolate": "no", "AssertResult": "yes", "AssertTimestamp": "Wed 2024-07-10 05:14:02 UTC", "AssertTimestampMonotonic": "962587608", "Before": "shutdown.target", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "CPUAccounting": "yes", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "38624000", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "no", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore", "CleanResult": "success", "CollectMode": "inactive", "ConditionResult": "yes", "ConditionTimestamp": "Wed 2024-07-10 05:14:02 UTC", "ConditionTimestampMonotonic": "962587604", "ConfigurationDirectoryMode": "0755", "Conflicts": "shutdown.target", "ControlGroupId": "0", "ControlPID": "0", "CoredumpFilter": "0x33", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "DefaultStartupMemoryLow": "0", "Delegate": "no", "Description": "quadlet-demo-mysql-volume.service", "DevicePolicy": "auto", "DropInPaths": "/usr/lib/systemd/system/service.d/10-timeout-abort.conf", "DynamicUser": "no", "ExecMainCode": "1", "ExecMainExitTimestamp": "Wed 2024-07-10 05:14:02 UTC", "ExecMainExitTimestampMonotonic": "962640687", "ExecMainPID": "7517", "ExecMainStartTimestamp": "Wed 2024-07-10 05:14:02 UTC", "ExecMainStartTimestampMonotonic": "962595096", "ExecMainStatus": "0", "ExecStart": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman volume create --ignore systemd-quadlet-demo-mysql ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStartEx": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman volume create --ignore systemd-quadlet-demo-mysql ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExitType": "main", "ExtensionImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FileDescriptorStorePreserve": "restart", "FinalKillSignal": "9", "FragmentPath": "/run/systemd/generator/quadlet-demo-mysql-volume.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOReadBytes": "18446744073709551615", "IOReadOperations": "18446744073709551615", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "IOWeight": "[not set]", "IOWriteBytes": "18446744073709551615", "IOWriteOperations": "18446744073709551615", "IPAccounting": "no", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "Id": "quadlet-demo-mysql-volume.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestamp": "Wed 2024-07-10 05:14:02 UTC", "InactiveExitTimestampMonotonic": "962595461", "InvocationID": "816a65673bcb4a1b9babf8b8555173f9", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "control-group", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "infinity", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "8388608", "LimitMEMLOCKSoft": "8388608", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "524288", "LimitNOFILESoft": "1024", "LimitNPROC": "14756", "LimitNPROCSoft": "14756", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "14756", "LimitSIGPENDINGSoft": "14756", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "0", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "ManagedOOMSwap": "auto", "MemoryAccounting": "yes", "MemoryAvailable": "infinity", "MemoryCurrent": "[not set]", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryKSM": "no", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemoryPressureThresholdUSec": "200ms", "MemoryPressureWatch": "auto", "MemorySwapMax": "infinity", "MemoryZSwapMax": "infinity", "MountAPIVFS": "no", "MountImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAPolicy": "n/a", "Names": "quadlet-demo-mysql-volume.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMPolicy": "stop", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "OnSuccessJobMode": "fail", "Perpetual": "no", "PrivateDevices": "no", "PrivateIPC": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProcSubset": "all", "ProtectClock": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectHostname": "no", "ProtectKernelLogs": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectProc": "default", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "ReloadResult": "success", "ReloadSignal": "1", "RemainAfterExit": "yes", "RemoveIPC": "no", "Requires": "system.slice sysinit.target -.mount", "RequiresMountsFor": "/run/containers", "Restart": "no", "RestartKillSignal": "15", "RestartMaxDelayUSec": "infinity", "RestartMode": "normal", "RestartSteps": "0", "RestartUSec": "100ms", "RestartUSecNext": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RootEphemeral": "no", "RootImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "Slice": "system.slice", "StandardError": "inherit", "StandardInput": "null", "StandardOutput": "journal", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StartupMemoryHigh": "18446744073709551615", "StartupMemoryLow": "0", "StartupMemoryMax": "18446744073709551615", "StartupMemorySwapMax": "18446744073709551615", "StartupMemoryZSwapMax": "18446744073709551615", "StateChangeTimestamp": "Wed 2024-07-10 05:14:02 UTC", "StateChangeTimestampMonotonic": "962640855", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "exited", "SuccessAction": "none", "SyslogFacility": "3", "SyslogIdentifier": "quadlet-demo-mysql-volume", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "2147483646", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "[not set]", "TasksMax": "4426", "TimeoutAbortUSec": "45s", "TimeoutCleanUSec": "infinity", "TimeoutStartFailureMode": "terminate", "TimeoutStartUSec": "infinity", "TimeoutStopFailureMode": "abort", "TimeoutStopUSec": "45s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "oneshot", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "disabled", "UnitFileState": "generated", "UtmpMode": "init", "WatchdogSignal": "6", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "0" } } TASK [fedora.linux_system_roles.podman : Remove quadlet file] ****************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:33 Wednesday 10 July 2024 05:15:26 +0000 (0:00:00.843) 0:03:31.356 ******** changed: [sut] => { "changed": true, "path": "/etc/containers/systemd/quadlet-demo-mysql.volume", "state": "absent" } TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:39 Wednesday 10 July 2024 05:15:26 +0000 (0:00:00.279) 0:03:31.636 ******** included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for sut TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Wednesday 10 July 2024 05:15:27 +0000 (0:00:00.035) 0:03:31.671 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Wednesday 10 July 2024 05:15:27 +0000 (0:00:00.021) 0:03:31.692 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Wednesday 10 July 2024 05:15:27 +0000 (0:00:00.020) 0:03:31.713 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Reload systemctl] ********************* task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:47 Wednesday 10 July 2024 05:15:27 +0000 (0:00:00.020) 0:03:31.734 ******** ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.podman : Prune images no longer in use] ******** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:56 Wednesday 10 July 2024 05:15:27 +0000 (0:00:00.786) 0:03:32.521 ******** changed: [sut] => { "changed": true, "cmd": [ "podman", "image", "prune", "--all", "-f" ], "delta": "0:00:00.033872", "end": "2024-07-10 05:15:28.088749", "rc": 0, "start": "2024-07-10 05:15:28.054877" } TASK [fedora.linux_system_roles.podman : Remove networks no longer in use] ***** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:64 Wednesday 10 July 2024 05:15:28 +0000 (0:00:00.272) 0:03:32.793 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_quadlet_type == \"network\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Remove volumes no longer in use] ****** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:73 Wednesday 10 July 2024 05:15:28 +0000 (0:00:00.023) 0:03:32.816 ******** changed: [sut] => { "changed": true, "cmd": [ "podman", "volume", "prune", "-f" ], "delta": "0:00:00.131415", "end": "2024-07-10 05:15:28.481581", "rc": 0, "start": "2024-07-10 05:15:28.350166" } STDOUT: systemd-quadlet-demo-mysql wp-pv-claim envoy-proxy-config envoy-certificates TASK [fedora.linux_system_roles.podman : For testing and debugging - images] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:82 Wednesday 10 July 2024 05:15:28 +0000 (0:00:00.365) 0:03:33.182 ******** ok: [sut] => { "changed": false, "cmd": [ "podman", "images", "-n" ], "delta": "0:00:00.035502", "end": "2024-07-10 05:15:28.754719", "rc": 0, "start": "2024-07-10 05:15:28.719217" } TASK [fedora.linux_system_roles.podman : For testing and debugging - volumes] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:92 Wednesday 10 July 2024 05:15:28 +0000 (0:00:00.274) 0:03:33.456 ******** ok: [sut] => { "changed": false, "cmd": [ "podman", "volume", "ls", "-n" ], "delta": "0:00:00.031639", "end": "2024-07-10 05:15:29.022810", "rc": 0, "start": "2024-07-10 05:15:28.991171" } TASK [fedora.linux_system_roles.podman : For testing and debugging - containers] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:102 Wednesday 10 July 2024 05:15:29 +0000 (0:00:00.269) 0:03:33.726 ******** ok: [sut] => { "changed": false, "cmd": [ "podman", "ps", "--noheading" ], "delta": "0:00:00.036405", "end": "2024-07-10 05:15:29.297830", "rc": 0, "start": "2024-07-10 05:15:29.261425" } TASK [fedora.linux_system_roles.podman : For testing and debugging - networks] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:112 Wednesday 10 July 2024 05:15:29 +0000 (0:00:00.273) 0:03:33.999 ******** ok: [sut] => { "changed": false, "cmd": [ "podman", "network", "ls", "-n", "-q" ], "delta": "0:00:00.037909", "end": "2024-07-10 05:15:29.574206", "rc": 0, "start": "2024-07-10 05:15:29.536297" } STDOUT: podman systemd-quadlet-demo TASK [fedora.linux_system_roles.podman : For testing and debugging - secrets] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:122 Wednesday 10 July 2024 05:15:29 +0000 (0:00:00.276) 0:03:34.276 ******** ok: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : For testing and debugging - services] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:133 Wednesday 10 July 2024 05:15:29 +0000 (0:00:00.271) 0:03:34.547 ******** ok: [sut] => { "ansible_facts": { "services": { "NetworkManager-dispatcher.service": { "name": "NetworkManager-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "NetworkManager-wait-online.service": { "name": "NetworkManager-wait-online.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "NetworkManager.service": { "name": "NetworkManager.service", "source": "systemd", "state": "running", "status": "enabled" }, "arp-ethers.service": { "name": "arp-ethers.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "auditd.service": { "name": "auditd.service", "source": "systemd", "state": "running", "status": "enabled" }, "auth-rpcgss-module.service": { "name": "auth-rpcgss-module.service", "source": "systemd", "state": "stopped", "status": "static" }, "autovt@.service": { "name": "autovt@.service", "source": "systemd", "state": "unknown", "status": "alias" }, "blk-availability.service": { "name": "blk-availability.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "bluetooth.service": { "name": "bluetooth.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "certmonger.service": { "name": "certmonger.service", "source": "systemd", "state": "running", "status": "enabled" }, "chrony-wait.service": { "name": "chrony-wait.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd-restricted.service": { "name": "chronyd-restricted.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd.service": { "name": "chronyd.service", "source": "systemd", "state": "running", "status": "enabled" }, "cloud-config.service": { "name": "cloud-config.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-final.service": { "name": "cloud-final.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init-hotplugd.service": { "name": "cloud-init-hotplugd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cloud-init-local.service": { "name": "cloud-init-local.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init.service": { "name": "cloud-init.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "console-getty.service": { "name": "console-getty.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "container-getty@.service": { "name": "container-getty@.service", "source": "systemd", "state": "unknown", "status": "static" }, "dbus-broker.service": { "name": "dbus-broker.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-org.bluez.service": { "name": "dbus-org.bluez.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.fedoraproject.FirewallD1.service": { "name": "dbus-org.fedoraproject.FirewallD1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.home1.service": { "name": "dbus-org.freedesktop.home1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.hostname1.service": { "name": "dbus-org.freedesktop.hostname1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.locale1.service": { "name": "dbus-org.freedesktop.locale1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.login1.service": { "name": "dbus-org.freedesktop.login1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.nm-dispatcher.service": { "name": "dbus-org.freedesktop.nm-dispatcher.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.oom1.service": { "name": "dbus-org.freedesktop.oom1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.portable1.service": { "name": "dbus-org.freedesktop.portable1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.resolve1.service": { "name": "dbus-org.freedesktop.resolve1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.timedate1.service": { "name": "dbus-org.freedesktop.timedate1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus.service": { "name": "dbus.service", "source": "systemd", "state": "active", "status": "alias" }, "debug-shell.service": { "name": "debug-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "display-manager.service": { "name": "display-manager.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "dm-event.service": { "name": "dm-event.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-makecache.service": { "name": "dnf-makecache.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-system-upgrade-cleanup.service": { "name": "dnf-system-upgrade-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "dnf-system-upgrade.service": { "name": "dnf-system-upgrade.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dracut-cmdline.service": { "name": "dracut-cmdline.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-initqueue.service": { "name": "dracut-initqueue.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-mount.service": { "name": "dracut-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-mount.service": { "name": "dracut-pre-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-pivot.service": { "name": "dracut-pre-pivot.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-trigger.service": { "name": "dracut-pre-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-udev.service": { "name": "dracut-pre-udev.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown-onfailure.service": { "name": "dracut-shutdown-onfailure.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown.service": { "name": "dracut-shutdown.service", "source": "systemd", "state": "stopped", "status": "static" }, "ebtables.service": { "name": "ebtables.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "emergency.service": { "name": "emergency.service", "source": "systemd", "state": "stopped", "status": "static" }, "firewalld.service": { "name": "firewalld.service", "source": "systemd", "state": "running", "status": "enabled" }, "fsidd.service": { "name": "fsidd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "fstrim.service": { "name": "fstrim.service", "source": "systemd", "state": "stopped", "status": "static" }, "fwupd-offline-update.service": { "name": "fwupd-offline-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd-refresh.service": { "name": "fwupd-refresh.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd.service": { "name": "fwupd.service", "source": "systemd", "state": "inactive", "status": "static" }, "getty@.service": { "name": "getty@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "getty@tty1.service": { "name": "getty@tty1.service", "source": "systemd", "state": "running", "status": "active" }, "grub-boot-indeterminate.service": { "name": "grub-boot-indeterminate.service", "source": "systemd", "state": "inactive", "status": "static" }, "grub2-systemd-integration.service": { "name": "grub2-systemd-integration.service", "source": "systemd", "state": "inactive", "status": "static" }, "gssproxy.service": { "name": "gssproxy.service", "source": "systemd", "state": "running", "status": "disabled" }, "hv_kvp_daemon.service": { "name": "hv_kvp_daemon.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "import-state.service": { "name": "import-state.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "initrd-cleanup.service": { "name": "initrd-cleanup.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-parse-etc.service": { "name": "initrd-parse-etc.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-switch-root.service": { "name": "initrd-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-udevadm-cleanup-db.service": { "name": "initrd-udevadm-cleanup-db.service", "source": "systemd", "state": "stopped", "status": "static" }, "ip6tables.service": { "name": "ip6tables.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ipset.service": { "name": "ipset.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "iptables.service": { "name": "iptables.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "kmod-static-nodes.service": { "name": "kmod-static-nodes.service", "source": "systemd", "state": "stopped", "status": "static" }, "ldconfig.service": { "name": "ldconfig.service", "source": "systemd", "state": "stopped", "status": "static" }, "loadmodules.service": { "name": "loadmodules.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "lvm2-lvmpolld.service": { "name": "lvm2-lvmpolld.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-monitor.service": { "name": "lvm2-monitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "man-db-cache-update.service": { "name": "man-db-cache-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "man-db-restart-cache-update.service": { "name": "man-db-restart-cache-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "mdadm-grow-continue@.service": { "name": "mdadm-grow-continue@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdadm-last-resort@.service": { "name": "mdadm-last-resort@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdcheck_continue.service": { "name": "mdcheck_continue.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdcheck_start.service": { "name": "mdcheck_start.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmon@.service": { "name": "mdmon@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdmonitor-oneshot.service": { "name": "mdmonitor-oneshot.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmonitor.service": { "name": "mdmonitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "modprobe@.service": { "name": "modprobe@.service", "source": "systemd", "state": "unknown", "status": "static" }, "modprobe@configfs.service": { "name": "modprobe@configfs.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "modprobe@dm_mod.service": { "name": "modprobe@dm_mod.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "modprobe@drm.service": { "name": "modprobe@drm.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "modprobe@efi_pstore.service": { "name": "modprobe@efi_pstore.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "modprobe@fuse.service": { "name": "modprobe@fuse.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "modprobe@loop.service": { "name": "modprobe@loop.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "netavark-dhcp-proxy.service": { "name": "netavark-dhcp-proxy.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "netavark-firewalld-reload.service": { "name": "netavark-firewalld-reload.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "network.service": { "name": "network.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "nfs-blkmap.service": { "name": "nfs-blkmap.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-idmapd.service": { "name": "nfs-idmapd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-mountd.service": { "name": "nfs-mountd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-server.service": { "name": "nfs-server.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-utils.service": { "name": "nfs-utils.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfsdcld.service": { "name": "nfsdcld.service", "source": "systemd", "state": "stopped", "status": "static" }, "nftables.service": { "name": "nftables.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nis-domainname.service": { "name": "nis-domainname.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nm-priv-helper.service": { "name": "nm-priv-helper.service", "source": "systemd", "state": "inactive", "status": "static" }, "ntpd.service": { "name": "ntpd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ntpdate.service": { "name": "ntpdate.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "pam_namespace.service": { "name": "pam_namespace.service", "source": "systemd", "state": "inactive", "status": "static" }, "pcscd.service": { "name": "pcscd.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "plymouth-halt.service": { "name": "plymouth-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-kexec.service": { "name": "plymouth-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-poweroff.service": { "name": "plymouth-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-quit-wait.service": { "name": "plymouth-quit-wait.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-quit.service": { "name": "plymouth-quit.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-read-write.service": { "name": "plymouth-read-write.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-reboot.service": { "name": "plymouth-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-start.service": { "name": "plymouth-start.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-switch-root-initramfs.service": { "name": "plymouth-switch-root-initramfs.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-switch-root.service": { "name": "plymouth-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "podman-auto-update.service": { "name": "podman-auto-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "podman-clean-transient.service": { "name": "podman-clean-transient.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "podman-kube@.service": { "name": "podman-kube@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "podman-restart.service": { "name": "podman-restart.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "podman.service": { "name": "podman.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "polkit.service": { "name": "polkit.service", "source": "systemd", "state": "running", "status": "static" }, "quadlet-demo-mysql.service": { "name": "quadlet-demo-mysql.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "quadlet-demo-network.service": { "name": "quadlet-demo-network.service", "source": "systemd", "state": "stopped", "status": "generated" }, "quotaon.service": { "name": "quotaon.service", "source": "systemd", "state": "inactive", "status": "static" }, "raid-check.service": { "name": "raid-check.service", "source": "systemd", "state": "stopped", "status": "static" }, "rc-local.service": { "name": "rc-local.service", "source": "systemd", "state": "stopped", "status": "static" }, "rescue.service": { "name": "rescue.service", "source": "systemd", "state": "stopped", "status": "static" }, "restraintd.service": { "name": "restraintd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rngd.service": { "name": "rngd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpc-gssd.service": { "name": "rpc-gssd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd-notify.service": { "name": "rpc-statd-notify.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd.service": { "name": "rpc-statd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-svcgssd.service": { "name": "rpc-svcgssd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "rpcbind.service": { "name": "rpcbind.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "rpmdb-migrate.service": { "name": "rpmdb-migrate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rpmdb-rebuild.service": { "name": "rpmdb-rebuild.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "selinux-autorelabel-mark.service": { "name": "selinux-autorelabel-mark.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "selinux-autorelabel.service": { "name": "selinux-autorelabel.service", "source": "systemd", "state": "inactive", "status": "static" }, "selinux-check-proper-disable.service": { "name": "selinux-check-proper-disable.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "serial-getty@.service": { "name": "serial-getty@.service", "source": "systemd", "state": "unknown", "status": "indirect" }, "serial-getty@ttyS0.service": { "name": "serial-getty@ttyS0.service", "source": "systemd", "state": "running", "status": "active" }, "snapd.seeded.service": { "name": "snapd.seeded.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sntp.service": { "name": "sntp.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ssh-host-keys-migration.service": { "name": "ssh-host-keys-migration.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "sshd-keygen.service": { "name": "sshd-keygen.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sshd-keygen@.service": { "name": "sshd-keygen@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@ecdsa.service": { "name": "sshd-keygen@ecdsa.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd-keygen@ed25519.service": { "name": "sshd-keygen@ed25519.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd-keygen@rsa.service": { "name": "sshd-keygen@rsa.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd.service": { "name": "sshd.service", "source": "systemd", "state": "running", "status": "enabled" }, "sshd@.service": { "name": "sshd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "sssd-autofs.service": { "name": "sssd-autofs.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-kcm.service": { "name": "sssd-kcm.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "sssd-nss.service": { "name": "sssd-nss.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pac.service": { "name": "sssd-pac.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pam.service": { "name": "sssd-pam.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-ssh.service": { "name": "sssd-ssh.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-sudo.service": { "name": "sssd-sudo.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd.service": { "name": "sssd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "syslog.service": { "name": "syslog.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "system-update-cleanup.service": { "name": "system-update-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-ask-password-console.service": { "name": "systemd-ask-password-console.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-plymouth.service": { "name": "systemd-ask-password-plymouth.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-wall.service": { "name": "systemd-ask-password-wall.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-backlight@.service": { "name": "systemd-backlight@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-battery-check.service": { "name": "systemd-battery-check.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-binfmt.service": { "name": "systemd-binfmt.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-bless-boot.service": { "name": "systemd-bless-boot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-boot-check-no-failures.service": { "name": "systemd-boot-check-no-failures.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-boot-random-seed.service": { "name": "systemd-boot-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-boot-update.service": { "name": "systemd-boot-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-confext.service": { "name": "systemd-confext.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-coredump@.service": { "name": "systemd-coredump@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-exit.service": { "name": "systemd-exit.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-firstboot.service": { "name": "systemd-firstboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck-root.service": { "name": "systemd-fsck-root.service", "source": "systemd", "state": "stopped", "status": "enabled-runtime" }, "systemd-fsck@.service": { "name": "systemd-fsck@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-growfs-root.service": { "name": "systemd-growfs-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-growfs@.service": { "name": "systemd-growfs@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-halt.service": { "name": "systemd-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hibernate.service": { "name": "systemd-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-homed-activate.service": { "name": "systemd-homed-activate.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-homed.service": { "name": "systemd-homed.service", "source": "systemd", "state": "running", "status": "enabled" }, "systemd-hostnamed.service": { "name": "systemd-hostnamed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hwdb-update.service": { "name": "systemd-hwdb-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-hybrid-sleep.service": { "name": "systemd-hybrid-sleep.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-initctl.service": { "name": "systemd-initctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-catalog-update.service": { "name": "systemd-journal-catalog-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-flush.service": { "name": "systemd-journal-flush.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journald.service": { "name": "systemd-journald.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-journald@.service": { "name": "systemd-journald@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-kexec.service": { "name": "systemd-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-localed.service": { "name": "systemd-localed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-logind.service": { "name": "systemd-logind.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-machine-id-commit.service": { "name": "systemd-machine-id-commit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-modules-load.service": { "name": "systemd-modules-load.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-network-generator.service": { "name": "systemd-network-generator.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-networkd-wait-online.service": { "name": "systemd-networkd-wait-online.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "systemd-networkd-wait-online@.service": { "name": "systemd-networkd-wait-online@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "systemd-networkd.service": { "name": "systemd-networkd.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "systemd-oomd.service": { "name": "systemd-oomd.service", "source": "systemd", "state": "running", "status": "enabled" }, "systemd-pcrfs-root.service": { "name": "systemd-pcrfs-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pcrfs@.service": { "name": "systemd-pcrfs@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-pcrmachine.service": { "name": "systemd-pcrmachine.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-pcrphase-initrd.service": { "name": "systemd-pcrphase-initrd.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-pcrphase-sysinit.service": { "name": "systemd-pcrphase-sysinit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-pcrphase.service": { "name": "systemd-pcrphase.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-portabled.service": { "name": "systemd-portabled.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-poweroff.service": { "name": "systemd-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pstore.service": { "name": "systemd-pstore.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-quotacheck.service": { "name": "systemd-quotacheck.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-random-seed.service": { "name": "systemd-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-reboot.service": { "name": "systemd-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-remount-fs.service": { "name": "systemd-remount-fs.service", "source": "systemd", "state": "stopped", "status": "enabled-runtime" }, "systemd-repart.service": { "name": "systemd-repart.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-resolved.service": { "name": "systemd-resolved.service", "source": "systemd", "state": "running", "status": "enabled" }, "systemd-rfkill.service": { "name": "systemd-rfkill.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-soft-reboot.service": { "name": "systemd-soft-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend-then-hibernate.service": { "name": "systemd-suspend-then-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend.service": { "name": "systemd-suspend.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-sysctl.service": { "name": "systemd-sysctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-sysext.service": { "name": "systemd-sysext.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "systemd-sysupdate-reboot.service": { "name": "systemd-sysupdate-reboot.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "systemd-sysupdate.service": { "name": "systemd-sysupdate.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "systemd-sysusers.service": { "name": "systemd-sysusers.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-time-wait-sync.service": { "name": "systemd-time-wait-sync.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-timedated.service": { "name": "systemd-timedated.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-timesyncd.service": { "name": "systemd-timesyncd.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "systemd-tmpfiles-clean.service": { "name": "systemd-tmpfiles-clean.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev-early.service": { "name": "systemd-tmpfiles-setup-dev-early.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev.service": { "name": "systemd-tmpfiles-setup-dev.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup.service": { "name": "systemd-tmpfiles-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-settle.service": { "name": "systemd-udev-settle.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-trigger.service": { "name": "systemd-udev-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udevd.service": { "name": "systemd-udevd.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-update-done.service": { "name": "systemd-update-done.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp-runlevel.service": { "name": "systemd-update-utmp-runlevel.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp.service": { "name": "systemd-update-utmp.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-user-sessions.service": { "name": "systemd-user-sessions.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-userdbd.service": { "name": "systemd-userdbd.service", "source": "systemd", "state": "running", "status": "indirect" }, "systemd-vconsole-setup.service": { "name": "systemd-vconsole-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-volatile-root.service": { "name": "systemd-volatile-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-zram-setup@.service": { "name": "systemd-zram-setup@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-zram-setup@zram0.service": { "name": "systemd-zram-setup@zram0.service", "source": "systemd", "state": "stopped", "status": "active" }, "udisks2.service": { "name": "udisks2.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "unbound-anchor.service": { "name": "unbound-anchor.service", "source": "systemd", "state": "stopped", "status": "static" }, "user-runtime-dir@.service": { "name": "user-runtime-dir@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user-runtime-dir@0.service": { "name": "user-runtime-dir@0.service", "source": "systemd", "state": "stopped", "status": "active" }, "user@.service": { "name": "user@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user@0.service": { "name": "user@0.service", "source": "systemd", "state": "running", "status": "active" } } }, "changed": false } TASK [fedora.linux_system_roles.podman : Create and update quadlets] *********** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:114 Wednesday 10 July 2024 05:15:32 +0000 (0:00:02.298) 0:03:36.845 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_state != \"absent\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14 Wednesday 10 July 2024 05:15:32 +0000 (0:00:00.022) 0:03:36.868 ******** ok: [sut] => { "ansible_facts": { "__podman_quadlet_file_src": "quadlet-demo.network", "__podman_quadlet_spec": {}, "__podman_quadlet_str": "[Network]\nSubnet=192.168.30.0/24\nGateway=192.168.30.1\nLabel=app=wordpress", "__podman_quadlet_template_src": "" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25 Wednesday 10 July 2024 05:15:32 +0000 (0:00:00.031) 0:03:36.899 ******** ok: [sut] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_pull_image": true, "__podman_state": "absent", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] ***** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35 Wednesday 10 July 2024 05:15:32 +0000 (0:00:00.027) 0:03:36.926 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_quadlet_file_src", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48 Wednesday 10 July 2024 05:15:32 +0000 (0:00:00.021) 0:03:36.948 ******** ok: [sut] => { "ansible_facts": { "__podman_quadlet_name": "quadlet-demo", "__podman_quadlet_type": "network", "__podman_rootless": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57 Wednesday 10 July 2024 05:15:32 +0000 (0:00:00.032) 0:03:36.981 ******** included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for sut TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Wednesday 10 July 2024 05:15:32 +0000 (0:00:00.041) 0:03:37.022 ******** skipping: [sut] => { "changed": false, "false_condition": "'getent_passwd' not in ansible_facts or __podman_user not in ansible_facts['getent_passwd']", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Wednesday 10 July 2024 05:15:32 +0000 (0:00:00.025) 0:03:37.048 ******** skipping: [sut] => { "changed": false, "false_condition": "not ansible_facts[\"getent_passwd\"][__podman_user]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Wednesday 10 July 2024 05:15:32 +0000 (0:00:00.026) 0:03:37.074 ******** ok: [sut] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get group information] **************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:28 Wednesday 10 July 2024 05:15:32 +0000 (0:00:00.031) 0:03:37.105 ******** ok: [sut] => { "ansible_facts": { "getent_group": { "root": [ "x", "0", "" ] } }, "changed": false } TASK [fedora.linux_system_roles.podman : Set group name] *********************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:35 Wednesday 10 July 2024 05:15:32 +0000 (0:00:00.237) 0:03:37.343 ******** ok: [sut] => { "ansible_facts": { "__podman_group_name": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:39 Wednesday 10 July 2024 05:15:32 +0000 (0:00:00.029) 0:03:37.372 ******** ok: [sut] => { "changed": false, "stat": { "atime": 1720588427.5164855, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "72ba28a3be19c467e89ab90f5d21f99b605159a7", "ctime": 1720588377.3007717, "dev": 51714, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 203603, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-pie-executable", "mode": "0755", "mtime": 1696291200.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 16032, "uid": 0, "version": "2092964490", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************ task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:50 Wednesday 10 July 2024 05:15:32 +0000 (0:00:00.238) 0:03:37.611 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check group with getsubids] *********** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:55 Wednesday 10 July 2024 05:15:33 +0000 (0:00:00.061) 0:03:37.673 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:60 Wednesday 10 July 2024 05:15:33 +0000 (0:00:00.022) 0:03:37.695 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subuid file] ********************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:74 Wednesday 10 July 2024 05:15:33 +0000 (0:00:00.020) 0:03:37.716 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subgid file] ********************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:79 Wednesday 10 July 2024 05:15:33 +0000 (0:00:00.020) 0:03:37.737 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:84 Wednesday 10 July 2024 05:15:33 +0000 (0:00:00.021) 0:03:37.758 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:94 Wednesday 10 July 2024 05:15:33 +0000 (0:00:00.020) 0:03:37.779 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] ***** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:101 Wednesday 10 July 2024 05:15:33 +0000 (0:00:00.019) 0:03:37.799 ******** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62 Wednesday 10 July 2024 05:15:33 +0000 (0:00:00.020) 0:03:37.819 ******** ok: [sut] => { "ansible_facts": { "__podman_activate_systemd_unit": true, "__podman_images_found": [], "__podman_kube_yamls_raw": "", "__podman_service_name": "quadlet-demo-network.service", "__podman_systemd_scope": "system", "__podman_user_home_dir": "/root", "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73 Wednesday 10 July 2024 05:15:33 +0000 (0:00:00.039) 0:03:37.858 ******** ok: [sut] => { "ansible_facts": { "__podman_quadlet_path": "/etc/containers/systemd" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get kube yaml contents] *************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77 Wednesday 10 July 2024 05:15:33 +0000 (0:00:00.023) 0:03:37.882 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_state != \"absent\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87 Wednesday 10 July 2024 05:15:33 +0000 (0:00:00.019) 0:03:37.901 ******** ok: [sut] => { "ansible_facts": { "__podman_images": [], "__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo.network", "__podman_volumes": [] }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 6] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:103 Wednesday 10 July 2024 05:15:33 +0000 (0:00:00.048) 0:03:37.950 ******** ok: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Cleanup quadlets] ********************* task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:110 Wednesday 10 July 2024 05:15:33 +0000 (0:00:00.025) 0:03:37.976 ******** included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml for sut TASK [fedora.linux_system_roles.podman : Stat XDG_RUNTIME_DIR] ***************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:4 Wednesday 10 July 2024 05:15:33 +0000 (0:00:00.047) 0:03:38.023 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Stop and disable service] ************* task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:12 Wednesday 10 July 2024 05:15:33 +0000 (0:00:00.061) 0:03:38.085 ******** changed: [sut] => { "changed": true, "enabled": false, "failed_when_result": false, "name": "quadlet-demo-network.service", "state": "stopped", "status": { "AccessSELinuxContext": "system_u:object_r:systemd_unit_file_t:s0", "ActiveEnterTimestamp": "Wed 2024-07-10 05:13:58 UTC", "ActiveEnterTimestampMonotonic": "959271314", "ActiveExitTimestampMonotonic": "0", "ActiveState": "active", "After": "basic.target -.mount systemd-journald.socket system.slice sysinit.target", "AllowIsolate": "no", "AssertResult": "yes", "AssertTimestamp": "Wed 2024-07-10 05:13:58 UTC", "AssertTimestampMonotonic": "959223551", "Before": "shutdown.target", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "CPUAccounting": "yes", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "37260000", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "no", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore", "CleanResult": "success", "CollectMode": "inactive", "ConditionResult": "yes", "ConditionTimestamp": "Wed 2024-07-10 05:13:58 UTC", "ConditionTimestampMonotonic": "959223548", "ConfigurationDirectoryMode": "0755", "Conflicts": "shutdown.target", "ControlGroupId": "0", "ControlPID": "0", "CoredumpFilter": "0x33", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "DefaultStartupMemoryLow": "0", "Delegate": "no", "Description": "quadlet-demo-network.service", "DevicePolicy": "auto", "DropInPaths": "/usr/lib/systemd/system/service.d/10-timeout-abort.conf", "DynamicUser": "no", "ExecMainCode": "1", "ExecMainExitTimestamp": "Wed 2024-07-10 05:13:58 UTC", "ExecMainExitTimestampMonotonic": "959271098", "ExecMainPID": "7247", "ExecMainStartTimestamp": "Wed 2024-07-10 05:13:58 UTC", "ExecMainStartTimestampMonotonic": "959231102", "ExecMainStatus": "0", "ExecStart": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman network create --ignore --subnet=192.168.30.0/24 --gateway=192.168.30.1 --label app=wordpress systemd-quadlet-demo ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStartEx": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman network create --ignore --subnet=192.168.30.0/24 --gateway=192.168.30.1 --label app=wordpress systemd-quadlet-demo ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExitType": "main", "ExtensionImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FileDescriptorStorePreserve": "restart", "FinalKillSignal": "9", "FragmentPath": "/run/systemd/generator/quadlet-demo-network.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOReadBytes": "18446744073709551615", "IOReadOperations": "18446744073709551615", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "IOWeight": "[not set]", "IOWriteBytes": "18446744073709551615", "IOWriteOperations": "18446744073709551615", "IPAccounting": "no", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "Id": "quadlet-demo-network.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestamp": "Wed 2024-07-10 05:13:58 UTC", "InactiveExitTimestampMonotonic": "959231435", "InvocationID": "a85c69adee0c4357a9d61aadc1100ae1", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "control-group", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "infinity", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "8388608", "LimitMEMLOCKSoft": "8388608", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "524288", "LimitNOFILESoft": "1024", "LimitNPROC": "14756", "LimitNPROCSoft": "14756", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "14756", "LimitSIGPENDINGSoft": "14756", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "0", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "ManagedOOMSwap": "auto", "MemoryAccounting": "yes", "MemoryAvailable": "infinity", "MemoryCurrent": "[not set]", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryKSM": "no", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemoryPressureThresholdUSec": "200ms", "MemoryPressureWatch": "auto", "MemorySwapMax": "infinity", "MemoryZSwapMax": "infinity", "MountAPIVFS": "no", "MountImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAPolicy": "n/a", "Names": "quadlet-demo-network.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMPolicy": "stop", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "OnSuccessJobMode": "fail", "Perpetual": "no", "PrivateDevices": "no", "PrivateIPC": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProcSubset": "all", "ProtectClock": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectHostname": "no", "ProtectKernelLogs": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectProc": "default", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "ReloadResult": "success", "ReloadSignal": "1", "RemainAfterExit": "yes", "RemoveIPC": "no", "Requires": "-.mount system.slice sysinit.target", "RequiresMountsFor": "/run/containers", "Restart": "no", "RestartKillSignal": "15", "RestartMaxDelayUSec": "infinity", "RestartMode": "normal", "RestartSteps": "0", "RestartUSec": "100ms", "RestartUSecNext": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RootEphemeral": "no", "RootImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "Slice": "system.slice", "StandardError": "inherit", "StandardInput": "null", "StandardOutput": "journal", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StartupMemoryHigh": "18446744073709551615", "StartupMemoryLow": "0", "StartupMemoryMax": "18446744073709551615", "StartupMemorySwapMax": "18446744073709551615", "StartupMemoryZSwapMax": "18446744073709551615", "StateChangeTimestamp": "Wed 2024-07-10 05:13:58 UTC", "StateChangeTimestampMonotonic": "959271314", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "exited", "SuccessAction": "none", "SyslogFacility": "3", "SyslogIdentifier": "quadlet-demo-network", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "2147483646", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "[not set]", "TasksMax": "4426", "TimeoutAbortUSec": "45s", "TimeoutCleanUSec": "infinity", "TimeoutStartFailureMode": "terminate", "TimeoutStartUSec": "infinity", "TimeoutStopFailureMode": "abort", "TimeoutStopUSec": "45s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "oneshot", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "disabled", "UnitFileState": "generated", "UtmpMode": "init", "WatchdogSignal": "6", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "0" } } TASK [fedora.linux_system_roles.podman : Remove quadlet file] ****************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:33 Wednesday 10 July 2024 05:15:34 +0000 (0:00:00.832) 0:03:38.918 ******** changed: [sut] => { "changed": true, "path": "/etc/containers/systemd/quadlet-demo.network", "state": "absent" } TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:39 Wednesday 10 July 2024 05:15:34 +0000 (0:00:00.241) 0:03:39.159 ******** included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for sut TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Wednesday 10 July 2024 05:15:34 +0000 (0:00:00.035) 0:03:39.195 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Wednesday 10 July 2024 05:15:34 +0000 (0:00:00.021) 0:03:39.216 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Wednesday 10 July 2024 05:15:34 +0000 (0:00:00.020) 0:03:39.237 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Reload systemctl] ********************* task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:47 Wednesday 10 July 2024 05:15:34 +0000 (0:00:00.020) 0:03:39.258 ******** ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.podman : Prune images no longer in use] ******** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:56 Wednesday 10 July 2024 05:15:35 +0000 (0:00:00.788) 0:03:40.046 ******** changed: [sut] => { "changed": true, "cmd": [ "podman", "image", "prune", "--all", "-f" ], "delta": "0:00:00.038657", "end": "2024-07-10 05:15:35.619304", "rc": 0, "start": "2024-07-10 05:15:35.580647" } TASK [fedora.linux_system_roles.podman : Remove networks no longer in use] ***** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:64 Wednesday 10 July 2024 05:15:35 +0000 (0:00:00.275) 0:03:40.322 ******** changed: [sut] => { "changed": true, "cmd": [ "podman", "network", "prune", "-f" ], "delta": "0:00:00.035443", "end": "2024-07-10 05:15:35.890721", "rc": 0, "start": "2024-07-10 05:15:35.855278" } STDOUT: systemd-quadlet-demo TASK [fedora.linux_system_roles.podman : Remove volumes no longer in use] ****** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:73 Wednesday 10 July 2024 05:15:35 +0000 (0:00:00.271) 0:03:40.594 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_quadlet_type == \"volume\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : For testing and debugging - images] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:82 Wednesday 10 July 2024 05:15:35 +0000 (0:00:00.023) 0:03:40.617 ******** ok: [sut] => { "changed": false, "cmd": [ "podman", "images", "-n" ], "delta": "0:00:00.033808", "end": "2024-07-10 05:15:36.188924", "rc": 0, "start": "2024-07-10 05:15:36.155116" } TASK [fedora.linux_system_roles.podman : For testing and debugging - volumes] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:92 Wednesday 10 July 2024 05:15:36 +0000 (0:00:00.272) 0:03:40.889 ******** ok: [sut] => { "changed": false, "cmd": [ "podman", "volume", "ls", "-n" ], "delta": "0:00:00.031044", "end": "2024-07-10 05:15:36.456793", "rc": 0, "start": "2024-07-10 05:15:36.425749" } TASK [fedora.linux_system_roles.podman : For testing and debugging - containers] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:102 Wednesday 10 July 2024 05:15:36 +0000 (0:00:00.269) 0:03:41.159 ******** ok: [sut] => { "changed": false, "cmd": [ "podman", "ps", "--noheading" ], "delta": "0:00:00.036209", "end": "2024-07-10 05:15:36.731388", "rc": 0, "start": "2024-07-10 05:15:36.695179" } TASK [fedora.linux_system_roles.podman : For testing and debugging - networks] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:112 Wednesday 10 July 2024 05:15:36 +0000 (0:00:00.317) 0:03:41.476 ******** ok: [sut] => { "changed": false, "cmd": [ "podman", "network", "ls", "-n", "-q" ], "delta": "0:00:00.034849", "end": "2024-07-10 05:15:37.049063", "rc": 0, "start": "2024-07-10 05:15:37.014214" } STDOUT: podman TASK [fedora.linux_system_roles.podman : For testing and debugging - secrets] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:122 Wednesday 10 July 2024 05:15:37 +0000 (0:00:00.276) 0:03:41.752 ******** ok: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : For testing and debugging - services] *** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:133 Wednesday 10 July 2024 05:15:37 +0000 (0:00:00.268) 0:03:42.020 ******** ok: [sut] => { "ansible_facts": { "services": { "NetworkManager-dispatcher.service": { "name": "NetworkManager-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "NetworkManager-wait-online.service": { "name": "NetworkManager-wait-online.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "NetworkManager.service": { "name": "NetworkManager.service", "source": "systemd", "state": "running", "status": "enabled" }, "arp-ethers.service": { "name": "arp-ethers.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "auditd.service": { "name": "auditd.service", "source": "systemd", "state": "running", "status": "enabled" }, "auth-rpcgss-module.service": { "name": "auth-rpcgss-module.service", "source": "systemd", "state": "stopped", "status": "static" }, "autovt@.service": { "name": "autovt@.service", "source": "systemd", "state": "unknown", "status": "alias" }, "blk-availability.service": { "name": "blk-availability.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "bluetooth.service": { "name": "bluetooth.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "certmonger.service": { "name": "certmonger.service", "source": "systemd", "state": "running", "status": "enabled" }, "chrony-wait.service": { "name": "chrony-wait.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd-restricted.service": { "name": "chronyd-restricted.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd.service": { "name": "chronyd.service", "source": "systemd", "state": "running", "status": "enabled" }, "cloud-config.service": { "name": "cloud-config.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-final.service": { "name": "cloud-final.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init-hotplugd.service": { "name": "cloud-init-hotplugd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cloud-init-local.service": { "name": "cloud-init-local.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init.service": { "name": "cloud-init.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "console-getty.service": { "name": "console-getty.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "container-getty@.service": { "name": "container-getty@.service", "source": "systemd", "state": "unknown", "status": "static" }, "dbus-broker.service": { "name": "dbus-broker.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-org.bluez.service": { "name": "dbus-org.bluez.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.fedoraproject.FirewallD1.service": { "name": "dbus-org.fedoraproject.FirewallD1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.home1.service": { "name": "dbus-org.freedesktop.home1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.hostname1.service": { "name": "dbus-org.freedesktop.hostname1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.locale1.service": { "name": "dbus-org.freedesktop.locale1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.login1.service": { "name": "dbus-org.freedesktop.login1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.nm-dispatcher.service": { "name": "dbus-org.freedesktop.nm-dispatcher.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.oom1.service": { "name": "dbus-org.freedesktop.oom1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.portable1.service": { "name": "dbus-org.freedesktop.portable1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.resolve1.service": { "name": "dbus-org.freedesktop.resolve1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.timedate1.service": { "name": "dbus-org.freedesktop.timedate1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus.service": { "name": "dbus.service", "source": "systemd", "state": "active", "status": "alias" }, "debug-shell.service": { "name": "debug-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "display-manager.service": { "name": "display-manager.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "dm-event.service": { "name": "dm-event.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-makecache.service": { "name": "dnf-makecache.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-system-upgrade-cleanup.service": { "name": "dnf-system-upgrade-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "dnf-system-upgrade.service": { "name": "dnf-system-upgrade.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dracut-cmdline.service": { "name": "dracut-cmdline.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-initqueue.service": { "name": "dracut-initqueue.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-mount.service": { "name": "dracut-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-mount.service": { "name": "dracut-pre-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-pivot.service": { "name": "dracut-pre-pivot.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-trigger.service": { "name": "dracut-pre-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-udev.service": { "name": "dracut-pre-udev.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown-onfailure.service": { "name": "dracut-shutdown-onfailure.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown.service": { "name": "dracut-shutdown.service", "source": "systemd", "state": "stopped", "status": "static" }, "ebtables.service": { "name": "ebtables.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "emergency.service": { "name": "emergency.service", "source": "systemd", "state": "stopped", "status": "static" }, "firewalld.service": { "name": "firewalld.service", "source": "systemd", "state": "running", "status": "enabled" }, "fsidd.service": { "name": "fsidd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "fstrim.service": { "name": "fstrim.service", "source": "systemd", "state": "stopped", "status": "static" }, "fwupd-offline-update.service": { "name": "fwupd-offline-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd-refresh.service": { "name": "fwupd-refresh.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd.service": { "name": "fwupd.service", "source": "systemd", "state": "inactive", "status": "static" }, "getty@.service": { "name": "getty@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "getty@tty1.service": { "name": "getty@tty1.service", "source": "systemd", "state": "running", "status": "active" }, "grub-boot-indeterminate.service": { "name": "grub-boot-indeterminate.service", "source": "systemd", "state": "inactive", "status": "static" }, "grub2-systemd-integration.service": { "name": "grub2-systemd-integration.service", "source": "systemd", "state": "inactive", "status": "static" }, "gssproxy.service": { "name": "gssproxy.service", "source": "systemd", "state": "running", "status": "disabled" }, "hv_kvp_daemon.service": { "name": "hv_kvp_daemon.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "import-state.service": { "name": "import-state.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "initrd-cleanup.service": { "name": "initrd-cleanup.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-parse-etc.service": { "name": "initrd-parse-etc.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-switch-root.service": { "name": "initrd-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-udevadm-cleanup-db.service": { "name": "initrd-udevadm-cleanup-db.service", "source": "systemd", "state": "stopped", "status": "static" }, "ip6tables.service": { "name": "ip6tables.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ipset.service": { "name": "ipset.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "iptables.service": { "name": "iptables.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "kmod-static-nodes.service": { "name": "kmod-static-nodes.service", "source": "systemd", "state": "stopped", "status": "static" }, "ldconfig.service": { "name": "ldconfig.service", "source": "systemd", "state": "stopped", "status": "static" }, "loadmodules.service": { "name": "loadmodules.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "lvm2-lvmpolld.service": { "name": "lvm2-lvmpolld.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-monitor.service": { "name": "lvm2-monitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "man-db-cache-update.service": { "name": "man-db-cache-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "man-db-restart-cache-update.service": { "name": "man-db-restart-cache-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "mdadm-grow-continue@.service": { "name": "mdadm-grow-continue@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdadm-last-resort@.service": { "name": "mdadm-last-resort@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdcheck_continue.service": { "name": "mdcheck_continue.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdcheck_start.service": { "name": "mdcheck_start.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmon@.service": { "name": "mdmon@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdmonitor-oneshot.service": { "name": "mdmonitor-oneshot.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmonitor.service": { "name": "mdmonitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "modprobe@.service": { "name": "modprobe@.service", "source": "systemd", "state": "unknown", "status": "static" }, "modprobe@configfs.service": { "name": "modprobe@configfs.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "modprobe@dm_mod.service": { "name": "modprobe@dm_mod.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "modprobe@drm.service": { "name": "modprobe@drm.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "modprobe@efi_pstore.service": { "name": "modprobe@efi_pstore.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "modprobe@fuse.service": { "name": "modprobe@fuse.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "modprobe@loop.service": { "name": "modprobe@loop.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "netavark-dhcp-proxy.service": { "name": "netavark-dhcp-proxy.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "netavark-firewalld-reload.service": { "name": "netavark-firewalld-reload.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "network.service": { "name": "network.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "nfs-blkmap.service": { "name": "nfs-blkmap.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-idmapd.service": { "name": "nfs-idmapd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-mountd.service": { "name": "nfs-mountd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-server.service": { "name": "nfs-server.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-utils.service": { "name": "nfs-utils.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfsdcld.service": { "name": "nfsdcld.service", "source": "systemd", "state": "stopped", "status": "static" }, "nftables.service": { "name": "nftables.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nis-domainname.service": { "name": "nis-domainname.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nm-priv-helper.service": { "name": "nm-priv-helper.service", "source": "systemd", "state": "inactive", "status": "static" }, "ntpd.service": { "name": "ntpd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ntpdate.service": { "name": "ntpdate.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "pam_namespace.service": { "name": "pam_namespace.service", "source": "systemd", "state": "inactive", "status": "static" }, "pcscd.service": { "name": "pcscd.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "plymouth-halt.service": { "name": "plymouth-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-kexec.service": { "name": "plymouth-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-poweroff.service": { "name": "plymouth-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-quit-wait.service": { "name": "plymouth-quit-wait.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-quit.service": { "name": "plymouth-quit.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-read-write.service": { "name": "plymouth-read-write.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-reboot.service": { "name": "plymouth-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-start.service": { "name": "plymouth-start.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-switch-root-initramfs.service": { "name": "plymouth-switch-root-initramfs.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-switch-root.service": { "name": "plymouth-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "podman-auto-update.service": { "name": "podman-auto-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "podman-clean-transient.service": { "name": "podman-clean-transient.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "podman-kube@.service": { "name": "podman-kube@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "podman-restart.service": { "name": "podman-restart.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "podman.service": { "name": "podman.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "polkit.service": { "name": "polkit.service", "source": "systemd", "state": "running", "status": "static" }, "quadlet-demo-mysql.service": { "name": "quadlet-demo-mysql.service", "source": "systemd", "state": "stopped", "status": "failed" }, "quotaon.service": { "name": "quotaon.service", "source": "systemd", "state": "inactive", "status": "static" }, "raid-check.service": { "name": "raid-check.service", "source": "systemd", "state": "stopped", "status": "static" }, "rc-local.service": { "name": "rc-local.service", "source": "systemd", "state": "stopped", "status": "static" }, "rescue.service": { "name": "rescue.service", "source": "systemd", "state": "stopped", "status": "static" }, "restraintd.service": { "name": "restraintd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rngd.service": { "name": "rngd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpc-gssd.service": { "name": "rpc-gssd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd-notify.service": { "name": "rpc-statd-notify.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd.service": { "name": "rpc-statd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-svcgssd.service": { "name": "rpc-svcgssd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "rpcbind.service": { "name": "rpcbind.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "rpmdb-migrate.service": { "name": "rpmdb-migrate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rpmdb-rebuild.service": { "name": "rpmdb-rebuild.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "selinux-autorelabel-mark.service": { "name": "selinux-autorelabel-mark.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "selinux-autorelabel.service": { "name": "selinux-autorelabel.service", "source": "systemd", "state": "inactive", "status": "static" }, "selinux-check-proper-disable.service": { "name": "selinux-check-proper-disable.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "serial-getty@.service": { "name": "serial-getty@.service", "source": "systemd", "state": "unknown", "status": "indirect" }, "serial-getty@ttyS0.service": { "name": "serial-getty@ttyS0.service", "source": "systemd", "state": "running", "status": "active" }, "snapd.seeded.service": { "name": "snapd.seeded.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sntp.service": { "name": "sntp.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ssh-host-keys-migration.service": { "name": "ssh-host-keys-migration.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "sshd-keygen.service": { "name": "sshd-keygen.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sshd-keygen@.service": { "name": "sshd-keygen@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@ecdsa.service": { "name": "sshd-keygen@ecdsa.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd-keygen@ed25519.service": { "name": "sshd-keygen@ed25519.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd-keygen@rsa.service": { "name": "sshd-keygen@rsa.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd.service": { "name": "sshd.service", "source": "systemd", "state": "running", "status": "enabled" }, "sshd@.service": { "name": "sshd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "sssd-autofs.service": { "name": "sssd-autofs.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-kcm.service": { "name": "sssd-kcm.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "sssd-nss.service": { "name": "sssd-nss.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pac.service": { "name": "sssd-pac.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pam.service": { "name": "sssd-pam.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-ssh.service": { "name": "sssd-ssh.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-sudo.service": { "name": "sssd-sudo.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd.service": { "name": "sssd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "syslog.service": { "name": "syslog.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "system-update-cleanup.service": { "name": "system-update-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-ask-password-console.service": { "name": "systemd-ask-password-console.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-plymouth.service": { "name": "systemd-ask-password-plymouth.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-wall.service": { "name": "systemd-ask-password-wall.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-backlight@.service": { "name": "systemd-backlight@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-battery-check.service": { "name": "systemd-battery-check.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-binfmt.service": { "name": "systemd-binfmt.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-bless-boot.service": { "name": "systemd-bless-boot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-boot-check-no-failures.service": { "name": "systemd-boot-check-no-failures.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-boot-random-seed.service": { "name": "systemd-boot-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-boot-update.service": { "name": "systemd-boot-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-confext.service": { "name": "systemd-confext.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-coredump@.service": { "name": "systemd-coredump@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-exit.service": { "name": "systemd-exit.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-firstboot.service": { "name": "systemd-firstboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck-root.service": { "name": "systemd-fsck-root.service", "source": "systemd", "state": "stopped", "status": "enabled-runtime" }, "systemd-fsck@.service": { "name": "systemd-fsck@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-growfs-root.service": { "name": "systemd-growfs-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-growfs@.service": { "name": "systemd-growfs@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-halt.service": { "name": "systemd-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hibernate.service": { "name": "systemd-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-homed-activate.service": { "name": "systemd-homed-activate.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-homed.service": { "name": "systemd-homed.service", "source": "systemd", "state": "running", "status": "enabled" }, "systemd-hostnamed.service": { "name": "systemd-hostnamed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hwdb-update.service": { "name": "systemd-hwdb-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-hybrid-sleep.service": { "name": "systemd-hybrid-sleep.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-initctl.service": { "name": "systemd-initctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-catalog-update.service": { "name": "systemd-journal-catalog-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-flush.service": { "name": "systemd-journal-flush.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journald.service": { "name": "systemd-journald.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-journald@.service": { "name": "systemd-journald@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-kexec.service": { "name": "systemd-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-localed.service": { "name": "systemd-localed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-logind.service": { "name": "systemd-logind.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-machine-id-commit.service": { "name": "systemd-machine-id-commit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-modules-load.service": { "name": "systemd-modules-load.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-network-generator.service": { "name": "systemd-network-generator.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-networkd-wait-online.service": { "name": "systemd-networkd-wait-online.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "systemd-networkd-wait-online@.service": { "name": "systemd-networkd-wait-online@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "systemd-networkd.service": { "name": "systemd-networkd.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "systemd-oomd.service": { "name": "systemd-oomd.service", "source": "systemd", "state": "running", "status": "enabled" }, "systemd-pcrfs-root.service": { "name": "systemd-pcrfs-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pcrfs@.service": { "name": "systemd-pcrfs@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-pcrmachine.service": { "name": "systemd-pcrmachine.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-pcrphase-initrd.service": { "name": "systemd-pcrphase-initrd.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-pcrphase-sysinit.service": { "name": "systemd-pcrphase-sysinit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-pcrphase.service": { "name": "systemd-pcrphase.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-portabled.service": { "name": "systemd-portabled.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-poweroff.service": { "name": "systemd-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pstore.service": { "name": "systemd-pstore.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-quotacheck.service": { "name": "systemd-quotacheck.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-random-seed.service": { "name": "systemd-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-reboot.service": { "name": "systemd-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-remount-fs.service": { "name": "systemd-remount-fs.service", "source": "systemd", "state": "stopped", "status": "enabled-runtime" }, "systemd-repart.service": { "name": "systemd-repart.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-resolved.service": { "name": "systemd-resolved.service", "source": "systemd", "state": "running", "status": "enabled" }, "systemd-rfkill.service": { "name": "systemd-rfkill.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-soft-reboot.service": { "name": "systemd-soft-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend-then-hibernate.service": { "name": "systemd-suspend-then-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend.service": { "name": "systemd-suspend.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-sysctl.service": { "name": "systemd-sysctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-sysext.service": { "name": "systemd-sysext.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "systemd-sysupdate-reboot.service": { "name": "systemd-sysupdate-reboot.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "systemd-sysupdate.service": { "name": "systemd-sysupdate.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "systemd-sysusers.service": { "name": "systemd-sysusers.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-time-wait-sync.service": { "name": "systemd-time-wait-sync.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-timedated.service": { "name": "systemd-timedated.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-timesyncd.service": { "name": "systemd-timesyncd.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "systemd-tmpfiles-clean.service": { "name": "systemd-tmpfiles-clean.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev-early.service": { "name": "systemd-tmpfiles-setup-dev-early.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev.service": { "name": "systemd-tmpfiles-setup-dev.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup.service": { "name": "systemd-tmpfiles-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-settle.service": { "name": "systemd-udev-settle.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-trigger.service": { "name": "systemd-udev-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udevd.service": { "name": "systemd-udevd.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-update-done.service": { "name": "systemd-update-done.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp-runlevel.service": { "name": "systemd-update-utmp-runlevel.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp.service": { "name": "systemd-update-utmp.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-user-sessions.service": { "name": "systemd-user-sessions.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-userdbd.service": { "name": "systemd-userdbd.service", "source": "systemd", "state": "running", "status": "indirect" }, "systemd-vconsole-setup.service": { "name": "systemd-vconsole-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-volatile-root.service": { "name": "systemd-volatile-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-zram-setup@.service": { "name": "systemd-zram-setup@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-zram-setup@zram0.service": { "name": "systemd-zram-setup@zram0.service", "source": "systemd", "state": "stopped", "status": "active" }, "udisks2.service": { "name": "udisks2.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "unbound-anchor.service": { "name": "unbound-anchor.service", "source": "systemd", "state": "stopped", "status": "static" }, "user-runtime-dir@.service": { "name": "user-runtime-dir@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user-runtime-dir@0.service": { "name": "user-runtime-dir@0.service", "source": "systemd", "state": "stopped", "status": "active" }, "user@.service": { "name": "user@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user@0.service": { "name": "user@0.service", "source": "systemd", "state": "running", "status": "active" } } }, "changed": false } TASK [fedora.linux_system_roles.podman : Create and update quadlets] *********** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:114 Wednesday 10 July 2024 05:15:39 +0000 (0:00:02.220) 0:03:44.241 ******** skipping: [sut] => { "changed": false, "false_condition": "__podman_state != \"absent\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Cancel linger] ************************ task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:143 Wednesday 10 July 2024 05:15:39 +0000 (0:00:00.020) 0:03:44.262 ******** skipping: [sut] => { "changed": false, "skipped_reason": "No items in the list" } TASK [fedora.linux_system_roles.podman : Handle credential files - absent] ***** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:149 Wednesday 10 July 2024 05:15:39 +0000 (0:00:00.019) 0:03:44.282 ******** skipping: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Handle certs.d files - absent] ******** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:158 Wednesday 10 July 2024 05:15:39 +0000 (0:00:00.019) 0:03:44.301 ******** skipping: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [Ensure no resources] ***************************************************** task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/tests/tests_quadlet_demo.yml:170 Wednesday 10 July 2024 05:15:39 +0000 (0:00:00.031) 0:03:44.332 ******** fatal: [sut]: FAILED! => { "assertion": "ansible_facts[\"services\"] | dict2items | rejectattr(\"value.status\", \"match\", \"not-found\") | selectattr(\"key\", \"match\", \"quadlet-demo\") | list | length == 0", "changed": false, "evaluated_to": false } MSG: Assertion failed TASK [Debug] ******************************************************************* task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/tests/tests_quadlet_demo.yml:181 Wednesday 10 July 2024 05:15:39 +0000 (0:00:00.036) 0:03:44.369 ******** ok: [sut] => { "changed": false, "cmd": "exec 1>&2\nset -x\nset -o pipefail\nsystemctl list-units --plain -l --all | grep quadlet || :\nsystemctl list-unit-files --all | grep quadlet || :\nsystemctl list-units --plain --failed -l --all | grep quadlet || :\njournalctl -u quadlet-demo-mysql.service || :\nsystemctl status quadlet-demo-mysql.service || :\n", "delta": "0:00:00.596940", "end": "2024-07-10 05:15:40.494700", "rc": 0, "start": "2024-07-10 05:15:39.897760" } STDERR: + set -o pipefail + systemctl list-units --plain -l --all + grep quadlet quadlet-demo-mysql.service not-found failed failed quadlet-demo-mysql.service + systemctl list-unit-files --all + grep quadlet + : + systemctl list-units --plain --failed -l --all + grep quadlet quadlet-demo-mysql.service not-found failed failed quadlet-demo-mysql.service + journalctl -u quadlet-demo-mysql.service Jul 10 05:14:12 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: Starting quadlet-demo-mysql.service... Jul 10 05:14:12 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm podman[7928]: 2024-07-10 05:14:12.80841545 +0000 UTC m=+0.052126110 container create 21b1326f57855b147efcb07cbd10c5f7b7edbabb6693b7752105182196a3f742 (image=quay.io/linux-system-roles/mysql:5.6, name=quadlet-demo-mysql, PODMAN_SYSTEMD_UNIT=quadlet-demo-mysql.service) Jul 10 05:14:12 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm podman[7928]: 2024-07-10 05:14:12.786708435 +0000 UTC m=+0.030419211 image pull dd3b2a5dcb48ff61113592ed5ddd762581be4387c7bc552375a2159422aa6bf5 quay.io/linux-system-roles/mysql:5.6 Jul 10 05:14:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm podman[7928]: 2024-07-10 05:14:13.195131115 +0000 UTC m=+0.438841882 container init 21b1326f57855b147efcb07cbd10c5f7b7edbabb6693b7752105182196a3f742 (image=quay.io/linux-system-roles/mysql:5.6, name=quadlet-demo-mysql, PODMAN_SYSTEMD_UNIT=quadlet-demo-mysql.service) Jul 10 05:14:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: Started quadlet-demo-mysql.service. Jul 10 05:14:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:13+00:00 [Note] [Entrypoint]: Entrypoint script for MySQL Server 5.6.51-1debian9 started. Jul 10 05:14:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm podman[7928]: 2024-07-10 05:14:13.232698056 +0000 UTC m=+0.476408784 container start 21b1326f57855b147efcb07cbd10c5f7b7edbabb6693b7752105182196a3f742 (image=quay.io/linux-system-roles/mysql:5.6, name=quadlet-demo-mysql, PODMAN_SYSTEMD_UNIT=quadlet-demo-mysql.service) Jul 10 05:14:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[7928]: 21b1326f57855b147efcb07cbd10c5f7b7edbabb6693b7752105182196a3f742 Jul 10 05:14:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:13+00:00 [Note] [Entrypoint]: Switching to dedicated user 'mysql' Jul 10 05:14:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:13+00:00 [Note] [Entrypoint]: Entrypoint script for MySQL Server 5.6.51-1debian9 started. Jul 10 05:14:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:13+00:00 [Note] [Entrypoint]: Initializing database files Jul 10 05:14:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:13 0 [Warning] TIMESTAMP with implicit DEFAULT value is deprecated. Please use --explicit_defaults_for_timestamp server option (see documentation for more details). Jul 10 05:14:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:13 0 [Note] Ignoring --secure-file-priv value as server is running with --bootstrap. Jul 10 05:14:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:13 0 [Note] /usr/sbin/mysqld (mysqld 5.6.51) starting as process 42 ... Jul 10 05:14:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:13 42 [Note] InnoDB: Using atomics to ref count buffer pool pages Jul 10 05:14:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:13 42 [Note] InnoDB: The InnoDB memory heap is disabled Jul 10 05:14:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:13 42 [Note] InnoDB: Mutexes and rw_locks use GCC atomic builtins Jul 10 05:14:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:13 42 [Note] InnoDB: Memory barrier is not used Jul 10 05:14:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:13 42 [Note] InnoDB: Compressed tables use zlib 1.2.11 Jul 10 05:14:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:13 42 [Note] InnoDB: Using Linux native AIO Jul 10 05:14:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:13 42 [Note] InnoDB: Using CPU crc32 instructions Jul 10 05:14:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:13 42 [Note] InnoDB: Initializing buffer pool, size = 128.0M Jul 10 05:14:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:13 42 [Note] InnoDB: Completed initialization of buffer pool Jul 10 05:14:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:13 42 [Note] InnoDB: The first specified data file ./ibdata1 did not exist: a new database to be created! Jul 10 05:14:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:13 42 [Note] InnoDB: Setting file ./ibdata1 size to 12 MB Jul 10 05:14:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:13 42 [Note] InnoDB: Database physically writes the file full: wait... Jul 10 05:14:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:13 42 [Note] InnoDB: Setting log file ./ib_logfile101 size to 48 MB Jul 10 05:14:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:13 42 [Note] InnoDB: Setting log file ./ib_logfile1 size to 48 MB Jul 10 05:14:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:14 42 [Note] InnoDB: Renaming log file ./ib_logfile101 to ./ib_logfile0 Jul 10 05:14:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:14 42 [Warning] InnoDB: New log files created, LSN=45781 Jul 10 05:14:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:14 42 [Note] InnoDB: Doublewrite buffer not found: creating new Jul 10 05:14:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:14 42 [Note] InnoDB: Doublewrite buffer created Jul 10 05:14:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:14 42 [Note] InnoDB: 128 rollback segment(s) are active. Jul 10 05:14:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:14 42 [Warning] InnoDB: Creating foreign key constraint system tables. Jul 10 05:14:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:14 42 [Note] InnoDB: Foreign key constraint system tables created Jul 10 05:14:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:14 42 [Note] InnoDB: Creating tablespace and datafile system tables. Jul 10 05:14:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:14 42 [Note] InnoDB: Tablespace and datafile system tables created. Jul 10 05:14:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:14 42 [Note] InnoDB: Waiting for purge to start Jul 10 05:14:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:14 42 [Note] InnoDB: 5.6.51 started; log sequence number 0 Jul 10 05:14:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:14 42 [Note] RSA private key file not found: /var/lib/mysql//private_key.pem. Some authentication plugins will not work. Jul 10 05:14:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:14 42 [Note] RSA public key file not found: /var/lib/mysql//public_key.pem. Some authentication plugins will not work. Jul 10 05:14:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:14 42 [Note] Binlog end Jul 10 05:14:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:14 42 [Note] InnoDB: FTS optimize thread exiting. Jul 10 05:14:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:14 42 [Note] InnoDB: Starting shutdown... Jul 10 05:14:16 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:16 42 [Note] InnoDB: Shutdown completed; log sequence number 1625977 Jul 10 05:14:16 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: Jul 10 05:14:16 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: Jul 10 05:14:16 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:16 0 [Warning] TIMESTAMP with implicit DEFAULT value is deprecated. Please use --explicit_defaults_for_timestamp server option (see documentation for more details). Jul 10 05:14:16 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:16 0 [Note] Ignoring --secure-file-priv value as server is running with --bootstrap. Jul 10 05:14:16 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:16 0 [Note] /usr/sbin/mysqld (mysqld 5.6.51) starting as process 65 ... Jul 10 05:14:16 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:16 65 [Note] InnoDB: Using atomics to ref count buffer pool pages Jul 10 05:14:16 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:16 65 [Note] InnoDB: The InnoDB memory heap is disabled Jul 10 05:14:16 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:16 65 [Note] InnoDB: Mutexes and rw_locks use GCC atomic builtins Jul 10 05:14:16 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:16 65 [Note] InnoDB: Memory barrier is not used Jul 10 05:14:16 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:16 65 [Note] InnoDB: Compressed tables use zlib 1.2.11 Jul 10 05:14:16 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:16 65 [Note] InnoDB: Using Linux native AIO Jul 10 05:14:16 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:16 65 [Note] InnoDB: Using CPU crc32 instructions Jul 10 05:14:16 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:16 65 [Note] InnoDB: Initializing buffer pool, size = 128.0M Jul 10 05:14:16 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:16 65 [Note] InnoDB: Completed initialization of buffer pool Jul 10 05:14:16 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:16 65 [Note] InnoDB: Highest supported file format is Barracuda. Jul 10 05:14:16 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:16 65 [Note] InnoDB: 128 rollback segment(s) are active. Jul 10 05:14:16 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:16 65 [Note] InnoDB: Waiting for purge to start Jul 10 05:14:16 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:16 65 [Note] InnoDB: 5.6.51 started; log sequence number 1625977 Jul 10 05:14:16 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:16 65 [Note] RSA private key file not found: /var/lib/mysql//private_key.pem. Some authentication plugins will not work. Jul 10 05:14:16 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:16 65 [Note] RSA public key file not found: /var/lib/mysql//public_key.pem. Some authentication plugins will not work. Jul 10 05:14:16 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:16 65 [Note] Binlog end Jul 10 05:14:16 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:16 65 [Note] InnoDB: FTS optimize thread exiting. Jul 10 05:14:16 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:16 65 [Note] InnoDB: Starting shutdown... Jul 10 05:14:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:18 65 [Note] InnoDB: Shutdown completed; log sequence number 1625987 Jul 10 05:14:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: Jul 10 05:14:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: Jul 10 05:14:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: Jul 10 05:14:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: Jul 10 05:14:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: PLEASE REMEMBER TO SET A PASSWORD FOR THE MySQL root USER ! Jul 10 05:14:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: To do so, start the server, then issue the following commands: Jul 10 05:14:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: Jul 10 05:14:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: /usr/bin/mysqladmin -u root password 'new-password' Jul 10 05:14:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: /usr/bin/mysqladmin -u root -h 21b1326f5785 password 'new-password' Jul 10 05:14:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: Jul 10 05:14:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: Alternatively you can run: Jul 10 05:14:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: Jul 10 05:14:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: /usr/bin/mysql_secure_installation Jul 10 05:14:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: Jul 10 05:14:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: which will also give you the option of removing the test Jul 10 05:14:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: databases and anonymous user created by default. This is Jul 10 05:14:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: strongly recommended for production servers. Jul 10 05:14:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: Jul 10 05:14:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: See the manual for more instructions. Jul 10 05:14:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: Jul 10 05:14:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: Please report any problems at http://bugs.mysql.com/ Jul 10 05:14:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: Jul 10 05:14:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: The latest information about MySQL is available on the web at Jul 10 05:14:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: Jul 10 05:14:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: http://www.mysql.com Jul 10 05:14:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: Jul 10 05:14:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: Support MySQL by buying support/licenses at http://shop.mysql.com Jul 10 05:14:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: Jul 10 05:14:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: Note: new default config file not created. Jul 10 05:14:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: Please make sure your config file is current Jul 10 05:14:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: Jul 10 05:14:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: WARNING: Default config file /etc/mysql/my.cnf exists on the system Jul 10 05:14:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: This file will be read by default by the MySQL server Jul 10 05:14:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: If you do not want to use this, either remove it, or use the Jul 10 05:14:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: --defaults-file argument to mysqld_safe when starting the server Jul 10 05:14:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: Jul 10 05:14:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:18+00:00 [Note] [Entrypoint]: Database files initialized Jul 10 05:14:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:18+00:00 [Note] [Entrypoint]: Starting temporary server Jul 10 05:14:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:18+00:00 [Note] [Entrypoint]: Waiting for server startup Jul 10 05:14:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:19 0 [Warning] TIMESTAMP with implicit DEFAULT value is deprecated. Please use --explicit_defaults_for_timestamp server option (see documentation for more details). Jul 10 05:14:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:19 0 [Note] mysqld (mysqld 5.6.51) starting as process 90 ... Jul 10 05:14:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:19 90 [Note] Plugin 'FEDERATED' is disabled. Jul 10 05:14:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:19 90 [Note] InnoDB: Using atomics to ref count buffer pool pages Jul 10 05:14:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:19 90 [Note] InnoDB: The InnoDB memory heap is disabled Jul 10 05:14:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:19 90 [Note] InnoDB: Mutexes and rw_locks use GCC atomic builtins Jul 10 05:14:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:19 90 [Note] InnoDB: Memory barrier is not used Jul 10 05:14:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:19 90 [Note] InnoDB: Compressed tables use zlib 1.2.11 Jul 10 05:14:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:19 90 [Note] InnoDB: Using Linux native AIO Jul 10 05:14:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:19 90 [Note] InnoDB: Using CPU crc32 instructions Jul 10 05:14:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:19 90 [Note] InnoDB: Initializing buffer pool, size = 128.0M Jul 10 05:14:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:19 90 [Note] InnoDB: Completed initialization of buffer pool Jul 10 05:14:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:19 90 [Note] InnoDB: Highest supported file format is Barracuda. Jul 10 05:14:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:19 90 [Note] InnoDB: 128 rollback segment(s) are active. Jul 10 05:14:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:19 90 [Note] InnoDB: Waiting for purge to start Jul 10 05:14:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:19 90 [Note] InnoDB: 5.6.51 started; log sequence number 1625987 Jul 10 05:14:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:19 90 [Warning] No existing UUID has been found, so we assume that this is the first time that this server has been started. Generating a new UUID: 42dc40c5-3e7b-11ef-9a49-2a8356b5d061. Jul 10 05:14:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:19 90 [Note] RSA private key file not found: /var/lib/mysql//private_key.pem. Some authentication plugins will not work. Jul 10 05:14:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:19 90 [Note] RSA public key file not found: /var/lib/mysql//public_key.pem. Some authentication plugins will not work. Jul 10 05:14:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:19 90 [Warning] Insecure configuration for --pid-file: Location '/var/run/mysqld' in the path is accessible to all OS users. Consider choosing a different directory. Jul 10 05:14:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:19 90 [Warning] 'user' entry 'root@21b1326f5785' ignored in --skip-name-resolve mode. Jul 10 05:14:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:19 90 [Warning] 'user' entry '@21b1326f5785' ignored in --skip-name-resolve mode. Jul 10 05:14:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:19 90 [Warning] 'proxies_priv' entry '@ root@21b1326f5785' ignored in --skip-name-resolve mode. Jul 10 05:14:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:19 90 [Note] Event Scheduler: Loaded 0 events Jul 10 05:14:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:19 90 [Note] mysqld: ready for connections. Jul 10 05:14:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: Version: '5.6.51' socket: '/var/run/mysqld/mysqld.sock' port: 0 MySQL Community Server (GPL) Jul 10 05:14:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:19+00:00 [Note] [Entrypoint]: Temporary server started. Jul 10 05:14:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: Warning: Unable to load '/usr/share/zoneinfo/iso3166.tab' as time zone. Skipping it. Jul 10 05:14:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: Warning: Unable to load '/usr/share/zoneinfo/leap-seconds.list' as time zone. Skipping it. Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: Warning: Unable to load '/usr/share/zoneinfo/zone.tab' as time zone. Skipping it. Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: Warning: Unable to load '/usr/share/zoneinfo/zone1970.tab' as time zone. Skipping it. Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:21 90 [Warning] 'proxies_priv' entry '@ root@21b1326f5785' ignored in --skip-name-resolve mode. Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:21+00:00 [Note] [Entrypoint]: Stopping temporary server Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:21 90 [Note] mysqld: Normal shutdown Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:21 90 [Note] Giving 0 client threads a chance to die gracefully Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:21 90 [Note] Event Scheduler: Purging the queue. 0 events Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:21 90 [Note] Shutting down slave threads Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:21 90 [Note] Forcefully disconnecting 0 remaining clients Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:21 90 [Note] Binlog end Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:21 90 [Note] Shutting down plugin 'partition' Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:21 90 [Note] Shutting down plugin 'PERFORMANCE_SCHEMA' Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:21 90 [Note] Shutting down plugin 'INNODB_SYS_DATAFILES' Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:21 90 [Note] Shutting down plugin 'INNODB_SYS_TABLESPACES' Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:21 90 [Note] Shutting down plugin 'INNODB_SYS_FOREIGN_COLS' Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:21 90 [Note] Shutting down plugin 'INNODB_SYS_FOREIGN' Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:21 90 [Note] Shutting down plugin 'INNODB_SYS_FIELDS' Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:21 90 [Note] Shutting down plugin 'INNODB_SYS_COLUMNS' Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:21 90 [Note] Shutting down plugin 'INNODB_SYS_INDEXES' Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:21 90 [Note] Shutting down plugin 'INNODB_SYS_TABLESTATS' Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:21 90 [Note] Shutting down plugin 'INNODB_SYS_TABLES' Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:21 90 [Note] Shutting down plugin 'INNODB_FT_INDEX_TABLE' Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:21 90 [Note] Shutting down plugin 'INNODB_FT_INDEX_CACHE' Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:21 90 [Note] Shutting down plugin 'INNODB_FT_CONFIG' Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:21 90 [Note] Shutting down plugin 'INNODB_FT_BEING_DELETED' Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:21 90 [Note] Shutting down plugin 'INNODB_FT_DELETED' Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:21 90 [Note] Shutting down plugin 'INNODB_FT_DEFAULT_STOPWORD' Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:21 90 [Note] Shutting down plugin 'INNODB_METRICS' Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:21 90 [Note] Shutting down plugin 'INNODB_BUFFER_POOL_STATS' Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:21 90 [Note] Shutting down plugin 'INNODB_BUFFER_PAGE_LRU' Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:21 90 [Note] Shutting down plugin 'INNODB_BUFFER_PAGE' Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:21 90 [Note] Shutting down plugin 'INNODB_CMP_PER_INDEX_RESET' Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:21 90 [Note] Shutting down plugin 'INNODB_CMP_PER_INDEX' Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:21 90 [Note] Shutting down plugin 'INNODB_CMPMEM_RESET' Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:21 90 [Note] Shutting down plugin 'INNODB_CMPMEM' Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:21 90 [Note] Shutting down plugin 'INNODB_CMP_RESET' Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:21 90 [Note] Shutting down plugin 'INNODB_CMP' Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:21 90 [Note] Shutting down plugin 'INNODB_LOCK_WAITS' Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:21 90 [Note] Shutting down plugin 'INNODB_LOCKS' Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:21 90 [Note] Shutting down plugin 'INNODB_TRX' Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:21 90 [Note] Shutting down plugin 'InnoDB' Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:21 90 [Note] InnoDB: FTS optimize thread exiting. Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:21 90 [Note] InnoDB: Starting shutdown... Jul 10 05:14:23 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:23 90 [Note] InnoDB: Shutdown completed; log sequence number 1625997 Jul 10 05:14:23 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:23 90 [Note] Shutting down plugin 'BLACKHOLE' Jul 10 05:14:23 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:23 90 [Note] Shutting down plugin 'ARCHIVE' Jul 10 05:14:23 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:23 90 [Note] Shutting down plugin 'MRG_MYISAM' Jul 10 05:14:23 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:23 90 [Note] Shutting down plugin 'MyISAM' Jul 10 05:14:23 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:23 90 [Note] Shutting down plugin 'MEMORY' Jul 10 05:14:23 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:23 90 [Note] Shutting down plugin 'CSV' Jul 10 05:14:23 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:23 90 [Note] Shutting down plugin 'sha256_password' Jul 10 05:14:23 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:23 90 [Note] Shutting down plugin 'mysql_old_password' Jul 10 05:14:23 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:23 90 [Note] Shutting down plugin 'mysql_native_password' Jul 10 05:14:23 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:23 90 [Note] Shutting down plugin 'binlog' Jul 10 05:14:23 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:23 90 [Note] mysqld: Shutdown complete Jul 10 05:14:23 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: Jul 10 05:14:23 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:23+00:00 [Note] [Entrypoint]: Temporary server stopped Jul 10 05:14:23 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: Jul 10 05:14:23 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:23+00:00 [Note] [Entrypoint]: MySQL init process done. Ready for start up. Jul 10 05:14:23 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: Jul 10 05:14:24 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:24 0 [Warning] TIMESTAMP with implicit DEFAULT value is deprecated. Please use --explicit_defaults_for_timestamp server option (see documentation for more details). Jul 10 05:14:24 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:24 0 [Note] mysqld (mysqld 5.6.51) starting as process 1 ... Jul 10 05:14:24 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:24 1 [Note] Plugin 'FEDERATED' is disabled. Jul 10 05:14:24 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:24 1 [Note] InnoDB: Using atomics to ref count buffer pool pages Jul 10 05:14:24 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:24 1 [Note] InnoDB: The InnoDB memory heap is disabled Jul 10 05:14:24 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:24 1 [Note] InnoDB: Mutexes and rw_locks use GCC atomic builtins Jul 10 05:14:24 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:24 1 [Note] InnoDB: Memory barrier is not used Jul 10 05:14:24 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:24 1 [Note] InnoDB: Compressed tables use zlib 1.2.11 Jul 10 05:14:24 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:24 1 [Note] InnoDB: Using Linux native AIO Jul 10 05:14:24 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:24 1 [Note] InnoDB: Using CPU crc32 instructions Jul 10 05:14:24 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:24 1 [Note] InnoDB: Initializing buffer pool, size = 128.0M Jul 10 05:14:24 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:24 1 [Note] InnoDB: Completed initialization of buffer pool Jul 10 05:14:24 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:24 1 [Note] InnoDB: Highest supported file format is Barracuda. Jul 10 05:14:24 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:24 1 [Note] InnoDB: 128 rollback segment(s) are active. Jul 10 05:14:24 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:24 1 [Note] InnoDB: Waiting for purge to start Jul 10 05:14:24 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:24 1 [Note] InnoDB: 5.6.51 started; log sequence number 1625997 Jul 10 05:14:24 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:24 1 [Note] RSA private key file not found: /var/lib/mysql//private_key.pem. Some authentication plugins will not work. Jul 10 05:14:24 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:24 1 [Note] RSA public key file not found: /var/lib/mysql//public_key.pem. Some authentication plugins will not work. Jul 10 05:14:24 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:24 1 [Note] Server hostname (bind-address): '*'; port: 3306 Jul 10 05:14:24 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:24 1 [Note] IPv6 is available. Jul 10 05:14:24 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:24 1 [Note] - '::' resolves to '::'; Jul 10 05:14:24 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:24 1 [Note] Server socket created on IP: '::'. Jul 10 05:14:24 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:24 1 [Warning] Insecure configuration for --pid-file: Location '/var/run/mysqld' in the path is accessible to all OS users. Consider choosing a different directory. Jul 10 05:14:24 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:24 1 [Warning] 'proxies_priv' entry '@ root@21b1326f5785' ignored in --skip-name-resolve mode. Jul 10 05:14:24 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:24 1 [Note] Event Scheduler: Loaded 0 events Jul 10 05:14:24 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:24 1 [Note] mysqld: ready for connections. Jul 10 05:14:24 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: Version: '5.6.51' socket: '/var/run/mysqld/mysqld.sock' port: 3306 MySQL Community Server (GPL) Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm podman[11232]: 2024-07-10 05:15:14.395138604 +0000 UTC m=+0.362193900 container remove 21b1326f57855b147efcb07cbd10c5f7b7edbabb6693b7752105182196a3f742 (image=quay.io/linux-system-roles/mysql:5.6, name=quadlet-demo-mysql, PODMAN_SYSTEMD_UNIT=quadlet-demo-mysql.service) Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: quadlet-demo-mysql.service: Main process exited, code=exited, status=137/n/a Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: quadlet-demo-mysql.service: Failed with result 'exit-code'. Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: quadlet-demo-mysql.service: Consumed 3.750s CPU time. + systemctl status quadlet-demo-mysql.service × quadlet-demo-mysql.service Loaded: not-found (Reason: Unit quadlet-demo-mysql.service not found.) Active: failed (Result: exit-code) since Wed 2024-07-10 05:15:14 UTC; 26s ago Duration: 1min 1.221s Main PID: 8040 (code=exited, status=137) CPU: 3.750s Jul 10 05:14:24 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:24 1 [Note] Server socket created on IP: '::'. Jul 10 05:14:24 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:24 1 [Warning] Insecure configuration for --pid-file: Location '/var/run/mysqld' in the path is accessible to all OS users. Consider choosing a different directory. Jul 10 05:14:24 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:24 1 [Warning] 'proxies_priv' entry '@ root@21b1326f5785' ignored in --skip-name-resolve mode. Jul 10 05:14:24 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:24 1 [Note] Event Scheduler: Loaded 0 events Jul 10 05:14:24 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:24 1 [Note] mysqld: ready for connections. Jul 10 05:14:24 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: Version: '5.6.51' socket: '/var/run/mysqld/mysqld.sock' port: 3306 MySQL Community Server (GPL) Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm podman[11232]: 2024-07-10 05:15:14.395138604 +0000 UTC m=+0.362193900 container remove 21b1326f57855b147efcb07cbd10c5f7b7edbabb6693b7752105182196a3f742 (image=quay.io/linux-system-roles/mysql:5.6, name=quadlet-demo-mysql, PODMAN_SYSTEMD_UNIT=quadlet-demo-mysql.service) Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: quadlet-demo-mysql.service: Main process exited, code=exited, status=137/n/a Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: quadlet-demo-mysql.service: Failed with result 'exit-code'. Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: quadlet-demo-mysql.service: Consumed 3.750s CPU time. + : TASK [Get journald] ************************************************************ task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/tests/tests_quadlet_demo.yml:193 Wednesday 10 July 2024 05:15:40 +0000 (0:00:00.828) 0:03:45.197 ******** fatal: [sut]: FAILED! => { "changed": false, "cmd": [ "journalctl", "-ex" ], "delta": "0:00:00.036562", "end": "2024-07-10 05:15:40.762072", "failed_when_result": true, "rc": 0, "start": "2024-07-10 05:15:40.725510" } STDOUT: Jul 10 05:14:54 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:14:54 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:14:54 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:14:54 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=10021 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success' Jul 10 05:14:54 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:14:54 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:14:55 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:14:55 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:14:55 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=10045 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success' Jul 10 05:14:55 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:14:55 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:14:55 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:14:55 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:14:55 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=10069 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success' Jul 10 05:14:55 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:14:55 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:14:56 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:14:56 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:14:56 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=10093 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success' Jul 10 05:14:56 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[10109]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None Jul 10 05:14:56 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:14:56 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:14:56 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:14:56 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:14:56 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=10111 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success' Jul 10 05:14:56 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[10127]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 10 05:14:56 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:14:56 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=10130 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success' Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[10146]: ansible-systemd Invoked with name=quadlet-demo.service scope=system state=stopped enabled=False force=True daemon_reload=False daemon_reexec=False no_block=False masked=None Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: Reloading requested from client PID 10149 ('systemctl') (unit session-6.scope)... Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: Reloading... Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: /run/systemd/transient/libpod-dce8c2f6a9a9de70b22786996e3a7fd320907983dffeb24910cd53361f6a0144.scope:14: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: /run/systemd/transient/libpod-af5eaff28e9fad6d7bd929af0607de207b40ae227252858bca105cbe91dcf058.scope:14: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: Reloading finished in 407 ms. Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=315 op=LOAD Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=292 op=UNLOAD Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=316 op=LOAD Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=293 op=UNLOAD Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=317 op=LOAD Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=318 op=LOAD Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=294 op=UNLOAD Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=295 op=UNLOAD Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=319 op=LOAD Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=296 op=UNLOAD Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=320 op=LOAD Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=297 op=UNLOAD Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=321 op=LOAD Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=322 op=LOAD Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=298 op=UNLOAD Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=299 op=UNLOAD Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=323 op=LOAD Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=300 op=UNLOAD Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=324 op=LOAD Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=325 op=LOAD Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=301 op=UNLOAD Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=302 op=UNLOAD Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=326 op=LOAD Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=303 op=UNLOAD Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=327 op=LOAD Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=304 op=UNLOAD Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=328 op=LOAD Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=305 op=UNLOAD Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=329 op=LOAD Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=330 op=LOAD Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=306 op=UNLOAD Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=307 op=UNLOAD Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=331 op=LOAD Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=308 op=UNLOAD Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=332 op=LOAD Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=333 op=LOAD Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=309 op=UNLOAD Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=310 op=UNLOAD Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: Stopping quadlet-demo.service... ░░ Subject: A stop job for unit quadlet-demo.service has begun execution ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ A stop job for unit quadlet-demo.service has begun execution. ░░ ░░ The job identifier is 3523. Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: libpod-9de2222e9853a1d7c250e11a1e4975a817569c5e9b9c202c5e84f41b7332711a.scope: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ The unit libpod-9de2222e9853a1d7c250e11a1e4975a817569c5e9b9c202c5e84f41b7332711a.scope has successfully entered the 'dead' state. Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=311 op=UNLOAD Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm podman[10200]: 2024-07-10 05:14:57.965328709 +0000 UTC m=+0.031740190 container died 9de2222e9853a1d7c250e11a1e4975a817569c5e9b9c202c5e84f41b7332711a (image=localhost/podman-pause:4.9.4-1711445992, name=a96f3a51b8d1-service, PODMAN_SYSTEMD_UNIT=quadlet-demo.service) Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: var-lib-containers-storage-overlay-2b7d88ad00bd6e4b28025299b957f2c60bd7968d96b676e2b876667ee5d7bc54-merged.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ The unit var-lib-containers-storage-overlay-2b7d88ad00bd6e4b28025299b957f2c60bd7968d96b676e2b876667ee5d7bc54-merged.mount has successfully entered the 'dead' state. Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-9de2222e9853a1d7c250e11a1e4975a817569c5e9b9c202c5e84f41b7332711a-userdata-shm.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ The unit var-lib-containers-storage-overlay\x2dcontainers-9de2222e9853a1d7c250e11a1e4975a817569c5e9b9c202c5e84f41b7332711a-userdata-shm.mount has successfully entered the 'dead' state. Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm podman[10200]: 2024-07-10 05:14:58.00693557 +0000 UTC m=+0.073347244 container cleanup 9de2222e9853a1d7c250e11a1e4975a817569c5e9b9c202c5e84f41b7332711a (image=localhost/podman-pause:4.9.4-1711445992, name=a96f3a51b8d1-service, PODMAN_SYSTEMD_UNIT=quadlet-demo.service) Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm podman[10208]: 2024-07-10 05:14:58.050049719 +0000 UTC m=+0.028193783 pod stop da66443b0d18df866aa4578267e9465681155ec0f758ec8d84214580c2123234 (image=, name=quadlet-demo) Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: libpod-ee4bc05257fc811692ed315d4bc9e3e8815ce0788437369ad146bb547ed5565f.scope: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ The unit libpod-ee4bc05257fc811692ed315d4bc9e3e8815ce0788437369ad146bb547ed5565f.scope has successfully entered the 'dead' state. Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=312 op=UNLOAD Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm podman[10208]: 2024-07-10 05:14:58.081849733 +0000 UTC m=+0.059993869 container died ee4bc05257fc811692ed315d4bc9e3e8815ce0788437369ad146bb547ed5565f (image=localhost/podman-pause:4.9.4-1711445992, name=da66443b0d18-infra, PODMAN_SYSTEMD_UNIT=quadlet-demo.service, io.buildah.version=1.33.7) Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: libpod-dce8c2f6a9a9de70b22786996e3a7fd320907983dffeb24910cd53361f6a0144.scope: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ The unit libpod-dce8c2f6a9a9de70b22786996e3a7fd320907983dffeb24910cd53361f6a0144.scope has successfully entered the 'dead' state. Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=314 op=UNLOAD Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm podman[10208]: 2024-07-10 05:14:58.113820223 +0000 UTC m=+0.091964394 container died dce8c2f6a9a9de70b22786996e3a7fd320907983dffeb24910cd53361f6a0144 (image=quay.io/linux-system-roles/envoyproxy:v1.25.0, name=quadlet-demo-envoy, PODMAN_SYSTEMD_UNIT=quadlet-demo.service) Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: libpod-af5eaff28e9fad6d7bd929af0607de207b40ae227252858bca105cbe91dcf058.scope: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ The unit libpod-af5eaff28e9fad6d7bd929af0607de207b40ae227252858bca105cbe91dcf058.scope has successfully entered the 'dead' state. Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: libpod-af5eaff28e9fad6d7bd929af0607de207b40ae227252858bca105cbe91dcf058.scope: Consumed 1.176s CPU time. ░░ Subject: Resources consumed by unit runtime ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ The unit libpod-af5eaff28e9fad6d7bd929af0607de207b40ae227252858bca105cbe91dcf058.scope completed and consumed the indicated resources. Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=313 op=UNLOAD Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm kernel: podman1: port 2(veth1) entered disabled state Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm kernel: veth1 (unregistering): left allmulticast mode Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm kernel: veth1 (unregistering): left promiscuous mode Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm kernel: podman1: port 2(veth1) entered disabled state Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: ANOM_PROMISCUOUS dev=veth1 prom=0 old_prom=256 auid=4294967295 uid=0 gid=0 ses=4294967295 Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm podman[10208]: 2024-07-10 05:14:58.144771474 +0000 UTC m=+0.122915565 container died af5eaff28e9fad6d7bd929af0607de207b40ae227252858bca105cbe91dcf058 (image=quay.io/linux-system-roles/wordpress:4.8-apache, name=quadlet-demo-wordpress, PODMAN_SYSTEMD_UNIT=quadlet-demo.service) Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[10247]: NETFILTER_CFG table=nat:42 family=2 entries=1 op=nft_unregister_rule pid=10247 subj=system_u:system_r:iptables_t:s0 comm="iptables" Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[10250]: NETFILTER_CFG table=nat:43 family=2 entries=1 op=nft_unregister_rule pid=10250 subj=system_u:system_r:iptables_t:s0 comm="iptables" Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm podman[10208]: 2024-07-10 05:14:58.218821937 +0000 UTC m=+0.196965972 container cleanup dce8c2f6a9a9de70b22786996e3a7fd320907983dffeb24910cd53361f6a0144 (image=quay.io/linux-system-roles/envoyproxy:v1.25.0, name=quadlet-demo-envoy, pod_id=da66443b0d18df866aa4578267e9465681155ec0f758ec8d84214580c2123234, PODMAN_SYSTEMD_UNIT=quadlet-demo.service) Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[10254]: NETFILTER_CFG table=nat:44 family=2 entries=1 op=nft_unregister_rule pid=10254 subj=system_u:system_r:iptables_t:s0 comm="iptables" Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm podman[10208]: 2024-07-10 05:14:58.224499975 +0000 UTC m=+0.202643952 container cleanup af5eaff28e9fad6d7bd929af0607de207b40ae227252858bca105cbe91dcf058 (image=quay.io/linux-system-roles/wordpress:4.8-apache, name=quadlet-demo-wordpress, pod_id=da66443b0d18df866aa4578267e9465681155ec0f758ec8d84214580c2123234, PODMAN_SYSTEMD_UNIT=quadlet-demo.service) Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[10256]: NETFILTER_CFG table=nat:45 family=2 entries=1 op=nft_unregister_rule pid=10256 subj=system_u:system_r:iptables_t:s0 comm="iptables" Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[10259]: NETFILTER_CFG table=nat:46 family=2 entries=1 op=nft_unregister_rule pid=10259 subj=system_u:system_r:iptables_t:s0 comm="iptables" Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[10263]: NETFILTER_CFG table=nat:47 family=2 entries=1 op=nft_unregister_rule pid=10263 subj=system_u:system_r:iptables_t:s0 comm="iptables" Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[10265]: NETFILTER_CFG table=nat:48 family=2 entries=1 op=nft_unregister_rule pid=10265 subj=system_u:system_r:iptables_t:s0 comm="iptables" Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[10267]: NETFILTER_CFG table=nat:49 family=2 entries=1 op=nft_unregister_rule pid=10267 subj=system_u:system_r:iptables_t:s0 comm="iptables" Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm podman[10208]: 2024-07-10 05:14:58.300925855 +0000 UTC m=+0.279069830 container cleanup ee4bc05257fc811692ed315d4bc9e3e8815ce0788437369ad146bb547ed5565f (image=localhost/podman-pause:4.9.4-1711445992, name=da66443b0d18-infra, pod_id=da66443b0d18df866aa4578267e9465681155ec0f758ec8d84214580c2123234, io.buildah.version=1.33.7, PODMAN_SYSTEMD_UNIT=quadlet-demo.service) Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: Removed slice machine-libpod_pod_da66443b0d18df866aa4578267e9465681155ec0f758ec8d84214580c2123234.slice - cgroup machine-libpod_pod_da66443b0d18df866aa4578267e9465681155ec0f758ec8d84214580c2123234.slice. ░░ Subject: A stop job for unit machine-libpod_pod_da66443b0d18df866aa4578267e9465681155ec0f758ec8d84214580c2123234.slice has finished ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ A stop job for unit machine-libpod_pod_da66443b0d18df866aa4578267e9465681155ec0f758ec8d84214580c2123234.slice has finished. ░░ ░░ The job identifier is 3525 and the job result is done. Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: machine-libpod_pod_da66443b0d18df866aa4578267e9465681155ec0f758ec8d84214580c2123234.slice: Consumed 1.276s CPU time. ░░ Subject: Resources consumed by unit runtime ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ The unit machine-libpod_pod_da66443b0d18df866aa4578267e9465681155ec0f758ec8d84214580c2123234.slice completed and consumed the indicated resources. Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm podman[10208]: 2024-07-10 05:14:58.311544683 +0000 UTC m=+0.289688767 pod stop da66443b0d18df866aa4578267e9465681155ec0f758ec8d84214580c2123234 (image=, name=quadlet-demo) Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: machine-libpod_pod_da66443b0d18df866aa4578267e9465681155ec0f758ec8d84214580c2123234.slice: Failed to open /run/systemd/transient/machine-libpod_pod_da66443b0d18df866aa4578267e9465681155ec0f758ec8d84214580c2123234.slice: No such file or directory Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo[10208]: time="2024-07-10T05:14:58Z" level=error msg="Checking if infra needs to be stopped: removing pod da66443b0d18df866aa4578267e9465681155ec0f758ec8d84214580c2123234 cgroup: Unit machine-libpod_pod_da66443b0d18df866aa4578267e9465681155ec0f758ec8d84214580c2123234.slice not loaded." Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm podman[10208]: 2024-07-10 05:14:58.381480436 +0000 UTC m=+0.359624442 container remove dce8c2f6a9a9de70b22786996e3a7fd320907983dffeb24910cd53361f6a0144 (image=quay.io/linux-system-roles/envoyproxy:v1.25.0, name=quadlet-demo-envoy, pod_id=da66443b0d18df866aa4578267e9465681155ec0f758ec8d84214580c2123234, PODMAN_SYSTEMD_UNIT=quadlet-demo.service) Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm podman[10208]: 2024-07-10 05:14:58.400712486 +0000 UTC m=+0.378856510 container remove af5eaff28e9fad6d7bd929af0607de207b40ae227252858bca105cbe91dcf058 (image=quay.io/linux-system-roles/wordpress:4.8-apache, name=quadlet-demo-wordpress, pod_id=da66443b0d18df866aa4578267e9465681155ec0f758ec8d84214580c2123234, PODMAN_SYSTEMD_UNIT=quadlet-demo.service) Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm podman[10208]: 2024-07-10 05:14:58.42903182 +0000 UTC m=+0.407175829 container remove ee4bc05257fc811692ed315d4bc9e3e8815ce0788437369ad146bb547ed5565f (image=localhost/podman-pause:4.9.4-1711445992, name=da66443b0d18-infra, pod_id=da66443b0d18df866aa4578267e9465681155ec0f758ec8d84214580c2123234, PODMAN_SYSTEMD_UNIT=quadlet-demo.service, io.buildah.version=1.33.7) Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: machine-libpod_pod_da66443b0d18df866aa4578267e9465681155ec0f758ec8d84214580c2123234.slice: Failed to open /run/systemd/transient/machine-libpod_pod_da66443b0d18df866aa4578267e9465681155ec0f758ec8d84214580c2123234.slice: No such file or directory Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm podman[10208]: 2024-07-10 05:14:58.437391183 +0000 UTC m=+0.415535161 pod remove da66443b0d18df866aa4578267e9465681155ec0f758ec8d84214580c2123234 (image=, name=quadlet-demo) Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm podman[10208]: 2024-07-10 05:14:58.456671982 +0000 UTC m=+0.434815999 container remove 9de2222e9853a1d7c250e11a1e4975a817569c5e9b9c202c5e84f41b7332711a (image=localhost/podman-pause:4.9.4-1711445992, name=a96f3a51b8d1-service, PODMAN_SYSTEMD_UNIT=quadlet-demo.service) Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo[10208]: Pods stopped: Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo[10208]: da66443b0d18df866aa4578267e9465681155ec0f758ec8d84214580c2123234 Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo[10208]: Pods removed: Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo[10208]: Error: removing pod da66443b0d18df866aa4578267e9465681155ec0f758ec8d84214580c2123234 cgroup: removing pod da66443b0d18df866aa4578267e9465681155ec0f758ec8d84214580c2123234 cgroup: Unit machine-libpod_pod_da66443b0d18df866aa4578267e9465681155ec0f758ec8d84214580c2123234.slice not loaded. Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo[10208]: Secrets removed: Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo[10208]: Error: %!s() Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo[10208]: Volumes removed: Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo[10208]: time="2024-07-10T05:14:58Z" level=error msg="Checking whether service of container 9de2222e9853a1d7c250e11a1e4975a817569c5e9b9c202c5e84f41b7332711a can be stopped: no container with ID 9de2222e9853a1d7c250e11a1e4975a817569c5e9b9c202c5e84f41b7332711a found in database: no such container" Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo[10208]: time="2024-07-10T05:14:58Z" level=error msg="Checking whether service of container 9de2222e9853a1d7c250e11a1e4975a817569c5e9b9c202c5e84f41b7332711a can be stopped: no container with ID 9de2222e9853a1d7c250e11a1e4975a817569c5e9b9c202c5e84f41b7332711a found in database: no such container" Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: quadlet-demo.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ The unit quadlet-demo.service has successfully entered the 'dead' state. Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: Stopped quadlet-demo.service. ░░ Subject: A stop job for unit quadlet-demo.service has finished ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ A stop job for unit quadlet-demo.service has finished. ░░ ░░ The job identifier is 3523 and the job result is done. Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:init_t:s0 msg='unit=quadlet-demo comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=10271 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success' Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[10287]: ansible-file Invoked with path=/etc/containers/systemd/quadlet-demo.kube state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=10288 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success' Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: tmp-crun.HDFMzq.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ The unit tmp-crun.HDFMzq.mount has successfully entered the 'dead' state. Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: var-lib-containers-storage-overlay-3016041ec40d0362707a4f866fdaa64e34e3c1e1999e0d6d8bd640ecbb918ce8-merged.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ The unit var-lib-containers-storage-overlay-3016041ec40d0362707a4f866fdaa64e34e3c1e1999e0d6d8bd640ecbb918ce8-merged.mount has successfully entered the 'dead' state. Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: var-lib-containers-storage-overlay-859ab31e620e0a4681f0f8f3ef5d32c5fe4e820f6d74606719739fdbfd66173e-merged.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ The unit var-lib-containers-storage-overlay-859ab31e620e0a4681f0f8f3ef5d32c5fe4e820f6d74606719739fdbfd66173e-merged.mount has successfully entered the 'dead' state. Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: var-lib-containers-storage-overlay-bac024e2d3f38a9994be6ed0f1964c96696ceaf11d0399758a993d5026385fae-merged.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ The unit var-lib-containers-storage-overlay-bac024e2d3f38a9994be6ed0f1964c96696ceaf11d0399758a993d5026385fae-merged.mount has successfully entered the 'dead' state. Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-ee4bc05257fc811692ed315d4bc9e3e8815ce0788437369ad146bb547ed5565f-userdata-shm.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ The unit var-lib-containers-storage-overlay\x2dcontainers-ee4bc05257fc811692ed315d4bc9e3e8815ce0788437369ad146bb547ed5565f-userdata-shm.mount has successfully entered the 'dead' state. Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: run-netns-netns\x2dc9c42157\x2d2a14\x2df2d5\x2d22c0\x2d3e88aa7c59f3.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ The unit run-netns-netns\x2dc9c42157\x2d2a14\x2df2d5\x2d22c0\x2d3e88aa7c59f3.mount has successfully entered the 'dead' state. Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[10304]: ansible-systemd Invoked with daemon_reload=True scope=system daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: Reloading requested from client PID 10305 ('systemctl') (unit session-6.scope)... Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: Reloading... Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: Reloading finished in 403 ms. Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=334 op=LOAD Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=315 op=UNLOAD Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=335 op=LOAD Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=316 op=UNLOAD Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=336 op=LOAD Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=337 op=LOAD Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=317 op=UNLOAD Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=318 op=UNLOAD Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=338 op=LOAD Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=319 op=UNLOAD Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=339 op=LOAD Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=320 op=UNLOAD Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=340 op=LOAD Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=341 op=LOAD Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=321 op=UNLOAD Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=322 op=UNLOAD Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=342 op=LOAD Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=323 op=UNLOAD Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=343 op=LOAD Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=344 op=LOAD Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=324 op=UNLOAD Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=325 op=UNLOAD Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=345 op=LOAD Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=326 op=UNLOAD Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=346 op=LOAD Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=327 op=UNLOAD Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=347 op=LOAD Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=328 op=UNLOAD Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=348 op=LOAD Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=349 op=LOAD Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=329 op=UNLOAD Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=330 op=UNLOAD Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=350 op=LOAD Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=331 op=UNLOAD Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=351 op=LOAD Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=352 op=LOAD Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=332 op=UNLOAD Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=333 op=UNLOAD Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=10355 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success' Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[10371]: ansible-ansible.legacy.command Invoked with _raw_params=podman image prune --all -f _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 10 05:15:00 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm podman[10372]: 2024-07-10 05:15:00.298256034 +0000 UTC m=+0.411657053 image untag fcf3e41b8864a14d75a6d0627d3d02154e28a153aa57e8baa392cd744ffa0d0b quay.io/linux-system-roles/wordpress:4.8-apache Jul 10 05:15:00 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm podman[10372]: 2024-07-10 05:14:59.912089645 +0000 UTC m=+0.025490984 image remove fcf3e41b8864a14d75a6d0627d3d02154e28a153aa57e8baa392cd744ffa0d0b Jul 10 05:15:00 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm podman[10372]: 2024-07-10 05:15:00.410901119 +0000 UTC m=+0.524302167 image untag 5af2585e22ed1562885d9407efab74010090427be79048c2cd6a226517cc1e1d quay.io/linux-system-roles/envoyproxy:v1.25.0 Jul 10 05:15:00 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm podman[10372]: 2024-07-10 05:15:00.29826683 +0000 UTC m=+0.411667812 image remove 5af2585e22ed1562885d9407efab74010090427be79048c2cd6a226517cc1e1d Jul 10 05:15:00 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm podman[10372]: 2024-07-10 05:15:00.417719687 +0000 UTC m=+0.531120696 image untag 6c68feb81ac7fb684f7b4d483de9bffaff110fa95445c248359b77bb5d3b2295 localhost/podman-pause:4.9.4-1711445992 Jul 10 05:15:00 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm podman[10372]: 2024-07-10 05:15:00.410910715 +0000 UTC m=+0.524311688 image remove 6c68feb81ac7fb684f7b4d483de9bffaff110fa95445c248359b77bb5d3b2295 Jul 10 05:15:00 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:00 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:00 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:00 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:00 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=10380 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success' Jul 10 05:15:00 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[10396]: ansible-ansible.legacy.command Invoked with _raw_params=podman images -n _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 10 05:15:00 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:00 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:00 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:00 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:00 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=10404 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success' Jul 10 05:15:00 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[10420]: ansible-ansible.legacy.command Invoked with _raw_params=podman volume ls -n _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 10 05:15:01 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:01 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:01 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:01 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:01 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=10428 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success' Jul 10 05:15:01 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[10444]: ansible-ansible.legacy.command Invoked with _raw_params=podman ps --noheading _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 10 05:15:01 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:01 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:01 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:01 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:01 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=10453 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success' Jul 10 05:15:01 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[10469]: ansible-ansible.legacy.command Invoked with _raw_params=podman network ls -n -q _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 10 05:15:01 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:01 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:01 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:01 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:01 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=10477 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success' Jul 10 05:15:01 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:01 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:01 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:01 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:01 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=10501 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success' Jul 10 05:15:02 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[10517]: ansible-service_facts Invoked Jul 10 05:15:04 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:04 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:04 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:04 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:04 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=10630 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success' Jul 10 05:15:04 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[10646]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None Jul 10 05:15:04 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:04 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:04 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:04 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:04 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=10648 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success' Jul 10 05:15:05 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[10664]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 10 05:15:05 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:05 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:05 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:05 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:05 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=10667 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success' Jul 10 05:15:05 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[10683]: ansible-file Invoked with path=/etc/containers/systemd/quadlet-demo.yml state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Jul 10 05:15:05 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:05 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:05 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:05 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:05 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=10684 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success' Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[10700]: ansible-systemd Invoked with daemon_reload=True scope=system daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: Reloading requested from client PID 10701 ('systemctl') (unit session-6.scope)... Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: Reloading... Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: Reloading finished in 387 ms. Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=353 op=LOAD Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=334 op=UNLOAD Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=354 op=LOAD Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=335 op=UNLOAD Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=355 op=LOAD Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=356 op=LOAD Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=336 op=UNLOAD Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=337 op=UNLOAD Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=357 op=LOAD Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=338 op=UNLOAD Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=358 op=LOAD Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=339 op=UNLOAD Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=359 op=LOAD Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=360 op=LOAD Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=340 op=UNLOAD Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=341 op=UNLOAD Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=361 op=LOAD Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=342 op=UNLOAD Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=362 op=LOAD Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=363 op=LOAD Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=343 op=UNLOAD Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=344 op=UNLOAD Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=364 op=LOAD Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=345 op=UNLOAD Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=365 op=LOAD Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=346 op=UNLOAD Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=366 op=LOAD Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=347 op=UNLOAD Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=367 op=LOAD Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=368 op=LOAD Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=348 op=UNLOAD Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=349 op=UNLOAD Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=369 op=LOAD Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=350 op=UNLOAD Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=370 op=LOAD Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=371 op=LOAD Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=351 op=UNLOAD Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=352 op=UNLOAD Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=10751 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success' Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[10767]: ansible-ansible.legacy.command Invoked with _raw_params=podman image prune --all -f _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:07 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:07 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:07 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=10775 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success' Jul 10 05:15:07 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[10791]: ansible-ansible.legacy.command Invoked with _raw_params=podman images -n _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 10 05:15:07 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:07 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:07 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:07 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:07 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=10800 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success' Jul 10 05:15:07 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[10816]: ansible-ansible.legacy.command Invoked with _raw_params=podman volume ls -n _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 10 05:15:07 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:07 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:07 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:07 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:07 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=10824 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success' Jul 10 05:15:07 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[10840]: ansible-ansible.legacy.command Invoked with _raw_params=podman ps --noheading _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 10 05:15:07 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:07 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:07 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:07 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:07 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=10848 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success' Jul 10 05:15:08 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[10864]: ansible-ansible.legacy.command Invoked with _raw_params=podman network ls -n -q _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 10 05:15:08 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:08 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:08 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:08 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:08 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=10872 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success' Jul 10 05:15:08 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:08 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:08 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:08 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:08 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=10896 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success' Jul 10 05:15:08 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[10912]: ansible-service_facts Invoked Jul 10 05:15:10 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:10 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:10 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:10 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:10 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=11025 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success' Jul 10 05:15:11 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[11041]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None Jul 10 05:15:11 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:11 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:11 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:11 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:11 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=11043 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success' Jul 10 05:15:11 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[11059]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 10 05:15:11 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:11 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:11 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:11 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:11 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=11062 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success' Jul 10 05:15:12 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[11078]: ansible-file Invoked with path=/etc/containers/systemd/envoy-proxy-configmap.yml state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Jul 10 05:15:12 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:12 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:12 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:12 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:12 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=11079 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success' Jul 10 05:15:12 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[11095]: ansible-systemd Invoked with daemon_reload=True scope=system daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None Jul 10 05:15:12 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: Reloading requested from client PID 11096 ('systemctl') (unit session-6.scope)... Jul 10 05:15:12 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: Reloading... Jul 10 05:15:12 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: Reloading finished in 384 ms. Jul 10 05:15:12 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=372 op=LOAD Jul 10 05:15:12 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=353 op=UNLOAD Jul 10 05:15:12 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=373 op=LOAD Jul 10 05:15:12 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=354 op=UNLOAD Jul 10 05:15:12 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=374 op=LOAD Jul 10 05:15:12 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=375 op=LOAD Jul 10 05:15:12 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=355 op=UNLOAD Jul 10 05:15:12 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=356 op=UNLOAD Jul 10 05:15:12 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=376 op=LOAD Jul 10 05:15:12 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=357 op=UNLOAD Jul 10 05:15:12 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=377 op=LOAD Jul 10 05:15:12 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=358 op=UNLOAD Jul 10 05:15:12 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=378 op=LOAD Jul 10 05:15:12 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=379 op=LOAD Jul 10 05:15:12 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=359 op=UNLOAD Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=360 op=UNLOAD Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=380 op=LOAD Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=361 op=UNLOAD Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=381 op=LOAD Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=382 op=LOAD Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=362 op=UNLOAD Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=363 op=UNLOAD Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=383 op=LOAD Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=364 op=UNLOAD Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=384 op=LOAD Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=365 op=UNLOAD Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=385 op=LOAD Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=366 op=UNLOAD Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=386 op=LOAD Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=387 op=LOAD Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=367 op=UNLOAD Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=368 op=UNLOAD Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=388 op=LOAD Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=369 op=UNLOAD Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=389 op=LOAD Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=390 op=LOAD Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=370 op=UNLOAD Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=371 op=UNLOAD Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=11146 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success' Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[11162]: ansible-ansible.legacy.command Invoked with _raw_params=podman image prune --all -f _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=11171 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success' Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[11187]: ansible-ansible.legacy.command Invoked with _raw_params=podman images -n _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=11195 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success' Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[11211]: ansible-ansible.legacy.command Invoked with _raw_params=podman volume ls -n _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:init_t:s0 msg='unit=21b1326f57855b147efcb07cbd10c5f7b7edbabb6693b7752105182196a3f742 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm podman[11212]: 2024-07-10 05:15:13.983498841 +0000 UTC m=+0.068110829 container health_status 21b1326f57855b147efcb07cbd10c5f7b7edbabb6693b7752105182196a3f742 (image=quay.io/linux-system-roles/mysql:5.6, name=quadlet-demo-mysql, health_status=unhealthy, PODMAN_SYSTEMD_UNIT=quadlet-demo-mysql.service) Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=11228 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success' Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: 21b1326f57855b147efcb07cbd10c5f7b7edbabb6693b7752105182196a3f742.service: Main process exited, code=exited, status=125/n/a ░░ Subject: Unit process exited ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ An ExecStart= process belonging to unit 21b1326f57855b147efcb07cbd10c5f7b7edbabb6693b7752105182196a3f742.service has exited. ░░ ░░ The process' exit code is 'exited' and its exit status is 125. Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: 21b1326f57855b147efcb07cbd10c5f7b7edbabb6693b7752105182196a3f742.service: Failed with result 'exit-code'. ░░ Subject: Unit failed ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ The unit 21b1326f57855b147efcb07cbd10c5f7b7edbabb6693b7752105182196a3f742.service has entered the 'failed' state with result 'exit-code'. Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:init_t:s0 msg='unit=21b1326f57855b147efcb07cbd10c5f7b7edbabb6693b7752105182196a3f742 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: 21b1326f57855b147efcb07cbd10c5f7b7edbabb6693b7752105182196a3f742.timer: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ The unit 21b1326f57855b147efcb07cbd10c5f7b7edbabb6693b7752105182196a3f742.timer has successfully entered the 'dead' state. Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: Stopped 21b1326f57855b147efcb07cbd10c5f7b7edbabb6693b7752105182196a3f742.timer - /usr/bin/podman healthcheck run 21b1326f57855b147efcb07cbd10c5f7b7edbabb6693b7752105182196a3f742. ░░ Subject: A stop job for unit 21b1326f57855b147efcb07cbd10c5f7b7edbabb6693b7752105182196a3f742.timer has finished ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ A stop job for unit 21b1326f57855b147efcb07cbd10c5f7b7edbabb6693b7752105182196a3f742.timer has finished. ░░ ░░ The job identifier is 3611 and the job result is done. Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: 21b1326f57855b147efcb07cbd10c5f7b7edbabb6693b7752105182196a3f742.service: Failed to open /run/systemd/transient/21b1326f57855b147efcb07cbd10c5f7b7edbabb6693b7752105182196a3f742.service: No such file or directory Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: run-r9a5a2edf77764b9e940f4c2cad5029cc.scope: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ The unit run-r9a5a2edf77764b9e940f4c2cad5029cc.scope has successfully entered the 'dead' state. Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm kernel: podman1: port 1(veth0) entered disabled state Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm kernel: veth0 (unregistering): left allmulticast mode Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm kernel: veth0 (unregistering): left promiscuous mode Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm kernel: podman1: port 1(veth0) entered disabled state Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: ANOM_PROMISCUOUS dev=veth0 prom=0 old_prom=256 auid=4294967295 uid=0 gid=0 ses=4294967295 Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm NetworkManager[657]: [1720588514.1347] device (podman1): state change: activated -> unmanaged (reason 'unmanaged', sys-iface-state: 'removed') Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[11260]: NETFILTER_CFG table=nat:50 family=2 entries=1 op=nft_unregister_rule pid=11260 subj=system_u:system_r:iptables_t:s0 comm="iptables" Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[11266]: NETFILTER_CFG table=nat:51 family=2 entries=1 op=nft_unregister_rule pid=11266 subj=system_u:system_r:iptables_t:s0 comm="iptables" Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: Starting NetworkManager-dispatcher.service - Network Manager Script Dispatcher Service... ░░ Subject: A start job for unit NetworkManager-dispatcher.service has begun execution ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ A start job for unit NetworkManager-dispatcher.service has begun execution. ░░ ░░ The job identifier is 3613. Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[11268]: NETFILTER_CFG table=nat:52 family=2 entries=1 op=nft_unregister_rule pid=11268 subj=system_u:system_r:iptables_t:s0 comm="iptables" Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[11270]: NETFILTER_CFG table=filter:53 family=2 entries=1 op=nft_unregister_rule pid=11270 subj=system_u:system_r:iptables_t:s0 comm="iptables" Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[11276]: NETFILTER_CFG table=filter:54 family=2 entries=1 op=nft_unregister_rule pid=11276 subj=system_u:system_r:iptables_t:s0 comm="iptables" Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: Started NetworkManager-dispatcher.service - Network Manager Script Dispatcher Service. ░░ Subject: A start job for unit NetworkManager-dispatcher.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ A start job for unit NetworkManager-dispatcher.service has finished successfully. ░░ ░░ The job identifier is 3613. Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:init_t:s0 msg='unit=NetworkManager-dispatcher comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[11283]: NETFILTER_CFG table=filter:55 family=2 entries=1 op=nft_unregister_rule pid=11283 subj=system_u:system_r:iptables_t:s0 comm="iptables" Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[11285]: NETFILTER_CFG table=filter:56 family=2 entries=1 op=nft_unregister_rule pid=11285 subj=system_u:system_r:iptables_t:s0 comm="iptables" Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[11287]: NETFILTER_CFG table=nat:57 family=2 entries=1 op=nft_unregister_chain pid=11287 subj=system_u:system_r:iptables_t:s0 comm="iptables" Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[11252]: ansible-ansible.legacy.command Invoked with _raw_params=podman ps --noheading _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[6763]: NETFILTER_CFG table=firewalld:58 family=1 entries=46 op=nft_unregister_rule pid=6763 subj=system_u:system_r:firewalld_t:s0 comm="firewalld" Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: run-netns-netns\x2de51cc4f9\x2d239a\x2d1eda\x2d2116\x2d2ee6febc9b1b.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ The unit run-netns-netns\x2de51cc4f9\x2d239a\x2d1eda\x2d2116\x2d2ee6febc9b1b.mount has successfully entered the 'dead' state. Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=253 op=UNLOAD Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: var-lib-containers-storage-overlay-365acc50bbd6897f3aabf4b51e9c692de2d02cbffe3822f80e928de6dd72014d-merged.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ The unit var-lib-containers-storage-overlay-365acc50bbd6897f3aabf4b51e9c692de2d02cbffe3822f80e928de6dd72014d-merged.mount has successfully entered the 'dead' state. Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-21b1326f57855b147efcb07cbd10c5f7b7edbabb6693b7752105182196a3f742-userdata-shm.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ The unit var-lib-containers-storage-overlay\x2dcontainers-21b1326f57855b147efcb07cbd10c5f7b7edbabb6693b7752105182196a3f742-userdata-shm.mount has successfully entered the 'dead' state. Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm podman[11232]: 2024-07-10 05:15:14.395138604 +0000 UTC m=+0.362193900 container remove 21b1326f57855b147efcb07cbd10c5f7b7edbabb6693b7752105182196a3f742 (image=quay.io/linux-system-roles/mysql:5.6, name=quadlet-demo-mysql, PODMAN_SYSTEMD_UNIT=quadlet-demo-mysql.service) Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: quadlet-demo-mysql.service: Main process exited, code=exited, status=137/n/a ░░ Subject: Unit process exited ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ An ExecStart= process belonging to unit quadlet-demo-mysql.service has exited. ░░ ░░ The process' exit code is 'exited' and its exit status is 137. Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: quadlet-demo-mysql.service: Failed with result 'exit-code'. ░░ Subject: Unit failed ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ The unit quadlet-demo-mysql.service has entered the 'failed' state with result 'exit-code'. Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:init_t:s0 msg='unit=quadlet-demo-mysql comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: quadlet-demo-mysql.service: Consumed 3.750s CPU time. ░░ Subject: Resources consumed by unit runtime ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ The unit quadlet-demo-mysql.service completed and consumed the indicated resources. Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=11307 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success' Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[11323]: ansible-ansible.legacy.command Invoked with _raw_params=podman network ls -n -q _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=11331 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success' Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=11356 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success' Jul 10 05:15:15 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[11372]: ansible-service_facts Invoked Jul 10 05:15:17 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:17 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:17 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:17 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:17 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=11484 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success' Jul 10 05:15:17 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[11500]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None Jul 10 05:15:17 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:17 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:17 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:17 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:17 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=11502 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success' Jul 10 05:15:17 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[11518]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 10 05:15:17 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:17 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=11521 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success' Jul 10 05:15:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[11537]: ansible-systemd Invoked with name=quadlet-demo-mysql.service scope=system state=stopped enabled=False force=True daemon_reload=False daemon_reexec=False no_block=False masked=None Jul 10 05:15:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: Reloading requested from client PID 11540 ('systemctl') (unit session-6.scope)... Jul 10 05:15:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: Reloading... Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: Reloading finished in 380 ms. Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=391 op=LOAD Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=372 op=UNLOAD Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=392 op=LOAD Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=373 op=UNLOAD Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=393 op=LOAD Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=394 op=LOAD Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=374 op=UNLOAD Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=375 op=UNLOAD Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=395 op=LOAD Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=376 op=UNLOAD Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=396 op=LOAD Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=377 op=UNLOAD Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=397 op=LOAD Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=398 op=LOAD Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=378 op=UNLOAD Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=379 op=UNLOAD Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=399 op=LOAD Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=380 op=UNLOAD Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=400 op=LOAD Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=401 op=LOAD Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=381 op=UNLOAD Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=382 op=UNLOAD Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=402 op=LOAD Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=383 op=UNLOAD Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=403 op=LOAD Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=384 op=UNLOAD Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=404 op=LOAD Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=385 op=UNLOAD Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=405 op=LOAD Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=406 op=LOAD Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=386 op=UNLOAD Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=387 op=UNLOAD Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=407 op=LOAD Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=388 op=UNLOAD Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=408 op=LOAD Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=409 op=LOAD Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=389 op=UNLOAD Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=390 op=UNLOAD Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=11591 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success' Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[11607]: ansible-file Invoked with path=/etc/containers/systemd/quadlet-demo-mysql.container state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=11608 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success' Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[11624]: ansible-systemd Invoked with daemon_reload=True scope=system daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: Reloading requested from client PID 11625 ('systemctl') (unit session-6.scope)... Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: Reloading... Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: Reloading finished in 372 ms. Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=410 op=LOAD Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=391 op=UNLOAD Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=411 op=LOAD Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=392 op=UNLOAD Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=412 op=LOAD Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=413 op=LOAD Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=393 op=UNLOAD Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=394 op=UNLOAD Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=414 op=LOAD Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=395 op=UNLOAD Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=415 op=LOAD Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=396 op=UNLOAD Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=416 op=LOAD Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=417 op=LOAD Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=397 op=UNLOAD Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=398 op=UNLOAD Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=418 op=LOAD Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=399 op=UNLOAD Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=419 op=LOAD Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=420 op=LOAD Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=400 op=UNLOAD Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=401 op=UNLOAD Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=421 op=LOAD Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=402 op=UNLOAD Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=422 op=LOAD Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=403 op=UNLOAD Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=423 op=LOAD Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=404 op=UNLOAD Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=424 op=LOAD Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=425 op=LOAD Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=405 op=UNLOAD Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=406 op=UNLOAD Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=426 op=LOAD Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=407 op=UNLOAD Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=427 op=LOAD Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=428 op=LOAD Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=408 op=UNLOAD Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=409 op=UNLOAD Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=11676 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success' Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[11692]: ansible-ansible.legacy.command Invoked with _raw_params=podman image prune --all -f _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm podman[11693]: 2024-07-10 05:15:20.967347867 +0000 UTC m=+0.212576529 image untag dd3b2a5dcb48ff61113592ed5ddd762581be4387c7bc552375a2159422aa6bf5 quay.io/linux-system-roles/mysql:5.6 Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm podman[11693]: 2024-07-10 05:15:20.779292153 +0000 UTC m=+0.024520898 image remove dd3b2a5dcb48ff61113592ed5ddd762581be4387c7bc552375a2159422aa6bf5 Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=11701 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success' Jul 10 05:15:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[11717]: ansible-ansible.legacy.command Invoked with _raw_params=podman images -n _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 10 05:15:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Jul 10 05:15:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=11726 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success' Jul 10 05:15:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[11742]: ansible-ansible.legacy.command Invoked with _raw_params=podman volume ls -n _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 10 05:15:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=11750 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success' Jul 10 05:15:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Jul 10 05:15:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[11766]: ansible-ansible.legacy.command Invoked with _raw_params=podman ps --noheading _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 10 05:15:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Jul 10 05:15:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=11774 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success' Jul 10 05:15:22 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[11790]: ansible-ansible.legacy.command Invoked with _raw_params=podman network ls -n -q _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 10 05:15:22 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Jul 10 05:15:22 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:22 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:22 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:22 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:22 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=11798 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success' Jul 10 05:15:22 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:22 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:22 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:22 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:22 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=11823 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success' Jul 10 05:15:22 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[11839]: ansible-service_facts Invoked Jul 10 05:15:24 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: NetworkManager-dispatcher.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ The unit NetworkManager-dispatcher.service has successfully entered the 'dead' state. Jul 10 05:15:24 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:init_t:s0 msg='unit=NetworkManager-dispatcher comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 10 05:15:24 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:24 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:24 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:24 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:24 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=11952 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success' Jul 10 05:15:25 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[11968]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None Jul 10 05:15:25 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:25 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:25 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:25 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:25 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=11970 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success' Jul 10 05:15:25 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[11986]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 10 05:15:25 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:25 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:25 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:25 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:25 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=11989 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success' Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[12005]: ansible-systemd Invoked with name=quadlet-demo-mysql-volume.service scope=system state=stopped enabled=False force=True daemon_reload=False daemon_reexec=False no_block=False masked=None Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: Reloading requested from client PID 12008 ('systemctl') (unit session-6.scope)... Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: Reloading... Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: Reloading finished in 370 ms. Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=429 op=LOAD Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=410 op=UNLOAD Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=430 op=LOAD Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=411 op=UNLOAD Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=431 op=LOAD Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=432 op=LOAD Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=412 op=UNLOAD Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=413 op=UNLOAD Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=433 op=LOAD Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=414 op=UNLOAD Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=434 op=LOAD Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=415 op=UNLOAD Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=435 op=LOAD Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=436 op=LOAD Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=416 op=UNLOAD Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=417 op=UNLOAD Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=437 op=LOAD Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=418 op=UNLOAD Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=438 op=LOAD Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=439 op=LOAD Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=419 op=UNLOAD Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=420 op=UNLOAD Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=440 op=LOAD Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=421 op=UNLOAD Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=441 op=LOAD Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=422 op=UNLOAD Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=442 op=LOAD Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=423 op=UNLOAD Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=443 op=LOAD Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=444 op=LOAD Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=424 op=UNLOAD Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=425 op=UNLOAD Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=445 op=LOAD Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=426 op=UNLOAD Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=446 op=LOAD Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=447 op=LOAD Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=427 op=UNLOAD Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=428 op=UNLOAD Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: quadlet-demo-mysql-volume.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ The unit quadlet-demo-mysql-volume.service has successfully entered the 'dead' state. Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: Stopped quadlet-demo-mysql-volume.service. ░░ Subject: A stop job for unit quadlet-demo-mysql-volume.service has finished ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ A stop job for unit quadlet-demo-mysql-volume.service has finished. ░░ ░░ The job identifier is 3699 and the job result is done. Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:init_t:s0 msg='unit=quadlet-demo-mysql-volume comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=12059 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success' Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[12075]: ansible-file Invoked with path=/etc/containers/systemd/quadlet-demo-mysql.volume state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=12076 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success' Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[12092]: ansible-systemd Invoked with daemon_reload=True scope=system daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: Reloading requested from client PID 12093 ('systemctl') (unit session-6.scope)... Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: Reloading... Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: Reloading finished in 363 ms. Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=448 op=LOAD Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=429 op=UNLOAD Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=449 op=LOAD Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=430 op=UNLOAD Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=450 op=LOAD Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=451 op=LOAD Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=431 op=UNLOAD Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=432 op=UNLOAD Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=452 op=LOAD Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=433 op=UNLOAD Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=453 op=LOAD Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=434 op=UNLOAD Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=454 op=LOAD Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=455 op=LOAD Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=435 op=UNLOAD Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=436 op=UNLOAD Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=456 op=LOAD Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=437 op=UNLOAD Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=457 op=LOAD Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=458 op=LOAD Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=438 op=UNLOAD Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=439 op=UNLOAD Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=459 op=LOAD Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=440 op=UNLOAD Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=460 op=LOAD Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=441 op=UNLOAD Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=461 op=LOAD Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=442 op=UNLOAD Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=462 op=LOAD Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=463 op=LOAD Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=443 op=UNLOAD Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=444 op=UNLOAD Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=464 op=LOAD Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=445 op=UNLOAD Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=465 op=LOAD Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=466 op=LOAD Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=446 op=UNLOAD Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=447 op=UNLOAD Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=12143 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success' Jul 10 05:15:28 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[12159]: ansible-ansible.legacy.command Invoked with _raw_params=podman image prune --all -f _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 10 05:15:28 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:28 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:28 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:28 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:28 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=12167 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success' Jul 10 05:15:28 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[12183]: ansible-ansible.legacy.command Invoked with _raw_params=podman volume prune -f _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 10 05:15:28 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm podman[12184]: 2024-07-10 05:15:28.410743708 +0000 UTC m=+0.051639497 volume remove systemd-quadlet-demo-mysql Jul 10 05:15:28 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm podman[12184]: 2024-07-10 05:15:28.411190479 +0000 UTC m=+0.052086323 volume prune systemd-quadlet-demo-mysql Jul 10 05:15:28 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm podman[12184]: 2024-07-10 05:15:28.45686457 +0000 UTC m=+0.097760616 volume remove wp-pv-claim Jul 10 05:15:28 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm podman[12184]: 2024-07-10 05:15:28.456915574 +0000 UTC m=+0.097811420 volume prune wp-pv-claim Jul 10 05:15:28 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm podman[12184]: 2024-07-10 05:15:28.470779425 +0000 UTC m=+0.111675245 volume remove envoy-proxy-config Jul 10 05:15:28 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm podman[12184]: 2024-07-10 05:15:28.47083971 +0000 UTC m=+0.111735606 volume prune envoy-proxy-config Jul 10 05:15:28 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm podman[12184]: 2024-07-10 05:15:28.477537993 +0000 UTC m=+0.118433739 volume remove envoy-certificates Jul 10 05:15:28 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm podman[12184]: 2024-07-10 05:15:28.477585664 +0000 UTC m=+0.118481540 volume prune envoy-certificates Jul 10 05:15:28 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Jul 10 05:15:28 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:28 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:28 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:28 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:28 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=12192 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success' Jul 10 05:15:28 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[12208]: ansible-ansible.legacy.command Invoked with _raw_params=podman images -n _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 10 05:15:28 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:28 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:28 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:28 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:28 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=12217 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success' Jul 10 05:15:28 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[12233]: ansible-ansible.legacy.command Invoked with _raw_params=podman volume ls -n _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 10 05:15:29 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:29 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:29 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:29 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:29 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=12242 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success' Jul 10 05:15:29 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[12258]: ansible-ansible.legacy.command Invoked with _raw_params=podman ps --noheading _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 10 05:15:29 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Jul 10 05:15:29 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:29 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:29 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:29 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:29 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=12267 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success' Jul 10 05:15:29 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[12283]: ansible-ansible.legacy.command Invoked with _raw_params=podman network ls -n -q _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 10 05:15:29 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Jul 10 05:15:29 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:29 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:29 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:29 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:29 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=12291 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success' Jul 10 05:15:29 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:29 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:29 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:29 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:29 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=12315 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success' Jul 10 05:15:30 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[12331]: ansible-service_facts Invoked Jul 10 05:15:32 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:32 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:32 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:32 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:32 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=12444 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success' Jul 10 05:15:32 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[12460]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None Jul 10 05:15:32 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:32 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:32 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:32 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:32 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=12462 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success' Jul 10 05:15:32 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[12478]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 10 05:15:32 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:32 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:33 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:33 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:33 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=12481 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success' Jul 10 05:15:33 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[12497]: ansible-systemd Invoked with name=quadlet-demo-network.service scope=system state=stopped enabled=False force=True daemon_reload=False daemon_reexec=False no_block=False masked=None Jul 10 05:15:33 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: Reloading requested from client PID 12500 ('systemctl') (unit session-6.scope)... Jul 10 05:15:33 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: Reloading... Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: Reloading finished in 361 ms. Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=467 op=LOAD Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=448 op=UNLOAD Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=468 op=LOAD Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=449 op=UNLOAD Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=469 op=LOAD Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=470 op=LOAD Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=450 op=UNLOAD Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=451 op=UNLOAD Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=471 op=LOAD Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=452 op=UNLOAD Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=472 op=LOAD Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=453 op=UNLOAD Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=473 op=LOAD Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=474 op=LOAD Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=454 op=UNLOAD Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=455 op=UNLOAD Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=475 op=LOAD Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=456 op=UNLOAD Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=476 op=LOAD Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=477 op=LOAD Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=457 op=UNLOAD Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=458 op=UNLOAD Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=478 op=LOAD Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=459 op=UNLOAD Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=479 op=LOAD Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=460 op=UNLOAD Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=480 op=LOAD Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=461 op=UNLOAD Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=481 op=LOAD Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=482 op=LOAD Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=462 op=UNLOAD Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=463 op=UNLOAD Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=483 op=LOAD Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=464 op=UNLOAD Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=484 op=LOAD Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=485 op=LOAD Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=465 op=UNLOAD Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=466 op=UNLOAD Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:init_t:s0 msg='unit=quadlet-demo-network comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: quadlet-demo-network.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ The unit quadlet-demo-network.service has successfully entered the 'dead' state. Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: Stopped quadlet-demo-network.service. ░░ Subject: A stop job for unit quadlet-demo-network.service has finished ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ A stop job for unit quadlet-demo-network.service has finished. ░░ ░░ The job identifier is 3700 and the job result is done. Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=12551 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success' Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[12567]: ansible-file Invoked with path=/etc/containers/systemd/quadlet-demo.network state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=12568 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success' Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[12584]: ansible-systemd Invoked with daemon_reload=True scope=system daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: Reloading requested from client PID 12585 ('systemctl') (unit session-6.scope)... Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: Reloading... Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: Reloading finished in 363 ms. Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=486 op=LOAD Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=467 op=UNLOAD Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=487 op=LOAD Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=468 op=UNLOAD Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=488 op=LOAD Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=489 op=LOAD Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=469 op=UNLOAD Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=470 op=UNLOAD Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=490 op=LOAD Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=471 op=UNLOAD Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=491 op=LOAD Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=472 op=UNLOAD Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=492 op=LOAD Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=493 op=LOAD Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=473 op=UNLOAD Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=474 op=UNLOAD Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=494 op=LOAD Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=475 op=UNLOAD Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=495 op=LOAD Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=496 op=LOAD Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=476 op=UNLOAD Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=477 op=UNLOAD Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=497 op=LOAD Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=478 op=UNLOAD Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=498 op=LOAD Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=479 op=UNLOAD Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=499 op=LOAD Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=480 op=UNLOAD Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=500 op=LOAD Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=501 op=LOAD Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=481 op=UNLOAD Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=482 op=UNLOAD Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=502 op=LOAD Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=483 op=UNLOAD Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=503 op=LOAD Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=504 op=LOAD Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=484 op=UNLOAD Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=485 op=UNLOAD Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=12636 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success' Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[12652]: ansible-ansible.legacy.command Invoked with _raw_params=podman image prune --all -f _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=12660 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success' Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[12676]: ansible-ansible.legacy.command Invoked with _raw_params=podman network prune -f _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=12684 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success' Jul 10 05:15:36 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[12700]: ansible-ansible.legacy.command Invoked with _raw_params=podman images -n _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 10 05:15:36 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:36 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:36 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:36 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:36 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=12708 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success' Jul 10 05:15:36 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[12724]: ansible-ansible.legacy.command Invoked with _raw_params=podman volume ls -n _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 10 05:15:36 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:36 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:36 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:36 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:36 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=12732 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success' Jul 10 05:15:36 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Jul 10 05:15:36 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[12748]: ansible-ansible.legacy.command Invoked with _raw_params=podman ps --noheading _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 10 05:15:36 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Jul 10 05:15:36 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:36 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:36 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:36 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:36 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=12756 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success' Jul 10 05:15:37 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[12772]: ansible-ansible.legacy.command Invoked with _raw_params=podman network ls -n -q _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 10 05:15:37 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel ░░ ░░ The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Jul 10 05:15:37 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:37 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:37 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:37 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:37 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=12780 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success' Jul 10 05:15:37 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:37 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:37 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:37 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:37 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=12804 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success' Jul 10 05:15:37 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[12820]: ansible-service_facts Invoked Jul 10 05:15:39 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:39 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:39 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:39 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:39 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=12933 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success' Jul 10 05:15:39 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[12949]: ansible-ansible.legacy.command Invoked with _raw_params=exec 1>&2 set -x set -o pipefail systemctl list-units --plain -l --all | grep quadlet || : systemctl list-unit-files --all | grep quadlet || : systemctl list-units --plain --failed -l --all | grep quadlet || : journalctl -u quadlet-demo-mysql.service || : systemctl status quadlet-demo-mysql.service || : _uses_shell=True expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 10 05:15:40 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:40 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:40 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:40 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success' Jul 10 05:15:40 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=12959 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success' Jul 10 05:15:40 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[12975]: ansible-ansible.legacy.command Invoked with _raw_params=journalctl -ex _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None to retry, use: --limit @/tmp/tests_quadlet_demo.retry PLAY RECAP ********************************************************************* sut : ok=374 changed=49 unreachable=0 failed=1 skipped=353 rescued=1 ignored=0 Wednesday 10 July 2024 05:15:40 +0000 (0:00:00.263) 0:03:45.461 ******** =============================================================================== fedora.linux_system_roles.podman : Ensure required packages are installed -- 94.89s /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:14 fedora.linux_system_roles.podman : Ensure container images are present -- 18.02s /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:18 fedora.linux_system_roles.podman : Ensure container images are present --- 6.86s /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:18 fedora.linux_system_roles.certificate : Ensure certificate role dependencies are installed --- 4.83s /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:5 fedora.linux_system_roles.certificate : Ensure provider packages are installed --- 4.47s /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:23 fedora.linux_system_roles.podman : For testing and debugging - services --- 2.30s /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:133 fedora.linux_system_roles.podman : For testing and debugging - services --- 2.30s /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:133 fedora.linux_system_roles.firewall : Install firewalld ------------------ 2.29s /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:19 fedora.linux_system_roles.podman : For testing and debugging - services --- 2.22s /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:133 fedora.linux_system_roles.podman : For testing and debugging - services --- 2.21s /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:133 fedora.linux_system_roles.podman : For testing and debugging - services --- 2.19s /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:133 fedora.linux_system_roles.podman : For testing and debugging - services --- 2.18s /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:133 fedora.linux_system_roles.firewall : Install firewalld ------------------ 2.18s /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:19 fedora.linux_system_roles.podman : Stop and disable service ------------- 1.42s /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:12 fedora.linux_system_roles.podman : Start service ------------------------ 1.34s /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:110 fedora.linux_system_roles.podman : Reload systemctl --------------------- 1.28s /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:82 fedora.linux_system_roles.firewall : Configure firewall ----------------- 1.17s /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:71 Check web --------------------------------------------------------------- 1.11s /WORKDIR/git-handle-quadlet-cleanup35__kigx/tests/tests_quadlet_demo.yml:104 -- fedora.linux_system_roles.firewall : Enable and start firewalld service --- 1.09s /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:28 fedora.linux_system_roles.certificate : Ensure provider service is running --- 1.01s /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:90 ---^---^---^---^---^--- # STDERR: ---v---v---v---v---v--- [DEPRECATION WARNING]: ANSIBLE_COLLECTIONS_PATHS option, does not fit var naming standard, use the singular form ANSIBLE_COLLECTIONS_PATH instead. This feature will be removed from ansible-core in version 2.19. Deprecation warnings can be disabled by setting deprecation_warnings=False in ansible.cfg. [WARNING]: Using a variable for a task's 'args' is unsafe in some situations (see https://docs.ansible.com/ansible/devel/reference_appendices/faq.html#argsplat- unsafe) ---^---^---^---^---^---