# STDOUT:
---v---v---v---v---v---
ansible-playbook [core 2.16.0]
config file = /etc/ansible/ansible.cfg
configured module search path = ['/home/jenkins/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules']
ansible python module location = /opt/ansible-2.16/lib/python3.11/site-packages/ansible
ansible collection location = /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection
executable location = /opt/ansible-2.16/bin/ansible-playbook
python version = 3.11.5 (main, Sep 7 2023, 00:00:00) [GCC 11.4.1 20230605 (Red Hat 11.4.1-2)] (/opt/ansible-2.16/bin/python)
jinja version = 3.1.2
libyaml = True
Using /etc/ansible/ansible.cfg as config file
Skipping callback 'debug', as we already have a stdout callback.
Skipping callback 'default', as we already have a stdout callback.
Skipping callback 'minimal', as we already have a stdout callback.
Skipping callback 'oneline', as we already have a stdout callback.
PLAYBOOK: tests_quadlet_demo.yml ***********************************************
2 plays in /WORKDIR/git-handle-quadlet-cleanup35__kigx/tests/tests_quadlet_demo.yml
PLAY [all] *********************************************************************
TASK [Include vault variables] *************************************************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/tests/tests_quadlet_demo.yml:4
Wednesday 10 July 2024 05:11:55 +0000 (0:00:00.011) 0:00:00.011 ********
ok: [sut] => {
"ansible_facts": {
"__podman_test_password": {
"__ansible_vault": "$ANSIBLE_VAULT;1.1;AES256\n35383939616163653333633431363463313831383037386236646138333162396161356130303461\n3932623930643263313563336163316337643562333936360a363538636631313039343233383732\n38666530383538656639363465313230343533386130303833336434303438333161656262346562\n3362626538613031640a663330613638366132356534363534353239616666653466353961323533\n6565\n"
},
"mysql_container_root_password": {
"__ansible_vault": "$ANSIBLE_VAULT;1.1;AES256\n61333932373230333539663035366431326163363166363036323963623131363530326231303634\n6635326161643165363366323062333334363730376631660a393566366139353861656364656661\n38653463363837336639363032646433666361646535366137303464623261313663643336306465\n6264663730656337310a343962353137386238383064646533366433333437303566656433386233\n34343235326665646661623131643335313236313131353661386338343366316261643634653633\n3832313034366536616531323963333234326461353130303532\n"
}
},
"ansible_included_var_files": [
"/WORKDIR/git-handle-quadlet-cleanup35__kigx/tests/vars/vault-variables.yml"
],
"changed": false
}
PLAY [Deploy the quadlet demo app] *********************************************
TASK [Gathering Facts] *********************************************************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/tests/tests_quadlet_demo.yml:9
Wednesday 10 July 2024 05:11:55 +0000 (0:00:00.012) 0:00:00.024 ********
ok: [sut]
TASK [Generate certificates] ***************************************************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/tests/tests_quadlet_demo.yml:39
Wednesday 10 July 2024 05:11:56 +0000 (0:00:00.974) 0:00:00.998 ********
TASK [fedora.linux_system_roles.certificate : Set version specific variables] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:2
Wednesday 10 July 2024 05:11:56 +0000 (0:00:00.023) 0:00:01.022 ********
included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/set_vars.yml for sut
TASK [fedora.linux_system_roles.certificate : Ensure ansible_facts used by role] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/set_vars.yml:2
Wednesday 10 July 2024 05:11:56 +0000 (0:00:00.012) 0:00:01.034 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__certificate_required_facts | difference(ansible_facts.keys() | list) | length > 0",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.certificate : Check if system is ostree] *******
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/set_vars.yml:10
Wednesday 10 July 2024 05:11:56 +0000 (0:00:00.017) 0:00:01.052 ********
ok: [sut] => {
"changed": false,
"stat": {
"exists": false
}
}
TASK [fedora.linux_system_roles.certificate : Set flag to indicate system is ostree] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/set_vars.yml:15
Wednesday 10 July 2024 05:11:56 +0000 (0:00:00.275) 0:00:01.327 ********
ok: [sut] => {
"ansible_facts": {
"__certificate_is_ostree": false
},
"changed": false
}
TASK [fedora.linux_system_roles.certificate : Set platform/version specific variables] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/set_vars.yml:19
Wednesday 10 July 2024 05:11:56 +0000 (0:00:00.014) 0:00:01.342 ********
skipping: [sut] => (item=RedHat.yml) => {
"ansible_loop_var": "item",
"changed": false,
"false_condition": "__vars_file is file",
"item": "RedHat.yml",
"skip_reason": "Conditional result was False"
}
ok: [sut] => (item=Fedora.yml) => {
"ansible_facts": {
"__certificate_certmonger_packages": [
"certmonger",
"python3-packaging"
]
},
"ansible_included_var_files": [
"/WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/vars/Fedora.yml"
],
"ansible_loop_var": "item",
"changed": false,
"item": "Fedora.yml"
}
skipping: [sut] => (item=Fedora_39.yml) => {
"ansible_loop_var": "item",
"changed": false,
"false_condition": "__vars_file is file",
"item": "Fedora_39.yml",
"skip_reason": "Conditional result was False"
}
skipping: [sut] => (item=Fedora_39.yml) => {
"ansible_loop_var": "item",
"changed": false,
"false_condition": "__vars_file is file",
"item": "Fedora_39.yml",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.certificate : Ensure certificate role dependencies are installed] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:5
Wednesday 10 July 2024 05:11:56 +0000 (0:00:00.023) 0:00:01.366 ********
changed: [sut] => {
"changed": true,
"rc": 0,
"results": [
"Installed: python3-ply-3.11-20.fc39.noarch",
"Installed: python3-pycparser-2.20-11.fc39.noarch",
"Installed: python3-cryptography-41.0.7-1.fc39.x86_64",
"Installed: python3-cffi-1.15.1-6.fc39.x86_64",
"Installed: python3-pyasn1-0.5.1-1.fc39.noarch"
]
}
TASK [fedora.linux_system_roles.certificate : Ensure provider packages are installed] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:23
Wednesday 10 July 2024 05:12:01 +0000 (0:00:04.834) 0:00:06.200 ********
changed: [sut] => (item=certmonger) => {
"__certificate_provider": "certmonger",
"ansible_loop_var": "__certificate_provider",
"changed": true,
"rc": 0,
"results": [
"Installed: certmonger-0.79.18-2.fc39.x86_64",
"Installed: dbus-tools-1:1.14.10-1.fc39.x86_64",
"Installed: python3-packaging-23.1-4.fc39.noarch"
]
}
TASK [fedora.linux_system_roles.certificate : Ensure pre-scripts hooks directory exists] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:35
Wednesday 10 July 2024 05:12:06 +0000 (0:00:04.469) 0:00:10.670 ********
changed: [sut] => (item=certmonger) => {
"__certificate_provider": "certmonger",
"ansible_loop_var": "__certificate_provider",
"changed": true,
"gid": 0,
"group": "root",
"mode": "0700",
"owner": "root",
"path": "/etc/certmonger//pre-scripts",
"secontext": "unconfined_u:object_r:etc_t:s0",
"size": 4096,
"state": "directory",
"uid": 0
}
TASK [fedora.linux_system_roles.certificate : Ensure post-scripts hooks directory exists] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:61
Wednesday 10 July 2024 05:12:06 +0000 (0:00:00.399) 0:00:11.069 ********
changed: [sut] => (item=certmonger) => {
"__certificate_provider": "certmonger",
"ansible_loop_var": "__certificate_provider",
"changed": true,
"gid": 0,
"group": "root",
"mode": "0700",
"owner": "root",
"path": "/etc/certmonger//post-scripts",
"secontext": "unconfined_u:object_r:etc_t:s0",
"size": 4096,
"state": "directory",
"uid": 0
}
TASK [fedora.linux_system_roles.certificate : Ensure provider service is running] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:90
Wednesday 10 July 2024 05:12:06 +0000 (0:00:00.248) 0:00:11.318 ********
changed: [sut] => (item=certmonger) => {
"__certificate_provider": "certmonger",
"ansible_loop_var": "__certificate_provider",
"changed": true,
"enabled": true,
"name": "certmonger",
"state": "started",
"status": {
"AccessSELinuxContext": "system_u:object_r:certmonger_unit_file_t:s0",
"ActiveEnterTimestampMonotonic": "0",
"ActiveExitTimestampMonotonic": "0",
"ActiveState": "inactive",
"After": "syslog.target systemd-journald.socket dbus-broker.service sysinit.target system.slice basic.target network.target dbus.socket",
"AllowIsolate": "no",
"AssertResult": "no",
"AssertTimestampMonotonic": "0",
"Before": "shutdown.target",
"BlockIOAccounting": "no",
"BlockIOWeight": "[not set]",
"BusName": "org.fedorahosted.certmonger",
"CPUAccounting": "yes",
"CPUAffinityFromNUMA": "no",
"CPUQuotaPerSecUSec": "infinity",
"CPUQuotaPeriodUSec": "infinity",
"CPUSchedulingPolicy": "0",
"CPUSchedulingPriority": "0",
"CPUSchedulingResetOnFork": "no",
"CPUShares": "[not set]",
"CPUUsageNSec": "[not set]",
"CPUWeight": "[not set]",
"CacheDirectoryMode": "0755",
"CanFreeze": "yes",
"CanIsolate": "no",
"CanReload": "no",
"CanStart": "yes",
"CanStop": "yes",
"CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore",
"CleanResult": "success",
"CollectMode": "inactive",
"ConditionResult": "no",
"ConditionTimestampMonotonic": "0",
"ConfigurationDirectoryMode": "0755",
"Conflicts": "shutdown.target",
"ControlGroupId": "0",
"ControlPID": "0",
"CoredumpFilter": "0x33",
"DefaultDependencies": "yes",
"DefaultMemoryLow": "0",
"DefaultMemoryMin": "0",
"DefaultStartupMemoryLow": "0",
"Delegate": "no",
"Description": "Certificate monitoring and PKI enrollment",
"DevicePolicy": "auto",
"DropInPaths": "/usr/lib/systemd/system/service.d/10-timeout-abort.conf",
"DynamicUser": "no",
"EnvironmentFiles": "/etc/sysconfig/certmonger (ignore_errors=yes)",
"ExecMainCode": "0",
"ExecMainExitTimestampMonotonic": "0",
"ExecMainPID": "0",
"ExecMainStartTimestampMonotonic": "0",
"ExecMainStatus": "0",
"ExecStart": "{ path=/usr/sbin/certmonger ; argv[]=/usr/sbin/certmonger -S -p /run/certmonger.pid -n $OPTS ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"ExecStartEx": "{ path=/usr/sbin/certmonger ; argv[]=/usr/sbin/certmonger -S -p /run/certmonger.pid -n $OPTS ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"ExitType": "main",
"ExtensionImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent",
"FailureAction": "none",
"FileDescriptorStoreMax": "0",
"FileDescriptorStorePreserve": "restart",
"FinalKillSignal": "9",
"FragmentPath": "/usr/lib/systemd/system/certmonger.service",
"FreezerState": "running",
"GID": "[not set]",
"GuessMainPID": "yes",
"IOAccounting": "no",
"IOReadBytes": "18446744073709551615",
"IOReadOperations": "18446744073709551615",
"IOSchedulingClass": "2",
"IOSchedulingPriority": "4",
"IOWeight": "[not set]",
"IOWriteBytes": "18446744073709551615",
"IOWriteOperations": "18446744073709551615",
"IPAccounting": "no",
"IPEgressBytes": "[no data]",
"IPEgressPackets": "[no data]",
"IPIngressBytes": "[no data]",
"IPIngressPackets": "[no data]",
"Id": "certmonger.service",
"IgnoreOnIsolate": "no",
"IgnoreSIGPIPE": "yes",
"InactiveEnterTimestampMonotonic": "0",
"InactiveExitTimestampMonotonic": "0",
"JobRunningTimeoutUSec": "infinity",
"JobTimeoutAction": "none",
"JobTimeoutUSec": "infinity",
"KeyringMode": "private",
"KillMode": "control-group",
"KillSignal": "15",
"LimitAS": "infinity",
"LimitASSoft": "infinity",
"LimitCORE": "infinity",
"LimitCORESoft": "infinity",
"LimitCPU": "infinity",
"LimitCPUSoft": "infinity",
"LimitDATA": "infinity",
"LimitDATASoft": "infinity",
"LimitFSIZE": "infinity",
"LimitFSIZESoft": "infinity",
"LimitLOCKS": "infinity",
"LimitLOCKSSoft": "infinity",
"LimitMEMLOCK": "8388608",
"LimitMEMLOCKSoft": "8388608",
"LimitMSGQUEUE": "819200",
"LimitMSGQUEUESoft": "819200",
"LimitNICE": "0",
"LimitNICESoft": "0",
"LimitNOFILE": "524288",
"LimitNOFILESoft": "1024",
"LimitNPROC": "14756",
"LimitNPROCSoft": "14756",
"LimitRSS": "infinity",
"LimitRSSSoft": "infinity",
"LimitRTPRIO": "0",
"LimitRTPRIOSoft": "0",
"LimitRTTIME": "infinity",
"LimitRTTIMESoft": "infinity",
"LimitSIGPENDING": "14756",
"LimitSIGPENDINGSoft": "14756",
"LimitSTACK": "infinity",
"LimitSTACKSoft": "8388608",
"LoadState": "loaded",
"LockPersonality": "no",
"LogLevelMax": "-1",
"LogRateLimitBurst": "0",
"LogRateLimitIntervalUSec": "0",
"LogsDirectoryMode": "0755",
"MainPID": "0",
"ManagedOOMMemoryPressure": "auto",
"ManagedOOMMemoryPressureLimit": "0",
"ManagedOOMPreference": "none",
"ManagedOOMSwap": "auto",
"MemoryAccounting": "yes",
"MemoryAvailable": "infinity",
"MemoryCurrent": "[not set]",
"MemoryDenyWriteExecute": "no",
"MemoryHigh": "infinity",
"MemoryKSM": "no",
"MemoryLimit": "infinity",
"MemoryLow": "0",
"MemoryMax": "infinity",
"MemoryMin": "0",
"MemoryPressureThresholdUSec": "200ms",
"MemoryPressureWatch": "auto",
"MemorySwapMax": "infinity",
"MemoryZSwapMax": "infinity",
"MountAPIVFS": "no",
"MountImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent",
"NFileDescriptorStore": "0",
"NRestarts": "0",
"NUMAPolicy": "n/a",
"Names": "certmonger.service",
"NeedDaemonReload": "no",
"Nice": "0",
"NoNewPrivileges": "no",
"NonBlocking": "no",
"NotifyAccess": "none",
"OOMPolicy": "stop",
"OOMScoreAdjust": "0",
"OnFailureJobMode": "replace",
"OnSuccessJobMode": "fail",
"PIDFile": "/run/certmonger.pid",
"PartOf": "dbus-broker.service",
"Perpetual": "no",
"PrivateDevices": "no",
"PrivateIPC": "no",
"PrivateMounts": "no",
"PrivateNetwork": "no",
"PrivateTmp": "no",
"PrivateUsers": "no",
"ProcSubset": "all",
"ProtectClock": "no",
"ProtectControlGroups": "no",
"ProtectHome": "no",
"ProtectHostname": "no",
"ProtectKernelLogs": "no",
"ProtectKernelModules": "no",
"ProtectKernelTunables": "no",
"ProtectProc": "default",
"ProtectSystem": "no",
"RefuseManualStart": "no",
"RefuseManualStop": "no",
"ReloadResult": "success",
"ReloadSignal": "1",
"RemainAfterExit": "no",
"RemoveIPC": "no",
"Requires": "dbus.socket sysinit.target system.slice",
"Restart": "no",
"RestartKillSignal": "15",
"RestartMaxDelayUSec": "infinity",
"RestartMode": "normal",
"RestartSteps": "0",
"RestartUSec": "100ms",
"RestartUSecNext": "100ms",
"RestrictNamespaces": "no",
"RestrictRealtime": "no",
"RestrictSUIDSGID": "no",
"Result": "success",
"RootDirectoryStartOnly": "no",
"RootEphemeral": "no",
"RootImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent",
"RuntimeDirectoryMode": "0755",
"RuntimeDirectoryPreserve": "no",
"RuntimeMaxUSec": "infinity",
"RuntimeRandomizedExtraUSec": "0",
"SameProcessGroup": "no",
"SecureBits": "0",
"SendSIGHUP": "no",
"SendSIGKILL": "yes",
"Slice": "system.slice",
"StandardError": "inherit",
"StandardInput": "null",
"StandardOutput": "journal",
"StartLimitAction": "none",
"StartLimitBurst": "5",
"StartLimitIntervalUSec": "10s",
"StartupBlockIOWeight": "[not set]",
"StartupCPUShares": "[not set]",
"StartupCPUWeight": "[not set]",
"StartupIOWeight": "[not set]",
"StartupMemoryHigh": "18446744073709551615",
"StartupMemoryLow": "0",
"StartupMemoryMax": "18446744073709551615",
"StartupMemorySwapMax": "18446744073709551615",
"StartupMemoryZSwapMax": "18446744073709551615",
"StateChangeTimestampMonotonic": "0",
"StateDirectoryMode": "0755",
"StatusErrno": "0",
"StopWhenUnneeded": "no",
"SubState": "dead",
"SuccessAction": "none",
"SyslogFacility": "3",
"SyslogLevel": "6",
"SyslogLevelPrefix": "yes",
"SyslogPriority": "30",
"SystemCallErrorNumber": "2147483646",
"TTYReset": "no",
"TTYVHangup": "no",
"TTYVTDisallocate": "no",
"TasksAccounting": "yes",
"TasksCurrent": "[not set]",
"TasksMax": "4426",
"TimeoutAbortUSec": "45s",
"TimeoutCleanUSec": "infinity",
"TimeoutStartFailureMode": "terminate",
"TimeoutStartUSec": "45s",
"TimeoutStopFailureMode": "abort",
"TimeoutStopUSec": "45s",
"TimerSlackNSec": "50000",
"Transient": "no",
"Type": "dbus",
"UID": "[not set]",
"UMask": "0022",
"UnitFilePreset": "disabled",
"UnitFileState": "disabled",
"UtmpMode": "init",
"WatchdogSignal": "6",
"WatchdogTimestampMonotonic": "0",
"WatchdogUSec": "infinity"
}
}
TASK [fedora.linux_system_roles.certificate : Ensure certificate requests] *****
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:101
Wednesday 10 July 2024 05:12:07 +0000 (0:00:01.009) 0:00:12.328 ********
changed: [sut] => (item={'name': 'quadlet_demo', 'dns': ['localhost'], 'ca': 'self-sign'}) => {
"ansible_loop_var": "item",
"changed": true,
"item": {
"ca": "self-sign",
"dns": [
"localhost"
],
"name": "quadlet_demo"
}
}
MSG:
Certificate requested (new).
TASK [fedora.linux_system_roles.certificate : Slurp the contents of the files] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:152
Wednesday 10 July 2024 05:12:08 +0000 (0:00:00.733) 0:00:13.061 ********
ok: [sut] => (item=['cert', {'name': 'quadlet_demo', 'dns': ['localhost'], 'ca': 'self-sign'}]) => {
"ansible_loop_var": "item",
"changed": false,
"content": "LS0tLS1CRUdJTiBDRVJUSUZJQ0FURS0tLS0tCk1JSURnakNDQW1xZ0F3SUJBZ0lRZDR1WUh2UExTSlc4ekdlQWtEWEV0akFOQmdrcWhraUc5dzBCQVFzRkFEQlEKTVNBd0hnWURWUVFEREJkTWIyTmhiQ0JUYVdkdWFXNW5JRUYxZEdodmNtbDBlVEVzTUNvR0ExVUVBd3dqTnpjNApZams0TVdVdFpqTmpZalE0T1RVdFltTmpZelkzT0RBdE9UQXpOV00wWWpVd0hoY05NalF3TnpFd01EVXhNakE0CldoY05NalV3TnpFd01EVXhNakEzV2pBVU1SSXdFQVlEVlFRREV3bHNiMk5oYkdodmMzUXdnZ0VpTUEwR0NTcUcKU0liM0RRRUJBUVVBQTRJQkR3QXdnZ0VLQW9JQkFRQ1lHWktkbWltVkVBNVJwRmhhVjRBTlBUWURkb0Rqd2gxNQprZGxOWmZHQW50dVJBT24wNTA4NSsrNTBRZmhvbU1yeVJsOVRwd0poZEoxT3Q2RkVPYU9pZUZadGFhSGluRmhZCk1WY0lQUFYyUmdGMVd6SXlQQzYzUTFqQ2xIVHQvMWdvRk1lTG52a3VKdkpGZUZvVk9scG0zckJUYllGT252enEKYkVyT0NDNVNxK0p5V3dYYkJtY25xUjRzczBYeGRUZzV5S0FxU3dtMXNTK0dtYVByVDJoYnpqaVNoQXpzaEZEWgpsOUJGNHc0NW1KL2pWc3JDc2k3OXBYL0VyOXgwMVZTeGdvaFJuTzMvaDBGUHUxTzVsbFBENm1HU1oxbUpIMzUvCkZ4WGlZcDRKZmw1V1Y0QkROOFBQWnJiay9uNGErYmNrRkh3cm1Ib3pLdnVUMkVMUlpxVGhBZ01CQUFHamdaTXcKZ1pBd0N3WURWUjBQQkFRREFnV2dNQlFHQTFVZEVRUU5NQXVDQ1d4dlkyRnNhRzl6ZERBZEJnTlZIU1VFRmpBVQpCZ2dyQmdFRkJRY0RBUVlJS3dZQkJRVUhBd0l3REFZRFZSMFRBUUgvQkFJd0FEQWRCZ05WSFE0RUZnUVV0OE5lCkdTSHVxOGcrakZTL3h2YS9FK3BPNXJzd0h3WURWUjBqQkJnd0ZvQVU2bGdLVGpnSGJSTEhxK3E0WWw4SGVtVmEKVzZnd0RRWUpLb1pJaHZjTkFRRUxCUUFEZ2dFQkFFYkhOb1dFVWZkVVZwTDBlMVhxMGRSclVtaUoraXJHRG5BYwpjWUxaUXdCQjRLSW9INkQzQ051ZUZydXVSNmxEN3d0bVNPYmZDU3ZLNGFmYXdST01nWE1yVVErQnBkZDdpSVRJClRCeUp6bDVpQ0M0NWY1c21kNS9nR01QN1Rqd2RGeVZhT25RV0ZyNHV5bzhkVHVxamNGcVZaWksxekVZZlQ3aGMKMFh6OGJJRkhzRWJIZ01SaEhxYzlQQ29MRHcxem0reDlVY2JLd3pKUWRNczZESlo5c2NSRWg5cmgxY2NCMW4xbQpVZHVrK2FUdWhRdTQ3Z2NHVUtiMzhqaWxYVllWbys4SE1BZDA1bnZTeHlOa2RrWUIzNVBEM2llQzBaSng5VzJSCk4xQkk5cXJ2emVLKzhpSElxZEU0MktlYksya1EyWmdOaktjSTQ0aEJLSnMvaW5EZGRFND0KLS0tLS1FTkQgQ0VSVElGSUNBVEUtLS0tLQo=",
"encoding": "base64",
"item": [
"cert",
{
"ca": "self-sign",
"dns": [
"localhost"
],
"name": "quadlet_demo"
}
],
"source": "/etc/pki/tls/certs/quadlet_demo.crt"
}
ok: [sut] => (item=['key', {'name': 'quadlet_demo', 'dns': ['localhost'], 'ca': 'self-sign'}]) => {
"ansible_loop_var": "item",
"changed": false,
"content": "LS0tLS1CRUdJTiBQUklWQVRFIEtFWS0tLS0tCk1JSUV2UUlCQURBTkJna3Foa2lHOXcwQkFRRUZBQVNDQktjd2dnU2pBZ0VBQW9JQkFRQ1lHWktkbWltVkVBNVIKcEZoYVY0QU5QVFlEZG9EandoMTVrZGxOWmZHQW50dVJBT24wNTA4NSsrNTBRZmhvbU1yeVJsOVRwd0poZEoxTwp0NkZFT2FPaWVGWnRhYUhpbkZoWU1WY0lQUFYyUmdGMVd6SXlQQzYzUTFqQ2xIVHQvMWdvRk1lTG52a3VKdkpGCmVGb1ZPbHBtM3JCVGJZRk9udnpxYkVyT0NDNVNxK0p5V3dYYkJtY25xUjRzczBYeGRUZzV5S0FxU3dtMXNTK0cKbWFQclQyaGJ6amlTaEF6c2hGRFpsOUJGNHc0NW1KL2pWc3JDc2k3OXBYL0VyOXgwMVZTeGdvaFJuTzMvaDBGUAp1MU81bGxQRDZtR1NaMW1KSDM1L0Z4WGlZcDRKZmw1V1Y0QkROOFBQWnJiay9uNGErYmNrRkh3cm1Ib3pLdnVUCjJFTFJacVRoQWdNQkFBRUNnZ0VBQUxTNkxpeDJjZEpkTWFadzBVc0hqWEF0dTNDQjdzU3ljck54U2ZOZG5vaG0KT2g4aUp0eUp2T241SUcyanRiY1RWcEpxYS9yS0Ixb2xab1dTVEZ6MDRDcTlNblR2Znh2MUNrUkEzRms5bGF6agpvZmdBVkxTOExnWTBUOU5rODB1WWpQZVRPVWY5cmJmWkx5YTh5czFPZVB5aXZvRmpMTEp1MTdiMDFpM2FLckN2CkZ0RHE1NmZNMGIxb2t1bDVBS1RhRUluU09MV3FyK0NBYnFaS2dKSDRHQ3ZuWURuRmVzb1daSC9iSWpQbmtOU2oKUE5EUGVRMGJKbFBJUVRmZ2JZM242U3hZckY3TWl5dWdVczROWXYyKzV4NTdhdkNnL3BFckt0c2xhczN2QUV5QgprOUlUbTdXQVBna09FaTBHMVB3TnFzbFpXUW02NUdIRXhZQVNlRFBRZ1FLQmdRRFFYbGpkN3poREF5QXdPVG5XCkNPUWVUcTVFSDBqUW82Ni80YytjdXlieWhzditlR1A0OUJnSkd0Ty84SllieHM2NFZkdTY0Y0E1NHRzMUwwMGcKK2xkLzBxVG5sbEJKWHRTOGFjOEp3N0VjMFROV2RHeWQraTJpSHlxK3BMYXJRbFhiUDFKSlE3MWdudUJtVUlkNwp4WktXTldJcitBbFpudFo1aFNzYmFSbitJUUtCZ1FDNjNtZDhmZHU3cGE4NGVqRzJGVEF5SjE5Y2ZZNzF3bCtYCmkxcU1Sb2xyZFBiVFgwWS8vMDlBQzQvVGVxcWNUYkZqZG9IUW5Qa20rTjVRcjBKUU8veGNuVzBrYWxIc1Bhc0EKcTFPeUJJTCsvY0NqV2gwVklLbzczN3o3WkV5LzdlcnEyeU5jaGhlektZcXlPRldNcU92clFocnMxTjRqS3ZHbQpmaHVWdldCT3dRS0JnUUM5QVlmTXNLeXhKSW16RklXZzMybkZFQXZGSVRlVWdaVUcydGJPelhpenI1QWs0S0dXCmRodkRUb0ZhN3VuS2dsaUxIdDUwczBnQmVnVHJSbC9KRlN2L2Qxdzc4YjRiLzVncTNET0FJUEZzT1ZzL0JkWnMKcWoxd2ZLTHlVVWlkS2t2azE1aFZlc3V5ZEFzSzQrc3hXM0x2SzdkRU9QYUtRNms2aTVBaGlnR1k0UUtCZ0JiRApKcG5zZnNxUnVSRkpHU3NUMDZZWVU2NWhnUHVBck1JNThGS3IrT3RFK1Jmd1FBOWxteEt6RmRidDEwQm8ybFhDClJyZlJseXZmSHIxM1M3MllRUHBMRnROYnZCVHB3aDIzd044clc0V1NCVmdnanZCVzVVV2IwRnBTOG1KaXFiM24KNlp2SW5iUnVOLzNUak9Bd1oyUmx6Zi9SUWluSWZsMEllWFc0OXJWQkFvR0FYWFQ3UkxualZDdlVCbzhtaWpUMwplUm5wNHVtSm9CUTk0Tk11NzBxV0pXbmIxWk1yUmlMNHZEd3JMZk11YUhYM2xjUUk3YklmSnUydkZqZGZOdjJOCjVUYlVMcy91aE51ZVhwaFlGMGlXb05VNmwveTVOaTYyN3pQUTdxMWphekU0RjhOaktxazZSOWhZY1U4NW1JZ3YKa2QwYkVYckRtM1lHem5SWmxWQW04RHM9Ci0tLS0tRU5EIFBSSVZBVEUgS0VZLS0tLS0K",
"encoding": "base64",
"item": [
"key",
{
"ca": "self-sign",
"dns": [
"localhost"
],
"name": "quadlet_demo"
}
],
"source": "/etc/pki/tls/private/quadlet_demo.key"
}
ok: [sut] => (item=['ca', {'name': 'quadlet_demo', 'dns': ['localhost'], 'ca': 'self-sign'}]) => {
"ansible_loop_var": "item",
"changed": false,
"content": "LS0tLS1CRUdJTiBDRVJUSUZJQ0FURS0tLS0tCk1JSURnakNDQW1xZ0F3SUJBZ0lRZDR1WUh2UExTSlc4ekdlQWtEWEV0akFOQmdrcWhraUc5dzBCQVFzRkFEQlEKTVNBd0hnWURWUVFEREJkTWIyTmhiQ0JUYVdkdWFXNW5JRUYxZEdodmNtbDBlVEVzTUNvR0ExVUVBd3dqTnpjNApZams0TVdVdFpqTmpZalE0T1RVdFltTmpZelkzT0RBdE9UQXpOV00wWWpVd0hoY05NalF3TnpFd01EVXhNakE0CldoY05NalV3TnpFd01EVXhNakEzV2pBVU1SSXdFQVlEVlFRREV3bHNiMk5oYkdodmMzUXdnZ0VpTUEwR0NTcUcKU0liM0RRRUJBUVVBQTRJQkR3QXdnZ0VLQW9JQkFRQ1lHWktkbWltVkVBNVJwRmhhVjRBTlBUWURkb0Rqd2gxNQprZGxOWmZHQW50dVJBT24wNTA4NSsrNTBRZmhvbU1yeVJsOVRwd0poZEoxT3Q2RkVPYU9pZUZadGFhSGluRmhZCk1WY0lQUFYyUmdGMVd6SXlQQzYzUTFqQ2xIVHQvMWdvRk1lTG52a3VKdkpGZUZvVk9scG0zckJUYllGT252enEKYkVyT0NDNVNxK0p5V3dYYkJtY25xUjRzczBYeGRUZzV5S0FxU3dtMXNTK0dtYVByVDJoYnpqaVNoQXpzaEZEWgpsOUJGNHc0NW1KL2pWc3JDc2k3OXBYL0VyOXgwMVZTeGdvaFJuTzMvaDBGUHUxTzVsbFBENm1HU1oxbUpIMzUvCkZ4WGlZcDRKZmw1V1Y0QkROOFBQWnJiay9uNGErYmNrRkh3cm1Ib3pLdnVUMkVMUlpxVGhBZ01CQUFHamdaTXcKZ1pBd0N3WURWUjBQQkFRREFnV2dNQlFHQTFVZEVRUU5NQXVDQ1d4dlkyRnNhRzl6ZERBZEJnTlZIU1VFRmpBVQpCZ2dyQmdFRkJRY0RBUVlJS3dZQkJRVUhBd0l3REFZRFZSMFRBUUgvQkFJd0FEQWRCZ05WSFE0RUZnUVV0OE5lCkdTSHVxOGcrakZTL3h2YS9FK3BPNXJzd0h3WURWUjBqQkJnd0ZvQVU2bGdLVGpnSGJSTEhxK3E0WWw4SGVtVmEKVzZnd0RRWUpLb1pJaHZjTkFRRUxCUUFEZ2dFQkFFYkhOb1dFVWZkVVZwTDBlMVhxMGRSclVtaUoraXJHRG5BYwpjWUxaUXdCQjRLSW9INkQzQ051ZUZydXVSNmxEN3d0bVNPYmZDU3ZLNGFmYXdST01nWE1yVVErQnBkZDdpSVRJClRCeUp6bDVpQ0M0NWY1c21kNS9nR01QN1Rqd2RGeVZhT25RV0ZyNHV5bzhkVHVxamNGcVZaWksxekVZZlQ3aGMKMFh6OGJJRkhzRWJIZ01SaEhxYzlQQ29MRHcxem0reDlVY2JLd3pKUWRNczZESlo5c2NSRWg5cmgxY2NCMW4xbQpVZHVrK2FUdWhRdTQ3Z2NHVUtiMzhqaWxYVllWbys4SE1BZDA1bnZTeHlOa2RrWUIzNVBEM2llQzBaSng5VzJSCk4xQkk5cXJ2emVLKzhpSElxZEU0MktlYksya1EyWmdOaktjSTQ0aEJLSnMvaW5EZGRFND0KLS0tLS1FTkQgQ0VSVElGSUNBVEUtLS0tLQo=",
"encoding": "base64",
"item": [
"ca",
{
"ca": "self-sign",
"dns": [
"localhost"
],
"name": "quadlet_demo"
}
],
"source": "/etc/pki/tls/certs/quadlet_demo.crt"
}
TASK [fedora.linux_system_roles.certificate : Create return data] **************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:160
Wednesday 10 July 2024 05:12:09 +0000 (0:00:00.696) 0:00:13.757 ********
ok: [sut] => {
"ansible_facts": {
"certificate_test_certs": {
"quadlet_demo": {
"ca": "/etc/pki/tls/certs/quadlet_demo.crt",
"ca_content": "-----BEGIN CERTIFICATE-----\nMIIDgjCCAmqgAwIBAgIQd4uYHvPLSJW8zGeAkDXEtjANBgkqhkiG9w0BAQsFADBQ\nMSAwHgYDVQQDDBdMb2NhbCBTaWduaW5nIEF1dGhvcml0eTEsMCoGA1UEAwwjNzc4\nYjk4MWUtZjNjYjQ4OTUtYmNjYzY3ODAtOTAzNWM0YjUwHhcNMjQwNzEwMDUxMjA4\nWhcNMjUwNzEwMDUxMjA3WjAUMRIwEAYDVQQDEwlsb2NhbGhvc3QwggEiMA0GCSqG\nSIb3DQEBAQUAA4IBDwAwggEKAoIBAQCYGZKdmimVEA5RpFhaV4ANPTYDdoDjwh15\nkdlNZfGAntuRAOn05085++50QfhomMryRl9TpwJhdJ1Ot6FEOaOieFZtaaHinFhY\nMVcIPPV2RgF1WzIyPC63Q1jClHTt/1goFMeLnvkuJvJFeFoVOlpm3rBTbYFOnvzq\nbErOCC5Sq+JyWwXbBmcnqR4ss0XxdTg5yKAqSwm1sS+GmaPrT2hbzjiShAzshFDZ\nl9BF4w45mJ/jVsrCsi79pX/Er9x01VSxgohRnO3/h0FPu1O5llPD6mGSZ1mJH35/\nFxXiYp4Jfl5WV4BDN8PPZrbk/n4a+bckFHwrmHozKvuT2ELRZqThAgMBAAGjgZMw\ngZAwCwYDVR0PBAQDAgWgMBQGA1UdEQQNMAuCCWxvY2FsaG9zdDAdBgNVHSUEFjAU\nBggrBgEFBQcDAQYIKwYBBQUHAwIwDAYDVR0TAQH/BAIwADAdBgNVHQ4EFgQUt8Ne\nGSHuq8g+jFS/xva/E+pO5rswHwYDVR0jBBgwFoAU6lgKTjgHbRLHq+q4Yl8HemVa\nW6gwDQYJKoZIhvcNAQELBQADggEBAEbHNoWEUfdUVpL0e1Xq0dRrUmiJ+irGDnAc\ncYLZQwBB4KIoH6D3CNueFruuR6lD7wtmSObfCSvK4afawROMgXMrUQ+Bpdd7iITI\nTByJzl5iCC45f5smd5/gGMP7TjwdFyVaOnQWFr4uyo8dTuqjcFqVZZK1zEYfT7hc\n0Xz8bIFHsEbHgMRhHqc9PCoLDw1zm+x9UcbKwzJQdMs6DJZ9scREh9rh1ccB1n1m\nUduk+aTuhQu47gcGUKb38jilXVYVo+8HMAd05nvSxyNkdkYB35PD3ieC0ZJx9W2R\nN1BI9qrvzeK+8iHIqdE42KebK2kQ2ZgNjKcI44hBKJs/inDddE4=\n-----END CERTIFICATE-----\n",
"cert": "/etc/pki/tls/certs/quadlet_demo.crt",
"cert_content": "-----BEGIN CERTIFICATE-----\nMIIDgjCCAmqgAwIBAgIQd4uYHvPLSJW8zGeAkDXEtjANBgkqhkiG9w0BAQsFADBQ\nMSAwHgYDVQQDDBdMb2NhbCBTaWduaW5nIEF1dGhvcml0eTEsMCoGA1UEAwwjNzc4\nYjk4MWUtZjNjYjQ4OTUtYmNjYzY3ODAtOTAzNWM0YjUwHhcNMjQwNzEwMDUxMjA4\nWhcNMjUwNzEwMDUxMjA3WjAUMRIwEAYDVQQDEwlsb2NhbGhvc3QwggEiMA0GCSqG\nSIb3DQEBAQUAA4IBDwAwggEKAoIBAQCYGZKdmimVEA5RpFhaV4ANPTYDdoDjwh15\nkdlNZfGAntuRAOn05085++50QfhomMryRl9TpwJhdJ1Ot6FEOaOieFZtaaHinFhY\nMVcIPPV2RgF1WzIyPC63Q1jClHTt/1goFMeLnvkuJvJFeFoVOlpm3rBTbYFOnvzq\nbErOCC5Sq+JyWwXbBmcnqR4ss0XxdTg5yKAqSwm1sS+GmaPrT2hbzjiShAzshFDZ\nl9BF4w45mJ/jVsrCsi79pX/Er9x01VSxgohRnO3/h0FPu1O5llPD6mGSZ1mJH35/\nFxXiYp4Jfl5WV4BDN8PPZrbk/n4a+bckFHwrmHozKvuT2ELRZqThAgMBAAGjgZMw\ngZAwCwYDVR0PBAQDAgWgMBQGA1UdEQQNMAuCCWxvY2FsaG9zdDAdBgNVHSUEFjAU\nBggrBgEFBQcDAQYIKwYBBQUHAwIwDAYDVR0TAQH/BAIwADAdBgNVHQ4EFgQUt8Ne\nGSHuq8g+jFS/xva/E+pO5rswHwYDVR0jBBgwFoAU6lgKTjgHbRLHq+q4Yl8HemVa\nW6gwDQYJKoZIhvcNAQELBQADggEBAEbHNoWEUfdUVpL0e1Xq0dRrUmiJ+irGDnAc\ncYLZQwBB4KIoH6D3CNueFruuR6lD7wtmSObfCSvK4afawROMgXMrUQ+Bpdd7iITI\nTByJzl5iCC45f5smd5/gGMP7TjwdFyVaOnQWFr4uyo8dTuqjcFqVZZK1zEYfT7hc\n0Xz8bIFHsEbHgMRhHqc9PCoLDw1zm+x9UcbKwzJQdMs6DJZ9scREh9rh1ccB1n1m\nUduk+aTuhQu47gcGUKb38jilXVYVo+8HMAd05nvSxyNkdkYB35PD3ieC0ZJx9W2R\nN1BI9qrvzeK+8iHIqdE42KebK2kQ2ZgNjKcI44hBKJs/inDddE4=\n-----END CERTIFICATE-----\n",
"key": "/etc/pki/tls/private/quadlet_demo.key",
"key_content": "-----BEGIN PRIVATE KEY-----\nMIIEvQIBADANBgkqhkiG9w0BAQEFAASCBKcwggSjAgEAAoIBAQCYGZKdmimVEA5R\npFhaV4ANPTYDdoDjwh15kdlNZfGAntuRAOn05085++50QfhomMryRl9TpwJhdJ1O\nt6FEOaOieFZtaaHinFhYMVcIPPV2RgF1WzIyPC63Q1jClHTt/1goFMeLnvkuJvJF\neFoVOlpm3rBTbYFOnvzqbErOCC5Sq+JyWwXbBmcnqR4ss0XxdTg5yKAqSwm1sS+G\nmaPrT2hbzjiShAzshFDZl9BF4w45mJ/jVsrCsi79pX/Er9x01VSxgohRnO3/h0FP\nu1O5llPD6mGSZ1mJH35/FxXiYp4Jfl5WV4BDN8PPZrbk/n4a+bckFHwrmHozKvuT\n2ELRZqThAgMBAAECggEAALS6Lix2cdJdMaZw0UsHjXAtu3CB7sSycrNxSfNdnohm\nOh8iJtyJvOn5IG2jtbcTVpJqa/rKB1olZoWSTFz04Cq9MnTvfxv1CkRA3Fk9lazj\nofgAVLS8LgY0T9Nk80uYjPeTOUf9rbfZLya8ys1OePyivoFjLLJu17b01i3aKrCv\nFtDq56fM0b1okul5AKTaEInSOLWqr+CAbqZKgJH4GCvnYDnFesoWZH/bIjPnkNSj\nPNDPeQ0bJlPIQTfgbY3n6SxYrF7MiyugUs4NYv2+5x57avCg/pErKtslas3vAEyB\nk9ITm7WAPgkOEi0G1PwNqslZWQm65GHExYASeDPQgQKBgQDQXljd7zhDAyAwOTnW\nCOQeTq5EH0jQo66/4c+cuybyhsv+eGP49BgJGtO/8JYbxs64Vdu64cA54ts1L00g\n+ld/0qTnllBJXtS8ac8Jw7Ec0TNWdGyd+i2iHyq+pLarQlXbP1JJQ71gnuBmUId7\nxZKWNWIr+AlZntZ5hSsbaRn+IQKBgQC63md8fdu7pa84ejG2FTAyJ19cfY71wl+X\ni1qMRolrdPbTX0Y//09AC4/TeqqcTbFjdoHQnPkm+N5Qr0JQO/xcnW0kalHsPasA\nq1OyBIL+/cCjWh0VIKo737z7ZEy/7erq2yNchhezKYqyOFWMqOvrQhrs1N4jKvGm\nfhuVvWBOwQKBgQC9AYfMsKyxJImzFIWg32nFEAvFITeUgZUG2tbOzXizr5Ak4KGW\ndhvDToFa7unKgliLHt50s0gBegTrRl/JFSv/d1w78b4b/5gq3DOAIPFsOVs/BdZs\nqj1wfKLyUUidKkvk15hVesuydAsK4+sxW3LvK7dEOPaKQ6k6i5AhigGY4QKBgBbD\nJpnsfsqRuRFJGSsT06YYU65hgPuArMI58FKr+OtE+RfwQA9lmxKzFdbt10Bo2lXC\nRrfRlyvfHr13S72YQPpLFtNbvBTpwh23wN8rW4WSBVggjvBW5UWb0FpS8mJiqb3n\n6ZvInbRuN/3TjOAwZ2Rlzf/RQinIfl0IeXW49rVBAoGAXXT7RLnjVCvUBo8mijT3\neRnp4umJoBQ94NMu70qWJWnb1ZMrRiL4vDwrLfMuaHX3lcQI7bIfJu2vFjdfNv2N\n5TbULs/uhNueXphYF0iWoNU6l/y5Ni627zPQ7q1jazE4F8NjKqk6R9hYcU85mIgv\nkd0bEXrDm3YGznRZlVAm8Ds=\n-----END PRIVATE KEY-----\n"
}
}
},
"changed": false
}
TASK [fedora.linux_system_roles.certificate : Stop tracking certificates] ******
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:176
Wednesday 10 July 2024 05:12:09 +0000 (0:00:00.022) 0:00:13.780 ********
ok: [sut] => (item={'cert': '/etc/pki/tls/certs/quadlet_demo.crt', 'cert_content': '-----BEGIN CERTIFICATE-----\nMIIDgjCCAmqgAwIBAgIQd4uYHvPLSJW8zGeAkDXEtjANBgkqhkiG9w0BAQsFADBQ\nMSAwHgYDVQQDDBdMb2NhbCBTaWduaW5nIEF1dGhvcml0eTEsMCoGA1UEAwwjNzc4\nYjk4MWUtZjNjYjQ4OTUtYmNjYzY3ODAtOTAzNWM0YjUwHhcNMjQwNzEwMDUxMjA4\nWhcNMjUwNzEwMDUxMjA3WjAUMRIwEAYDVQQDEwlsb2NhbGhvc3QwggEiMA0GCSqG\nSIb3DQEBAQUAA4IBDwAwggEKAoIBAQCYGZKdmimVEA5RpFhaV4ANPTYDdoDjwh15\nkdlNZfGAntuRAOn05085++50QfhomMryRl9TpwJhdJ1Ot6FEOaOieFZtaaHinFhY\nMVcIPPV2RgF1WzIyPC63Q1jClHTt/1goFMeLnvkuJvJFeFoVOlpm3rBTbYFOnvzq\nbErOCC5Sq+JyWwXbBmcnqR4ss0XxdTg5yKAqSwm1sS+GmaPrT2hbzjiShAzshFDZ\nl9BF4w45mJ/jVsrCsi79pX/Er9x01VSxgohRnO3/h0FPu1O5llPD6mGSZ1mJH35/\nFxXiYp4Jfl5WV4BDN8PPZrbk/n4a+bckFHwrmHozKvuT2ELRZqThAgMBAAGjgZMw\ngZAwCwYDVR0PBAQDAgWgMBQGA1UdEQQNMAuCCWxvY2FsaG9zdDAdBgNVHSUEFjAU\nBggrBgEFBQcDAQYIKwYBBQUHAwIwDAYDVR0TAQH/BAIwADAdBgNVHQ4EFgQUt8Ne\nGSHuq8g+jFS/xva/E+pO5rswHwYDVR0jBBgwFoAU6lgKTjgHbRLHq+q4Yl8HemVa\nW6gwDQYJKoZIhvcNAQELBQADggEBAEbHNoWEUfdUVpL0e1Xq0dRrUmiJ+irGDnAc\ncYLZQwBB4KIoH6D3CNueFruuR6lD7wtmSObfCSvK4afawROMgXMrUQ+Bpdd7iITI\nTByJzl5iCC45f5smd5/gGMP7TjwdFyVaOnQWFr4uyo8dTuqjcFqVZZK1zEYfT7hc\n0Xz8bIFHsEbHgMRhHqc9PCoLDw1zm+x9UcbKwzJQdMs6DJZ9scREh9rh1ccB1n1m\nUduk+aTuhQu47gcGUKb38jilXVYVo+8HMAd05nvSxyNkdkYB35PD3ieC0ZJx9W2R\nN1BI9qrvzeK+8iHIqdE42KebK2kQ2ZgNjKcI44hBKJs/inDddE4=\n-----END CERTIFICATE-----\n', 'key': '/etc/pki/tls/private/quadlet_demo.key', 'key_content': '-----BEGIN PRIVATE KEY-----\nMIIEvQIBADANBgkqhkiG9w0BAQEFAASCBKcwggSjAgEAAoIBAQCYGZKdmimVEA5R\npFhaV4ANPTYDdoDjwh15kdlNZfGAntuRAOn05085++50QfhomMryRl9TpwJhdJ1O\nt6FEOaOieFZtaaHinFhYMVcIPPV2RgF1WzIyPC63Q1jClHTt/1goFMeLnvkuJvJF\neFoVOlpm3rBTbYFOnvzqbErOCC5Sq+JyWwXbBmcnqR4ss0XxdTg5yKAqSwm1sS+G\nmaPrT2hbzjiShAzshFDZl9BF4w45mJ/jVsrCsi79pX/Er9x01VSxgohRnO3/h0FP\nu1O5llPD6mGSZ1mJH35/FxXiYp4Jfl5WV4BDN8PPZrbk/n4a+bckFHwrmHozKvuT\n2ELRZqThAgMBAAECggEAALS6Lix2cdJdMaZw0UsHjXAtu3CB7sSycrNxSfNdnohm\nOh8iJtyJvOn5IG2jtbcTVpJqa/rKB1olZoWSTFz04Cq9MnTvfxv1CkRA3Fk9lazj\nofgAVLS8LgY0T9Nk80uYjPeTOUf9rbfZLya8ys1OePyivoFjLLJu17b01i3aKrCv\nFtDq56fM0b1okul5AKTaEInSOLWqr+CAbqZKgJH4GCvnYDnFesoWZH/bIjPnkNSj\nPNDPeQ0bJlPIQTfgbY3n6SxYrF7MiyugUs4NYv2+5x57avCg/pErKtslas3vAEyB\nk9ITm7WAPgkOEi0G1PwNqslZWQm65GHExYASeDPQgQKBgQDQXljd7zhDAyAwOTnW\nCOQeTq5EH0jQo66/4c+cuybyhsv+eGP49BgJGtO/8JYbxs64Vdu64cA54ts1L00g\n+ld/0qTnllBJXtS8ac8Jw7Ec0TNWdGyd+i2iHyq+pLarQlXbP1JJQ71gnuBmUId7\nxZKWNWIr+AlZntZ5hSsbaRn+IQKBgQC63md8fdu7pa84ejG2FTAyJ19cfY71wl+X\ni1qMRolrdPbTX0Y//09AC4/TeqqcTbFjdoHQnPkm+N5Qr0JQO/xcnW0kalHsPasA\nq1OyBIL+/cCjWh0VIKo737z7ZEy/7erq2yNchhezKYqyOFWMqOvrQhrs1N4jKvGm\nfhuVvWBOwQKBgQC9AYfMsKyxJImzFIWg32nFEAvFITeUgZUG2tbOzXizr5Ak4KGW\ndhvDToFa7unKgliLHt50s0gBegTrRl/JFSv/d1w78b4b/5gq3DOAIPFsOVs/BdZs\nqj1wfKLyUUidKkvk15hVesuydAsK4+sxW3LvK7dEOPaKQ6k6i5AhigGY4QKBgBbD\nJpnsfsqRuRFJGSsT06YYU65hgPuArMI58FKr+OtE+RfwQA9lmxKzFdbt10Bo2lXC\nRrfRlyvfHr13S72YQPpLFtNbvBTpwh23wN8rW4WSBVggjvBW5UWb0FpS8mJiqb3n\n6ZvInbRuN/3TjOAwZ2Rlzf/RQinIfl0IeXW49rVBAoGAXXT7RLnjVCvUBo8mijT3\neRnp4umJoBQ94NMu70qWJWnb1ZMrRiL4vDwrLfMuaHX3lcQI7bIfJu2vFjdfNv2N\n5TbULs/uhNueXphYF0iWoNU6l/y5Ni627zPQ7q1jazE4F8NjKqk6R9hYcU85mIgv\nkd0bEXrDm3YGznRZlVAm8Ds=\n-----END PRIVATE KEY-----\n', 'ca': '/etc/pki/tls/certs/quadlet_demo.crt', 'ca_content': '-----BEGIN CERTIFICATE-----\nMIIDgjCCAmqgAwIBAgIQd4uYHvPLSJW8zGeAkDXEtjANBgkqhkiG9w0BAQsFADBQ\nMSAwHgYDVQQDDBdMb2NhbCBTaWduaW5nIEF1dGhvcml0eTEsMCoGA1UEAwwjNzc4\nYjk4MWUtZjNjYjQ4OTUtYmNjYzY3ODAtOTAzNWM0YjUwHhcNMjQwNzEwMDUxMjA4\nWhcNMjUwNzEwMDUxMjA3WjAUMRIwEAYDVQQDEwlsb2NhbGhvc3QwggEiMA0GCSqG\nSIb3DQEBAQUAA4IBDwAwggEKAoIBAQCYGZKdmimVEA5RpFhaV4ANPTYDdoDjwh15\nkdlNZfGAntuRAOn05085++50QfhomMryRl9TpwJhdJ1Ot6FEOaOieFZtaaHinFhY\nMVcIPPV2RgF1WzIyPC63Q1jClHTt/1goFMeLnvkuJvJFeFoVOlpm3rBTbYFOnvzq\nbErOCC5Sq+JyWwXbBmcnqR4ss0XxdTg5yKAqSwm1sS+GmaPrT2hbzjiShAzshFDZ\nl9BF4w45mJ/jVsrCsi79pX/Er9x01VSxgohRnO3/h0FPu1O5llPD6mGSZ1mJH35/\nFxXiYp4Jfl5WV4BDN8PPZrbk/n4a+bckFHwrmHozKvuT2ELRZqThAgMBAAGjgZMw\ngZAwCwYDVR0PBAQDAgWgMBQGA1UdEQQNMAuCCWxvY2FsaG9zdDAdBgNVHSUEFjAU\nBggrBgEFBQcDAQYIKwYBBQUHAwIwDAYDVR0TAQH/BAIwADAdBgNVHQ4EFgQUt8Ne\nGSHuq8g+jFS/xva/E+pO5rswHwYDVR0jBBgwFoAU6lgKTjgHbRLHq+q4Yl8HemVa\nW6gwDQYJKoZIhvcNAQELBQADggEBAEbHNoWEUfdUVpL0e1Xq0dRrUmiJ+irGDnAc\ncYLZQwBB4KIoH6D3CNueFruuR6lD7wtmSObfCSvK4afawROMgXMrUQ+Bpdd7iITI\nTByJzl5iCC45f5smd5/gGMP7TjwdFyVaOnQWFr4uyo8dTuqjcFqVZZK1zEYfT7hc\n0Xz8bIFHsEbHgMRhHqc9PCoLDw1zm+x9UcbKwzJQdMs6DJZ9scREh9rh1ccB1n1m\nUduk+aTuhQu47gcGUKb38jilXVYVo+8HMAd05nvSxyNkdkYB35PD3ieC0ZJx9W2R\nN1BI9qrvzeK+8iHIqdE42KebK2kQ2ZgNjKcI44hBKJs/inDddE4=\n-----END CERTIFICATE-----\n'}) => {
"ansible_loop_var": "item",
"changed": false,
"cmd": [
"getcert",
"stop-tracking",
"-f",
"/etc/pki/tls/certs/quadlet_demo.crt"
],
"delta": "0:00:00.030336",
"end": "2024-07-10 05:12:09.403801",
"item": {
"ca": "/etc/pki/tls/certs/quadlet_demo.crt",
"ca_content": "-----BEGIN CERTIFICATE-----\nMIIDgjCCAmqgAwIBAgIQd4uYHvPLSJW8zGeAkDXEtjANBgkqhkiG9w0BAQsFADBQ\nMSAwHgYDVQQDDBdMb2NhbCBTaWduaW5nIEF1dGhvcml0eTEsMCoGA1UEAwwjNzc4\nYjk4MWUtZjNjYjQ4OTUtYmNjYzY3ODAtOTAzNWM0YjUwHhcNMjQwNzEwMDUxMjA4\nWhcNMjUwNzEwMDUxMjA3WjAUMRIwEAYDVQQDEwlsb2NhbGhvc3QwggEiMA0GCSqG\nSIb3DQEBAQUAA4IBDwAwggEKAoIBAQCYGZKdmimVEA5RpFhaV4ANPTYDdoDjwh15\nkdlNZfGAntuRAOn05085++50QfhomMryRl9TpwJhdJ1Ot6FEOaOieFZtaaHinFhY\nMVcIPPV2RgF1WzIyPC63Q1jClHTt/1goFMeLnvkuJvJFeFoVOlpm3rBTbYFOnvzq\nbErOCC5Sq+JyWwXbBmcnqR4ss0XxdTg5yKAqSwm1sS+GmaPrT2hbzjiShAzshFDZ\nl9BF4w45mJ/jVsrCsi79pX/Er9x01VSxgohRnO3/h0FPu1O5llPD6mGSZ1mJH35/\nFxXiYp4Jfl5WV4BDN8PPZrbk/n4a+bckFHwrmHozKvuT2ELRZqThAgMBAAGjgZMw\ngZAwCwYDVR0PBAQDAgWgMBQGA1UdEQQNMAuCCWxvY2FsaG9zdDAdBgNVHSUEFjAU\nBggrBgEFBQcDAQYIKwYBBQUHAwIwDAYDVR0TAQH/BAIwADAdBgNVHQ4EFgQUt8Ne\nGSHuq8g+jFS/xva/E+pO5rswHwYDVR0jBBgwFoAU6lgKTjgHbRLHq+q4Yl8HemVa\nW6gwDQYJKoZIhvcNAQELBQADggEBAEbHNoWEUfdUVpL0e1Xq0dRrUmiJ+irGDnAc\ncYLZQwBB4KIoH6D3CNueFruuR6lD7wtmSObfCSvK4afawROMgXMrUQ+Bpdd7iITI\nTByJzl5iCC45f5smd5/gGMP7TjwdFyVaOnQWFr4uyo8dTuqjcFqVZZK1zEYfT7hc\n0Xz8bIFHsEbHgMRhHqc9PCoLDw1zm+x9UcbKwzJQdMs6DJZ9scREh9rh1ccB1n1m\nUduk+aTuhQu47gcGUKb38jilXVYVo+8HMAd05nvSxyNkdkYB35PD3ieC0ZJx9W2R\nN1BI9qrvzeK+8iHIqdE42KebK2kQ2ZgNjKcI44hBKJs/inDddE4=\n-----END CERTIFICATE-----\n",
"cert": "/etc/pki/tls/certs/quadlet_demo.crt",
"cert_content": "-----BEGIN CERTIFICATE-----\nMIIDgjCCAmqgAwIBAgIQd4uYHvPLSJW8zGeAkDXEtjANBgkqhkiG9w0BAQsFADBQ\nMSAwHgYDVQQDDBdMb2NhbCBTaWduaW5nIEF1dGhvcml0eTEsMCoGA1UEAwwjNzc4\nYjk4MWUtZjNjYjQ4OTUtYmNjYzY3ODAtOTAzNWM0YjUwHhcNMjQwNzEwMDUxMjA4\nWhcNMjUwNzEwMDUxMjA3WjAUMRIwEAYDVQQDEwlsb2NhbGhvc3QwggEiMA0GCSqG\nSIb3DQEBAQUAA4IBDwAwggEKAoIBAQCYGZKdmimVEA5RpFhaV4ANPTYDdoDjwh15\nkdlNZfGAntuRAOn05085++50QfhomMryRl9TpwJhdJ1Ot6FEOaOieFZtaaHinFhY\nMVcIPPV2RgF1WzIyPC63Q1jClHTt/1goFMeLnvkuJvJFeFoVOlpm3rBTbYFOnvzq\nbErOCC5Sq+JyWwXbBmcnqR4ss0XxdTg5yKAqSwm1sS+GmaPrT2hbzjiShAzshFDZ\nl9BF4w45mJ/jVsrCsi79pX/Er9x01VSxgohRnO3/h0FPu1O5llPD6mGSZ1mJH35/\nFxXiYp4Jfl5WV4BDN8PPZrbk/n4a+bckFHwrmHozKvuT2ELRZqThAgMBAAGjgZMw\ngZAwCwYDVR0PBAQDAgWgMBQGA1UdEQQNMAuCCWxvY2FsaG9zdDAdBgNVHSUEFjAU\nBggrBgEFBQcDAQYIKwYBBQUHAwIwDAYDVR0TAQH/BAIwADAdBgNVHQ4EFgQUt8Ne\nGSHuq8g+jFS/xva/E+pO5rswHwYDVR0jBBgwFoAU6lgKTjgHbRLHq+q4Yl8HemVa\nW6gwDQYJKoZIhvcNAQELBQADggEBAEbHNoWEUfdUVpL0e1Xq0dRrUmiJ+irGDnAc\ncYLZQwBB4KIoH6D3CNueFruuR6lD7wtmSObfCSvK4afawROMgXMrUQ+Bpdd7iITI\nTByJzl5iCC45f5smd5/gGMP7TjwdFyVaOnQWFr4uyo8dTuqjcFqVZZK1zEYfT7hc\n0Xz8bIFHsEbHgMRhHqc9PCoLDw1zm+x9UcbKwzJQdMs6DJZ9scREh9rh1ccB1n1m\nUduk+aTuhQu47gcGUKb38jilXVYVo+8HMAd05nvSxyNkdkYB35PD3ieC0ZJx9W2R\nN1BI9qrvzeK+8iHIqdE42KebK2kQ2ZgNjKcI44hBKJs/inDddE4=\n-----END CERTIFICATE-----\n",
"key": "/etc/pki/tls/private/quadlet_demo.key",
"key_content": "-----BEGIN PRIVATE KEY-----\nMIIEvQIBADANBgkqhkiG9w0BAQEFAASCBKcwggSjAgEAAoIBAQCYGZKdmimVEA5R\npFhaV4ANPTYDdoDjwh15kdlNZfGAntuRAOn05085++50QfhomMryRl9TpwJhdJ1O\nt6FEOaOieFZtaaHinFhYMVcIPPV2RgF1WzIyPC63Q1jClHTt/1goFMeLnvkuJvJF\neFoVOlpm3rBTbYFOnvzqbErOCC5Sq+JyWwXbBmcnqR4ss0XxdTg5yKAqSwm1sS+G\nmaPrT2hbzjiShAzshFDZl9BF4w45mJ/jVsrCsi79pX/Er9x01VSxgohRnO3/h0FP\nu1O5llPD6mGSZ1mJH35/FxXiYp4Jfl5WV4BDN8PPZrbk/n4a+bckFHwrmHozKvuT\n2ELRZqThAgMBAAECggEAALS6Lix2cdJdMaZw0UsHjXAtu3CB7sSycrNxSfNdnohm\nOh8iJtyJvOn5IG2jtbcTVpJqa/rKB1olZoWSTFz04Cq9MnTvfxv1CkRA3Fk9lazj\nofgAVLS8LgY0T9Nk80uYjPeTOUf9rbfZLya8ys1OePyivoFjLLJu17b01i3aKrCv\nFtDq56fM0b1okul5AKTaEInSOLWqr+CAbqZKgJH4GCvnYDnFesoWZH/bIjPnkNSj\nPNDPeQ0bJlPIQTfgbY3n6SxYrF7MiyugUs4NYv2+5x57avCg/pErKtslas3vAEyB\nk9ITm7WAPgkOEi0G1PwNqslZWQm65GHExYASeDPQgQKBgQDQXljd7zhDAyAwOTnW\nCOQeTq5EH0jQo66/4c+cuybyhsv+eGP49BgJGtO/8JYbxs64Vdu64cA54ts1L00g\n+ld/0qTnllBJXtS8ac8Jw7Ec0TNWdGyd+i2iHyq+pLarQlXbP1JJQ71gnuBmUId7\nxZKWNWIr+AlZntZ5hSsbaRn+IQKBgQC63md8fdu7pa84ejG2FTAyJ19cfY71wl+X\ni1qMRolrdPbTX0Y//09AC4/TeqqcTbFjdoHQnPkm+N5Qr0JQO/xcnW0kalHsPasA\nq1OyBIL+/cCjWh0VIKo737z7ZEy/7erq2yNchhezKYqyOFWMqOvrQhrs1N4jKvGm\nfhuVvWBOwQKBgQC9AYfMsKyxJImzFIWg32nFEAvFITeUgZUG2tbOzXizr5Ak4KGW\ndhvDToFa7unKgliLHt50s0gBegTrRl/JFSv/d1w78b4b/5gq3DOAIPFsOVs/BdZs\nqj1wfKLyUUidKkvk15hVesuydAsK4+sxW3LvK7dEOPaKQ6k6i5AhigGY4QKBgBbD\nJpnsfsqRuRFJGSsT06YYU65hgPuArMI58FKr+OtE+RfwQA9lmxKzFdbt10Bo2lXC\nRrfRlyvfHr13S72YQPpLFtNbvBTpwh23wN8rW4WSBVggjvBW5UWb0FpS8mJiqb3n\n6ZvInbRuN/3TjOAwZ2Rlzf/RQinIfl0IeXW49rVBAoGAXXT7RLnjVCvUBo8mijT3\neRnp4umJoBQ94NMu70qWJWnb1ZMrRiL4vDwrLfMuaHX3lcQI7bIfJu2vFjdfNv2N\n5TbULs/uhNueXphYF0iWoNU6l/y5Ni627zPQ7q1jazE4F8NjKqk6R9hYcU85mIgv\nkd0bEXrDm3YGznRZlVAm8Ds=\n-----END PRIVATE KEY-----\n"
},
"rc": 0,
"start": "2024-07-10 05:12:09.373465"
}
STDOUT:
Request "20240710051208" removed.
TASK [fedora.linux_system_roles.certificate : Remove files] ********************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:181
Wednesday 10 July 2024 05:12:09 +0000 (0:00:00.314) 0:00:14.094 ********
changed: [sut] => (item=/etc/pki/tls/certs/quadlet_demo.crt) => {
"ansible_loop_var": "item",
"changed": true,
"item": "/etc/pki/tls/certs/quadlet_demo.crt",
"path": "/etc/pki/tls/certs/quadlet_demo.crt",
"state": "absent"
}
changed: [sut] => (item=/etc/pki/tls/private/quadlet_demo.key) => {
"ansible_loop_var": "item",
"changed": true,
"item": "/etc/pki/tls/private/quadlet_demo.key",
"path": "/etc/pki/tls/private/quadlet_demo.key",
"state": "absent"
}
ok: [sut] => (item=/etc/pki/tls/certs/quadlet_demo.crt) => {
"ansible_loop_var": "item",
"changed": false,
"item": "/etc/pki/tls/certs/quadlet_demo.crt",
"path": "/etc/pki/tls/certs/quadlet_demo.crt",
"state": "absent"
}
TASK [Run the role] ************************************************************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/tests/tests_quadlet_demo.yml:50
Wednesday 10 July 2024 05:12:10 +0000 (0:00:00.673) 0:00:14.768 ********
TASK [fedora.linux_system_roles.podman : Set platform/version specific variables] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:3
Wednesday 10 July 2024 05:12:10 +0000 (0:00:00.032) 0:00:14.800 ********
included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml for sut
TASK [fedora.linux_system_roles.podman : Ensure ansible_facts used by role] ****
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:3
Wednesday 10 July 2024 05:12:10 +0000 (0:00:00.015) 0:00:14.816 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_required_facts | difference(ansible_facts.keys() | list) | length > 0",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Check if system is ostree] ************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:11
Wednesday 10 July 2024 05:12:10 +0000 (0:00:00.017) 0:00:14.833 ********
ok: [sut] => {
"changed": false,
"stat": {
"exists": false
}
}
TASK [fedora.linux_system_roles.podman : Set flag to indicate system is ostree] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:16
Wednesday 10 July 2024 05:12:10 +0000 (0:00:00.222) 0:00:15.056 ********
ok: [sut] => {
"ansible_facts": {
"__podman_is_ostree": false
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set platform/version specific variables] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:20
Wednesday 10 July 2024 05:12:10 +0000 (0:00:00.014) 0:00:15.071 ********
ok: [sut] => (item=RedHat.yml) => {
"ansible_facts": {
"__podman_packages": [
"podman",
"shadow-utils-subid"
]
},
"ansible_included_var_files": [
"/WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/vars/RedHat.yml"
],
"ansible_loop_var": "item",
"changed": false,
"item": "RedHat.yml"
}
ok: [sut] => (item=Fedora.yml) => {
"ansible_facts": {
"__podman_packages": [
"iptables-nft",
"podman",
"shadow-utils-subid"
]
},
"ansible_included_var_files": [
"/WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/vars/Fedora.yml"
],
"ansible_loop_var": "item",
"changed": false,
"item": "Fedora.yml"
}
skipping: [sut] => (item=Fedora_39.yml) => {
"ansible_loop_var": "item",
"changed": false,
"false_condition": "__vars_file is file",
"item": "Fedora_39.yml",
"skip_reason": "Conditional result was False"
}
skipping: [sut] => (item=Fedora_39.yml) => {
"ansible_loop_var": "item",
"changed": false,
"false_condition": "__vars_file is file",
"item": "Fedora_39.yml",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Gather the package facts] *************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:6
Wednesday 10 July 2024 05:12:10 +0000 (0:00:00.023) 0:00:15.095 ********
ok: [sut] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : Enable copr if requested] *************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:10
Wednesday 10 July 2024 05:12:11 +0000 (0:00:00.892) 0:00:15.987 ********
skipping: [sut] => {
"changed": false,
"false_condition": "podman_use_copr | d(false)",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Ensure required packages are installed] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:14
Wednesday 10 July 2024 05:12:11 +0000 (0:00:00.025) 0:00:16.013 ********
changed: [sut] => {
"changed": true,
"rc": 0,
"results": [
"Installed: passt-0^20240624.g1ee2eca-1.fc39.x86_64",
"Installed: passt-selinux-0^20240624.g1ee2eca-1.fc39.noarch",
"Installed: shadow-utils-subid-2:4.14.0-2.fc39.x86_64",
"Installed: catatonit-0.1.7-18.fc39.x86_64",
"Installed: fuse-overlayfs-1.13-1.fc39.x86_64",
"Installed: libslirp-4.7.0-4.fc39.x86_64",
"Installed: fuse-common-3.16.1-1.fc39.x86_64",
"Installed: qemu-user-static-2:8.1.3-5.fc39.x86_64",
"Installed: qemu-user-static-aarch64-2:8.1.3-5.fc39.x86_64",
"Installed: aardvark-dns-1.11.0-1.fc39.x86_64",
"Installed: podman-5:4.9.4-1.fc39.x86_64",
"Installed: qemu-user-static-alpha-2:8.1.3-5.fc39.x86_64",
"Installed: netavark-1.11.0-1.fc39.x86_64",
"Installed: qemu-user-static-arm-2:8.1.3-5.fc39.x86_64",
"Installed: criu-3.19-2.fc39.x86_64",
"Installed: fuse3-3.16.1-1.fc39.x86_64",
"Installed: criu-libs-3.19-2.fc39.x86_64",
"Installed: qemu-user-static-cris-2:8.1.3-5.fc39.x86_64",
"Installed: qemu-user-static-hexagon-2:8.1.3-5.fc39.x86_64",
"Installed: fuse3-libs-3.16.1-1.fc39.x86_64",
"Installed: qemu-user-static-hppa-2:8.1.3-5.fc39.x86_64",
"Installed: qemu-user-static-loongarch64-2:8.1.3-5.fc39.x86_64",
"Installed: qemu-user-static-m68k-2:8.1.3-5.fc39.x86_64",
"Installed: qemu-user-static-microblaze-2:8.1.3-5.fc39.x86_64",
"Installed: qemu-user-static-mips-2:8.1.3-5.fc39.x86_64",
"Installed: qemu-user-static-nios2-2:8.1.3-5.fc39.x86_64",
"Installed: qemu-user-static-or1k-2:8.1.3-5.fc39.x86_64",
"Installed: qemu-user-static-ppc-2:8.1.3-5.fc39.x86_64",
"Installed: qemu-user-static-riscv-2:8.1.3-5.fc39.x86_64",
"Installed: qemu-user-static-s390x-2:8.1.3-5.fc39.x86_64",
"Installed: crun-1.15-1.fc39.x86_64",
"Installed: qemu-user-static-sh4-2:8.1.3-5.fc39.x86_64",
"Installed: qemu-user-static-sparc-2:8.1.3-5.fc39.x86_64",
"Installed: qemu-user-static-x86-2:8.1.3-5.fc39.x86_64",
"Installed: qemu-user-static-xtensa-2:8.1.3-5.fc39.x86_64",
"Installed: yajl-2.1.0-22.fc39.x86_64",
"Installed: libnet-1.3-1.fc39.x86_64",
"Installed: conmon-2:2.1.12-1.fc39.x86_64",
"Installed: slirp4netns-1.2.2-1.fc39.x86_64",
"Installed: container-selinux-2:2.232.1-1.fc39.noarch",
"Installed: containers-common-4:1-99.fc39.noarch",
"Installed: containers-common-extra-4:1-99.fc39.noarch",
"Installed: gvisor-tap-vsock-6:0.7.3-1.fc39.x86_64",
"Installed: gvisor-tap-vsock-gvforwarder-6:0.7.3-1.fc39.x86_64",
"Installed: libbsd-0.12.2-3.fc39.x86_64",
"Installed: libmd-1.1.0-2.fc39.x86_64"
]
}
TASK [fedora.linux_system_roles.podman : Get podman version] *******************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:22
Wednesday 10 July 2024 05:13:46 +0000 (0:01:34.888) 0:01:50.901 ********
ok: [sut] => {
"changed": false,
"cmd": [
"podman",
"--version"
],
"delta": "0:00:00.039043",
"end": "2024-07-10 05:13:46.499533",
"rc": 0,
"start": "2024-07-10 05:13:46.460490"
}
STDOUT:
podman version 4.9.4
TASK [fedora.linux_system_roles.podman : Set podman version] *******************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:28
Wednesday 10 July 2024 05:13:46 +0000 (0:00:00.294) 0:01:51.196 ********
ok: [sut] => {
"ansible_facts": {
"podman_version": "4.9.4"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Podman package version must be 4.2 or later] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:32
Wednesday 10 July 2024 05:13:46 +0000 (0:00:00.019) 0:01:51.216 ********
skipping: [sut] => {
"changed": false,
"false_condition": "podman_version is version(\"4.2\", \"<\")",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Podman package version must be 4.4 or later for quadlet, secrets] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:39
Wednesday 10 July 2024 05:13:46 +0000 (0:00:00.017) 0:01:51.234 ********
skipping: [sut] => {
"changed": false,
"false_condition": "podman_version is version(\"4.4\", \"<\")",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Podman package version must be 4.4 or later for quadlet, secrets] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:49
Wednesday 10 July 2024 05:13:46 +0000 (0:00:00.045) 0:01:51.279 ********
META: end_host conditional evaluated to False, continuing execution for sut
skipping: [sut] => {
"skip_reason": "end_host conditional evaluated to False, continuing execution for sut"
}
MSG:
end_host conditional evaluated to false, continuing execution for sut
TASK [fedora.linux_system_roles.podman : Check user and group information] *****
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:56
Wednesday 10 July 2024 05:13:46 +0000 (0:00:00.045) 0:01:51.325 ********
included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for sut
TASK [fedora.linux_system_roles.podman : Get user information] *****************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2
Wednesday 10 July 2024 05:13:46 +0000 (0:00:00.053) 0:01:51.378 ********
ok: [sut] => {
"ansible_facts": {
"getent_passwd": {
"root": [
"x",
"0",
"0",
"Super User",
"/root",
"/bin/bash"
]
}
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Fail if user does not exist] **********
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9
Wednesday 10 July 2024 05:13:47 +0000 (0:00:00.301) 0:01:51.680 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not ansible_facts[\"getent_passwd\"][__podman_user]",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set group for podman user] ************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16
Wednesday 10 July 2024 05:13:47 +0000 (0:00:00.020) 0:01:51.701 ********
ok: [sut] => {
"ansible_facts": {
"__podman_group": "0"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Get group information] ****************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:28
Wednesday 10 July 2024 05:13:47 +0000 (0:00:00.025) 0:01:51.726 ********
ok: [sut] => {
"ansible_facts": {
"getent_group": {
"root": [
"x",
"0",
""
]
}
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set group name] ***********************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:35
Wednesday 10 July 2024 05:13:47 +0000 (0:00:00.235) 0:01:51.962 ********
ok: [sut] => {
"ansible_facts": {
"__podman_group_name": "root"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : See if getsubids exists] **************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:39
Wednesday 10 July 2024 05:13:47 +0000 (0:00:00.023) 0:01:51.986 ********
ok: [sut] => {
"changed": false,
"stat": {
"atime": 1696291200.0,
"attr_flags": "e",
"attributes": [
"extents"
],
"block_size": 4096,
"blocks": 32,
"charset": "binary",
"checksum": "72ba28a3be19c467e89ab90f5d21f99b605159a7",
"ctime": 1720588377.3007717,
"dev": 51714,
"device_type": 0,
"executable": true,
"exists": true,
"gid": 0,
"gr_name": "root",
"inode": 203603,
"isblk": false,
"ischr": false,
"isdir": false,
"isfifo": false,
"isgid": false,
"islnk": false,
"isreg": true,
"issock": false,
"isuid": false,
"mimetype": "application/x-pie-executable",
"mode": "0755",
"mtime": 1696291200.0,
"nlink": 1,
"path": "/usr/bin/getsubids",
"pw_name": "root",
"readable": true,
"rgrp": true,
"roth": true,
"rusr": true,
"size": 16032,
"uid": 0,
"version": "2092964490",
"wgrp": false,
"woth": false,
"writeable": true,
"wusr": true,
"xgrp": true,
"xoth": true,
"xusr": true
}
}
TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:50
Wednesday 10 July 2024 05:13:47 +0000 (0:00:00.234) 0:01:52.221 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_user not in [\"root\", \"0\"]",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Check group with getsubids] ***********
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:55
Wednesday 10 July 2024 05:13:47 +0000 (0:00:00.017) 0:01:52.239 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_user not in [\"root\", \"0\"]",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ******
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:60
Wednesday 10 July 2024 05:13:47 +0000 (0:00:00.017) 0:01:52.256 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_user not in [\"root\", \"0\"]",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Get subuid file] **********************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:74
Wednesday 10 July 2024 05:13:47 +0000 (0:00:00.017) 0:01:52.273 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_stat_getsubids.stat.exists",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Get subgid file] **********************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:79
Wednesday 10 July 2024 05:13:47 +0000 (0:00:00.015) 0:01:52.289 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_stat_getsubids.stat.exists",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ******
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:84
Wednesday 10 July 2024 05:13:47 +0000 (0:00:00.015) 0:01:52.305 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_stat_getsubids.stat.exists",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ******
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:94
Wednesday 10 July 2024 05:13:47 +0000 (0:00:00.015) 0:01:52.320 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_stat_getsubids.stat.exists",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] *****
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:101
Wednesday 10 July 2024 05:13:47 +0000 (0:00:00.015) 0:01:52.336 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_stat_getsubids.stat.exists",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set config file paths] ****************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:62
Wednesday 10 July 2024 05:13:47 +0000 (0:00:00.015) 0:01:52.352 ********
ok: [sut] => {
"ansible_facts": {
"__podman_container_conf_file": "/etc/containers/containers.conf.d/50-systemroles.conf",
"__podman_policy_json_file": "/etc/containers/policy.json",
"__podman_registries_conf_file": "/etc/containers/registries.conf.d/50-systemroles.conf",
"__podman_storage_conf_file": "/etc/containers/storage.conf"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Handle container.conf.d] **************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:71
Wednesday 10 July 2024 05:13:47 +0000 (0:00:00.034) 0:01:52.387 ********
included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_container_conf_d.yml for sut
TASK [fedora.linux_system_roles.podman : Ensure containers.d exists] ***********
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_container_conf_d.yml:5
Wednesday 10 July 2024 05:13:47 +0000 (0:00:00.033) 0:01:52.420 ********
skipping: [sut] => {
"changed": false,
"false_condition": "podman_containers_conf | length > 0",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Update container config file] *********
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_container_conf_d.yml:13
Wednesday 10 July 2024 05:13:47 +0000 (0:00:00.038) 0:01:52.458 ********
skipping: [sut] => {
"changed": false,
"false_condition": "podman_containers_conf | length > 0",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Handle registries.conf.d] *************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:74
Wednesday 10 July 2024 05:13:47 +0000 (0:00:00.018) 0:01:52.477 ********
included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_registries_conf_d.yml for sut
TASK [fedora.linux_system_roles.podman : Ensure registries.d exists] ***********
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_registries_conf_d.yml:5
Wednesday 10 July 2024 05:13:47 +0000 (0:00:00.033) 0:01:52.511 ********
skipping: [sut] => {
"changed": false,
"false_condition": "podman_registries_conf | length > 0",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Update registries config file] ********
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_registries_conf_d.yml:13
Wednesday 10 July 2024 05:13:47 +0000 (0:00:00.017) 0:01:52.528 ********
skipping: [sut] => {
"changed": false,
"false_condition": "podman_registries_conf | length > 0",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Handle storage.conf] ******************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:77
Wednesday 10 July 2024 05:13:47 +0000 (0:00:00.016) 0:01:52.545 ********
included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_storage_conf.yml for sut
TASK [fedora.linux_system_roles.podman : Ensure storage.conf parent dir exists] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_storage_conf.yml:5
Wednesday 10 July 2024 05:13:47 +0000 (0:00:00.034) 0:01:52.579 ********
skipping: [sut] => {
"changed": false,
"false_condition": "podman_storage_conf | length > 0",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Update storage config file] ***********
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_storage_conf.yml:13
Wednesday 10 July 2024 05:13:47 +0000 (0:00:00.017) 0:01:52.596 ********
skipping: [sut] => {
"changed": false,
"false_condition": "podman_storage_conf | length > 0",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Handle policy.json] *******************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:80
Wednesday 10 July 2024 05:13:47 +0000 (0:00:00.016) 0:01:52.613 ********
included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml for sut
TASK [fedora.linux_system_roles.podman : Ensure policy.json parent dir exists] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:6
Wednesday 10 July 2024 05:13:47 +0000 (0:00:00.036) 0:01:52.649 ********
skipping: [sut] => {
"changed": false,
"false_condition": "podman_policy_json | length > 0",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Stat the policy.json file] ************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:14
Wednesday 10 July 2024 05:13:48 +0000 (0:00:00.016) 0:01:52.666 ********
skipping: [sut] => {
"changed": false,
"false_condition": "podman_policy_json | length > 0",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Get the existing policy.json] *********
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:19
Wednesday 10 July 2024 05:13:48 +0000 (0:00:00.017) 0:01:52.683 ********
skipping: [sut] => {
"changed": false,
"false_condition": "podman_policy_json | length > 0",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Write new policy.json file] ***********
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:25
Wednesday 10 July 2024 05:13:48 +0000 (0:00:00.016) 0:01:52.700 ********
skipping: [sut] => {
"changed": false,
"false_condition": "podman_policy_json | length > 0",
"skip_reason": "Conditional result was False"
}
TASK [Manage firewall for specified ports] *************************************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:86
Wednesday 10 July 2024 05:13:48 +0000 (0:00:00.017) 0:01:52.717 ********
TASK [fedora.linux_system_roles.firewall : Setup firewalld] ********************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:2
Wednesday 10 July 2024 05:13:48 +0000 (0:00:00.082) 0:01:52.799 ********
included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml for sut
TASK [fedora.linux_system_roles.firewall : Ensure ansible_facts used by role] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:2
Wednesday 10 July 2024 05:13:48 +0000 (0:00:00.030) 0:01:52.829 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__firewall_required_facts | difference(ansible_facts.keys() | list) | length > 0",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.firewall : Check if system is ostree] **********
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:10
Wednesday 10 July 2024 05:13:48 +0000 (0:00:00.023) 0:01:52.853 ********
ok: [sut] => {
"changed": false,
"stat": {
"exists": false
}
}
TASK [fedora.linux_system_roles.firewall : Set flag to indicate system is ostree] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:15
Wednesday 10 July 2024 05:13:48 +0000 (0:00:00.231) 0:01:53.085 ********
ok: [sut] => {
"ansible_facts": {
"__firewall_is_ostree": false
},
"changed": false
}
TASK [fedora.linux_system_roles.firewall : Install firewalld] ******************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:19
Wednesday 10 July 2024 05:13:48 +0000 (0:00:00.022) 0:01:53.107 ********
ok: [sut] => {
"changed": false,
"rc": 0,
"results": []
}
MSG:
Nothing to do
TASK [fedora.linux_system_roles.firewall : Collect service facts] **************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:5
Wednesday 10 July 2024 05:13:50 +0000 (0:00:02.180) 0:01:55.288 ********
skipping: [sut] => {
"changed": false,
"false_condition": "firewall_disable_conflicting_services | bool",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.firewall : Attempt to stop and disable conflicting services] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:9
Wednesday 10 July 2024 05:13:50 +0000 (0:00:00.017) 0:01:55.305 ********
skipping: [sut] => (item=nftables) => {
"ansible_loop_var": "item",
"changed": false,
"false_condition": "firewall_disable_conflicting_services | bool",
"item": "nftables",
"skip_reason": "Conditional result was False"
}
skipping: [sut] => (item=iptables) => {
"ansible_loop_var": "item",
"changed": false,
"false_condition": "firewall_disable_conflicting_services | bool",
"item": "iptables",
"skip_reason": "Conditional result was False"
}
skipping: [sut] => (item=ufw) => {
"ansible_loop_var": "item",
"changed": false,
"false_condition": "firewall_disable_conflicting_services | bool",
"item": "ufw",
"skip_reason": "Conditional result was False"
}
skipping: [sut] => {
"changed": false
}
MSG:
All items skipped
TASK [fedora.linux_system_roles.firewall : Unmask firewalld service] ***********
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:22
Wednesday 10 July 2024 05:13:50 +0000 (0:00:00.020) 0:01:55.326 ********
ok: [sut] => {
"changed": false,
"name": "firewalld",
"status": {
"AccessSELinuxContext": "system_u:object_r:firewalld_unit_file_t:s0",
"ActiveEnterTimestampMonotonic": "0",
"ActiveExitTimestampMonotonic": "0",
"ActiveState": "inactive",
"After": "sysinit.target dbus-broker.service basic.target dbus.socket system.slice polkit.service",
"AllowIsolate": "no",
"AssertResult": "no",
"AssertTimestampMonotonic": "0",
"Before": "shutdown.target network-pre.target",
"BlockIOAccounting": "no",
"BlockIOWeight": "[not set]",
"BusName": "org.fedoraproject.FirewallD1",
"CPUAccounting": "yes",
"CPUAffinityFromNUMA": "no",
"CPUQuotaPerSecUSec": "infinity",
"CPUQuotaPeriodUSec": "infinity",
"CPUSchedulingPolicy": "0",
"CPUSchedulingPriority": "0",
"CPUSchedulingResetOnFork": "no",
"CPUShares": "[not set]",
"CPUUsageNSec": "[not set]",
"CPUWeight": "[not set]",
"CacheDirectoryMode": "0755",
"CanFreeze": "yes",
"CanIsolate": "no",
"CanReload": "yes",
"CanStart": "yes",
"CanStop": "yes",
"CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore",
"CleanResult": "success",
"CollectMode": "inactive",
"ConditionResult": "no",
"ConditionTimestampMonotonic": "0",
"ConfigurationDirectoryMode": "0755",
"Conflicts": "ip6tables.service iptables.service nftables.service shutdown.target ebtables.service ipset.service",
"ControlGroupId": "0",
"ControlPID": "0",
"CoredumpFilter": "0x33",
"DefaultDependencies": "yes",
"DefaultMemoryLow": "0",
"DefaultMemoryMin": "0",
"DefaultStartupMemoryLow": "0",
"Delegate": "no",
"Description": "firewalld - dynamic firewall daemon",
"DevicePolicy": "auto",
"Documentation": "\"man:firewalld(1)\"",
"DropInPaths": "/usr/lib/systemd/system/service.d/10-timeout-abort.conf",
"DynamicUser": "no",
"EnvironmentFiles": "/etc/sysconfig/firewalld (ignore_errors=yes)",
"ExecMainCode": "0",
"ExecMainExitTimestampMonotonic": "0",
"ExecMainPID": "0",
"ExecMainStartTimestampMonotonic": "0",
"ExecMainStatus": "0",
"ExecReload": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"ExecReloadEx": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"ExecStart": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"ExecStartEx": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"ExitType": "main",
"ExtensionImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent",
"FailureAction": "none",
"FileDescriptorStoreMax": "0",
"FileDescriptorStorePreserve": "restart",
"FinalKillSignal": "9",
"FragmentPath": "/usr/lib/systemd/system/firewalld.service",
"FreezerState": "running",
"GID": "[not set]",
"GuessMainPID": "yes",
"IOAccounting": "no",
"IOReadBytes": "18446744073709551615",
"IOReadOperations": "18446744073709551615",
"IOSchedulingClass": "2",
"IOSchedulingPriority": "4",
"IOWeight": "[not set]",
"IOWriteBytes": "18446744073709551615",
"IOWriteOperations": "18446744073709551615",
"IPAccounting": "no",
"IPEgressBytes": "[no data]",
"IPEgressPackets": "[no data]",
"IPIngressBytes": "[no data]",
"IPIngressPackets": "[no data]",
"Id": "firewalld.service",
"IgnoreOnIsolate": "no",
"IgnoreSIGPIPE": "yes",
"InactiveEnterTimestampMonotonic": "0",
"InactiveExitTimestampMonotonic": "0",
"JobRunningTimeoutUSec": "infinity",
"JobTimeoutAction": "none",
"JobTimeoutUSec": "infinity",
"KeyringMode": "private",
"KillMode": "mixed",
"KillSignal": "15",
"LimitAS": "infinity",
"LimitASSoft": "infinity",
"LimitCORE": "infinity",
"LimitCORESoft": "infinity",
"LimitCPU": "infinity",
"LimitCPUSoft": "infinity",
"LimitDATA": "infinity",
"LimitDATASoft": "infinity",
"LimitFSIZE": "infinity",
"LimitFSIZESoft": "infinity",
"LimitLOCKS": "infinity",
"LimitLOCKSSoft": "infinity",
"LimitMEMLOCK": "8388608",
"LimitMEMLOCKSoft": "8388608",
"LimitMSGQUEUE": "819200",
"LimitMSGQUEUESoft": "819200",
"LimitNICE": "0",
"LimitNICESoft": "0",
"LimitNOFILE": "524288",
"LimitNOFILESoft": "1024",
"LimitNPROC": "14756",
"LimitNPROCSoft": "14756",
"LimitRSS": "infinity",
"LimitRSSSoft": "infinity",
"LimitRTPRIO": "0",
"LimitRTPRIOSoft": "0",
"LimitRTTIME": "infinity",
"LimitRTTIMESoft": "infinity",
"LimitSIGPENDING": "14756",
"LimitSIGPENDINGSoft": "14756",
"LimitSTACK": "infinity",
"LimitSTACKSoft": "8388608",
"LoadState": "loaded",
"LockPersonality": "no",
"LogLevelMax": "-1",
"LogRateLimitBurst": "0",
"LogRateLimitIntervalUSec": "0",
"LogsDirectoryMode": "0755",
"MainPID": "0",
"ManagedOOMMemoryPressure": "auto",
"ManagedOOMMemoryPressureLimit": "0",
"ManagedOOMPreference": "none",
"ManagedOOMSwap": "auto",
"MemoryAccounting": "yes",
"MemoryAvailable": "infinity",
"MemoryCurrent": "[not set]",
"MemoryDenyWriteExecute": "no",
"MemoryHigh": "infinity",
"MemoryKSM": "no",
"MemoryLimit": "infinity",
"MemoryLow": "0",
"MemoryMax": "infinity",
"MemoryMin": "0",
"MemoryPressureThresholdUSec": "200ms",
"MemoryPressureWatch": "auto",
"MemorySwapMax": "infinity",
"MemoryZSwapMax": "infinity",
"MountAPIVFS": "no",
"MountImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent",
"NFileDescriptorStore": "0",
"NRestarts": "0",
"NUMAPolicy": "n/a",
"Names": "firewalld.service",
"NeedDaemonReload": "no",
"Nice": "0",
"NoNewPrivileges": "no",
"NonBlocking": "no",
"NotifyAccess": "none",
"OOMPolicy": "stop",
"OOMScoreAdjust": "0",
"OnFailureJobMode": "replace",
"OnSuccessJobMode": "fail",
"Perpetual": "no",
"PrivateDevices": "no",
"PrivateIPC": "no",
"PrivateMounts": "no",
"PrivateNetwork": "no",
"PrivateTmp": "no",
"PrivateUsers": "no",
"ProcSubset": "all",
"ProtectClock": "no",
"ProtectControlGroups": "no",
"ProtectHome": "no",
"ProtectHostname": "no",
"ProtectKernelLogs": "no",
"ProtectKernelModules": "no",
"ProtectKernelTunables": "no",
"ProtectProc": "default",
"ProtectSystem": "no",
"RefuseManualStart": "no",
"RefuseManualStop": "no",
"ReloadResult": "success",
"ReloadSignal": "1",
"RemainAfterExit": "no",
"RemoveIPC": "no",
"Requires": "dbus.socket system.slice sysinit.target",
"Restart": "no",
"RestartKillSignal": "15",
"RestartMaxDelayUSec": "infinity",
"RestartMode": "normal",
"RestartSteps": "0",
"RestartUSec": "100ms",
"RestartUSecNext": "100ms",
"RestrictNamespaces": "no",
"RestrictRealtime": "no",
"RestrictSUIDSGID": "no",
"Result": "success",
"RootDirectoryStartOnly": "no",
"RootEphemeral": "no",
"RootImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent",
"RuntimeDirectoryMode": "0755",
"RuntimeDirectoryPreserve": "no",
"RuntimeMaxUSec": "infinity",
"RuntimeRandomizedExtraUSec": "0",
"SameProcessGroup": "no",
"SecureBits": "0",
"SendSIGHUP": "no",
"SendSIGKILL": "yes",
"Slice": "system.slice",
"StandardError": "null",
"StandardInput": "null",
"StandardOutput": "null",
"StartLimitAction": "none",
"StartLimitBurst": "5",
"StartLimitIntervalUSec": "10s",
"StartupBlockIOWeight": "[not set]",
"StartupCPUShares": "[not set]",
"StartupCPUWeight": "[not set]",
"StartupIOWeight": "[not set]",
"StartupMemoryHigh": "18446744073709551615",
"StartupMemoryLow": "0",
"StartupMemoryMax": "18446744073709551615",
"StartupMemorySwapMax": "18446744073709551615",
"StartupMemoryZSwapMax": "18446744073709551615",
"StateChangeTimestampMonotonic": "0",
"StateDirectoryMode": "0755",
"StatusErrno": "0",
"StopWhenUnneeded": "no",
"SubState": "dead",
"SuccessAction": "none",
"SyslogFacility": "3",
"SyslogLevel": "6",
"SyslogLevelPrefix": "yes",
"SyslogPriority": "30",
"SystemCallErrorNumber": "2147483646",
"TTYReset": "no",
"TTYVHangup": "no",
"TTYVTDisallocate": "no",
"TasksAccounting": "yes",
"TasksCurrent": "[not set]",
"TasksMax": "4426",
"TimeoutAbortUSec": "45s",
"TimeoutCleanUSec": "infinity",
"TimeoutStartFailureMode": "terminate",
"TimeoutStartUSec": "45s",
"TimeoutStopFailureMode": "abort",
"TimeoutStopUSec": "45s",
"TimerSlackNSec": "50000",
"Transient": "no",
"Type": "dbus",
"UID": "[not set]",
"UMask": "0022",
"UnitFilePreset": "enabled",
"UnitFileState": "disabled",
"UtmpMode": "init",
"Wants": "network-pre.target",
"WatchdogSignal": "6",
"WatchdogTimestampMonotonic": "0",
"WatchdogUSec": "infinity"
}
}
TASK [fedora.linux_system_roles.firewall : Enable and start firewalld service] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:28
Wednesday 10 July 2024 05:13:51 +0000 (0:00:00.416) 0:01:55.742 ********
changed: [sut] => {
"changed": true,
"enabled": true,
"name": "firewalld",
"state": "started",
"status": {
"AccessSELinuxContext": "system_u:object_r:firewalld_unit_file_t:s0",
"ActiveEnterTimestampMonotonic": "0",
"ActiveExitTimestampMonotonic": "0",
"ActiveState": "inactive",
"After": "system.slice polkit.service dbus-broker.service basic.target dbus.socket sysinit.target",
"AllowIsolate": "no",
"AssertResult": "no",
"AssertTimestampMonotonic": "0",
"Before": "shutdown.target network-pre.target",
"BlockIOAccounting": "no",
"BlockIOWeight": "[not set]",
"BusName": "org.fedoraproject.FirewallD1",
"CPUAccounting": "yes",
"CPUAffinityFromNUMA": "no",
"CPUQuotaPerSecUSec": "infinity",
"CPUQuotaPeriodUSec": "infinity",
"CPUSchedulingPolicy": "0",
"CPUSchedulingPriority": "0",
"CPUSchedulingResetOnFork": "no",
"CPUShares": "[not set]",
"CPUUsageNSec": "[not set]",
"CPUWeight": "[not set]",
"CacheDirectoryMode": "0755",
"CanFreeze": "yes",
"CanIsolate": "no",
"CanReload": "yes",
"CanStart": "yes",
"CanStop": "yes",
"CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore",
"CleanResult": "success",
"CollectMode": "inactive",
"ConditionResult": "no",
"ConditionTimestampMonotonic": "0",
"ConfigurationDirectoryMode": "0755",
"Conflicts": "ebtables.service ipset.service nftables.service ip6tables.service shutdown.target iptables.service",
"ControlGroupId": "0",
"ControlPID": "0",
"CoredumpFilter": "0x33",
"DefaultDependencies": "yes",
"DefaultMemoryLow": "0",
"DefaultMemoryMin": "0",
"DefaultStartupMemoryLow": "0",
"Delegate": "no",
"Description": "firewalld - dynamic firewall daemon",
"DevicePolicy": "auto",
"Documentation": "\"man:firewalld(1)\"",
"DropInPaths": "/usr/lib/systemd/system/service.d/10-timeout-abort.conf",
"DynamicUser": "no",
"EnvironmentFiles": "/etc/sysconfig/firewalld (ignore_errors=yes)",
"ExecMainCode": "0",
"ExecMainExitTimestampMonotonic": "0",
"ExecMainPID": "0",
"ExecMainStartTimestampMonotonic": "0",
"ExecMainStatus": "0",
"ExecReload": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"ExecReloadEx": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"ExecStart": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"ExecStartEx": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"ExitType": "main",
"ExtensionImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent",
"FailureAction": "none",
"FileDescriptorStoreMax": "0",
"FileDescriptorStorePreserve": "restart",
"FinalKillSignal": "9",
"FragmentPath": "/usr/lib/systemd/system/firewalld.service",
"FreezerState": "running",
"GID": "[not set]",
"GuessMainPID": "yes",
"IOAccounting": "no",
"IOReadBytes": "18446744073709551615",
"IOReadOperations": "18446744073709551615",
"IOSchedulingClass": "2",
"IOSchedulingPriority": "4",
"IOWeight": "[not set]",
"IOWriteBytes": "18446744073709551615",
"IOWriteOperations": "18446744073709551615",
"IPAccounting": "no",
"IPEgressBytes": "[no data]",
"IPEgressPackets": "[no data]",
"IPIngressBytes": "[no data]",
"IPIngressPackets": "[no data]",
"Id": "firewalld.service",
"IgnoreOnIsolate": "no",
"IgnoreSIGPIPE": "yes",
"InactiveEnterTimestampMonotonic": "0",
"InactiveExitTimestampMonotonic": "0",
"JobRunningTimeoutUSec": "infinity",
"JobTimeoutAction": "none",
"JobTimeoutUSec": "infinity",
"KeyringMode": "private",
"KillMode": "mixed",
"KillSignal": "15",
"LimitAS": "infinity",
"LimitASSoft": "infinity",
"LimitCORE": "infinity",
"LimitCORESoft": "infinity",
"LimitCPU": "infinity",
"LimitCPUSoft": "infinity",
"LimitDATA": "infinity",
"LimitDATASoft": "infinity",
"LimitFSIZE": "infinity",
"LimitFSIZESoft": "infinity",
"LimitLOCKS": "infinity",
"LimitLOCKSSoft": "infinity",
"LimitMEMLOCK": "8388608",
"LimitMEMLOCKSoft": "8388608",
"LimitMSGQUEUE": "819200",
"LimitMSGQUEUESoft": "819200",
"LimitNICE": "0",
"LimitNICESoft": "0",
"LimitNOFILE": "524288",
"LimitNOFILESoft": "1024",
"LimitNPROC": "14756",
"LimitNPROCSoft": "14756",
"LimitRSS": "infinity",
"LimitRSSSoft": "infinity",
"LimitRTPRIO": "0",
"LimitRTPRIOSoft": "0",
"LimitRTTIME": "infinity",
"LimitRTTIMESoft": "infinity",
"LimitSIGPENDING": "14756",
"LimitSIGPENDINGSoft": "14756",
"LimitSTACK": "infinity",
"LimitSTACKSoft": "8388608",
"LoadState": "loaded",
"LockPersonality": "no",
"LogLevelMax": "-1",
"LogRateLimitBurst": "0",
"LogRateLimitIntervalUSec": "0",
"LogsDirectoryMode": "0755",
"MainPID": "0",
"ManagedOOMMemoryPressure": "auto",
"ManagedOOMMemoryPressureLimit": "0",
"ManagedOOMPreference": "none",
"ManagedOOMSwap": "auto",
"MemoryAccounting": "yes",
"MemoryAvailable": "infinity",
"MemoryCurrent": "[not set]",
"MemoryDenyWriteExecute": "no",
"MemoryHigh": "infinity",
"MemoryKSM": "no",
"MemoryLimit": "infinity",
"MemoryLow": "0",
"MemoryMax": "infinity",
"MemoryMin": "0",
"MemoryPressureThresholdUSec": "200ms",
"MemoryPressureWatch": "auto",
"MemorySwapMax": "infinity",
"MemoryZSwapMax": "infinity",
"MountAPIVFS": "no",
"MountImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent",
"NFileDescriptorStore": "0",
"NRestarts": "0",
"NUMAPolicy": "n/a",
"Names": "firewalld.service",
"NeedDaemonReload": "no",
"Nice": "0",
"NoNewPrivileges": "no",
"NonBlocking": "no",
"NotifyAccess": "none",
"OOMPolicy": "stop",
"OOMScoreAdjust": "0",
"OnFailureJobMode": "replace",
"OnSuccessJobMode": "fail",
"Perpetual": "no",
"PrivateDevices": "no",
"PrivateIPC": "no",
"PrivateMounts": "no",
"PrivateNetwork": "no",
"PrivateTmp": "no",
"PrivateUsers": "no",
"ProcSubset": "all",
"ProtectClock": "no",
"ProtectControlGroups": "no",
"ProtectHome": "no",
"ProtectHostname": "no",
"ProtectKernelLogs": "no",
"ProtectKernelModules": "no",
"ProtectKernelTunables": "no",
"ProtectProc": "default",
"ProtectSystem": "no",
"RefuseManualStart": "no",
"RefuseManualStop": "no",
"ReloadResult": "success",
"ReloadSignal": "1",
"RemainAfterExit": "no",
"RemoveIPC": "no",
"Requires": "system.slice dbus.socket sysinit.target",
"Restart": "no",
"RestartKillSignal": "15",
"RestartMaxDelayUSec": "infinity",
"RestartMode": "normal",
"RestartSteps": "0",
"RestartUSec": "100ms",
"RestartUSecNext": "100ms",
"RestrictNamespaces": "no",
"RestrictRealtime": "no",
"RestrictSUIDSGID": "no",
"Result": "success",
"RootDirectoryStartOnly": "no",
"RootEphemeral": "no",
"RootImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent",
"RuntimeDirectoryMode": "0755",
"RuntimeDirectoryPreserve": "no",
"RuntimeMaxUSec": "infinity",
"RuntimeRandomizedExtraUSec": "0",
"SameProcessGroup": "no",
"SecureBits": "0",
"SendSIGHUP": "no",
"SendSIGKILL": "yes",
"Slice": "system.slice",
"StandardError": "null",
"StandardInput": "null",
"StandardOutput": "null",
"StartLimitAction": "none",
"StartLimitBurst": "5",
"StartLimitIntervalUSec": "10s",
"StartupBlockIOWeight": "[not set]",
"StartupCPUShares": "[not set]",
"StartupCPUWeight": "[not set]",
"StartupIOWeight": "[not set]",
"StartupMemoryHigh": "18446744073709551615",
"StartupMemoryLow": "0",
"StartupMemoryMax": "18446744073709551615",
"StartupMemorySwapMax": "18446744073709551615",
"StartupMemoryZSwapMax": "18446744073709551615",
"StateChangeTimestampMonotonic": "0",
"StateDirectoryMode": "0755",
"StatusErrno": "0",
"StopWhenUnneeded": "no",
"SubState": "dead",
"SuccessAction": "none",
"SyslogFacility": "3",
"SyslogLevel": "6",
"SyslogLevelPrefix": "yes",
"SyslogPriority": "30",
"SystemCallErrorNumber": "2147483646",
"TTYReset": "no",
"TTYVHangup": "no",
"TTYVTDisallocate": "no",
"TasksAccounting": "yes",
"TasksCurrent": "[not set]",
"TasksMax": "4426",
"TimeoutAbortUSec": "45s",
"TimeoutCleanUSec": "infinity",
"TimeoutStartFailureMode": "terminate",
"TimeoutStartUSec": "45s",
"TimeoutStopFailureMode": "abort",
"TimeoutStopUSec": "45s",
"TimerSlackNSec": "50000",
"Transient": "no",
"Type": "dbus",
"UID": "[not set]",
"UMask": "0022",
"UnitFilePreset": "enabled",
"UnitFileState": "disabled",
"UtmpMode": "init",
"Wants": "network-pre.target",
"WatchdogSignal": "6",
"WatchdogTimestampMonotonic": "0",
"WatchdogUSec": "infinity"
}
}
TASK [fedora.linux_system_roles.firewall : Check if previous replaced is defined] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:34
Wednesday 10 July 2024 05:13:52 +0000 (0:00:01.093) 0:01:56.836 ********
ok: [sut] => {
"ansible_facts": {
"__firewall_previous_replaced": false,
"__firewall_python_cmd": "/usr/bin/python3",
"__firewall_report_changed": true
},
"changed": false
}
TASK [fedora.linux_system_roles.firewall : Get config files, checksums before and remove] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:43
Wednesday 10 July 2024 05:13:52 +0000 (0:00:00.025) 0:01:56.861 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__firewall_previous_replaced | bool",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.firewall : Tell firewall module it is able to report changed] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:55
Wednesday 10 July 2024 05:13:52 +0000 (0:00:00.017) 0:01:56.878 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__firewall_previous_replaced | bool",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.firewall : Configure firewall] *****************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:71
Wednesday 10 July 2024 05:13:52 +0000 (0:00:00.016) 0:01:56.894 ********
changed: [sut] => (item={'port': '8000/tcp', 'state': 'enabled'}) => {
"__firewall_changed": true,
"ansible_loop_var": "item",
"changed": true,
"item": {
"port": "8000/tcp",
"state": "enabled"
}
}
changed: [sut] => (item={'port': '9000/tcp', 'state': 'enabled'}) => {
"__firewall_changed": true,
"ansible_loop_var": "item",
"changed": true,
"item": {
"port": "9000/tcp",
"state": "enabled"
}
}
TASK [fedora.linux_system_roles.firewall : Gather firewall config information] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:120
Wednesday 10 July 2024 05:13:53 +0000 (0:00:01.168) 0:01:58.063 ********
skipping: [sut] => (item={'port': '8000/tcp', 'state': 'enabled'}) => {
"ansible_loop_var": "item",
"changed": false,
"false_condition": "firewall | length == 1",
"item": {
"port": "8000/tcp",
"state": "enabled"
},
"skip_reason": "Conditional result was False"
}
skipping: [sut] => (item={'port': '9000/tcp', 'state': 'enabled'}) => {
"ansible_loop_var": "item",
"changed": false,
"false_condition": "firewall | length == 1",
"item": {
"port": "9000/tcp",
"state": "enabled"
},
"skip_reason": "Conditional result was False"
}
skipping: [sut] => {
"changed": false
}
MSG:
All items skipped
TASK [fedora.linux_system_roles.firewall : Update firewalld_config fact] *******
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:130
Wednesday 10 July 2024 05:13:53 +0000 (0:00:00.028) 0:01:58.092 ********
skipping: [sut] => {
"changed": false,
"false_condition": "firewall | length == 1",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.firewall : Gather firewall config if no arguments] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:139
Wednesday 10 July 2024 05:13:53 +0000 (0:00:00.019) 0:01:58.111 ********
skipping: [sut] => {
"changed": false,
"false_condition": "firewall == None or firewall | length == 0",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.firewall : Update firewalld_config fact] *******
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:144
Wednesday 10 July 2024 05:13:53 +0000 (0:00:00.018) 0:01:58.129 ********
skipping: [sut] => {
"changed": false,
"false_condition": "firewall == None or firewall | length == 0",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.firewall : Get config files, checksums after] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:153
Wednesday 10 July 2024 05:13:53 +0000 (0:00:00.017) 0:01:58.147 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__firewall_previous_replaced | bool",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.firewall : Calculate what has changed] *********
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:163
Wednesday 10 July 2024 05:13:53 +0000 (0:00:00.040) 0:01:58.187 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__firewall_previous_replaced | bool",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.firewall : Show diffs] *************************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:169
Wednesday 10 July 2024 05:13:53 +0000 (0:00:00.017) 0:01:58.205 ********
skipping: [sut] => {
"false_condition": "__firewall_previous_replaced | bool"
}
TASK [Manage selinux for specified ports] **************************************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:93
Wednesday 10 July 2024 05:13:53 +0000 (0:00:00.026) 0:01:58.231 ********
skipping: [sut] => {
"changed": false,
"false_condition": "podman_selinux_ports | length > 0",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Keep track of users that need to cancel linger] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:100
Wednesday 10 July 2024 05:13:53 +0000 (0:00:00.017) 0:01:58.249 ********
ok: [sut] => {
"ansible_facts": {
"__podman_cancel_user_linger": []
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Handle certs.d files - present] *******
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:104
Wednesday 10 July 2024 05:13:53 +0000 (0:00:00.018) 0:01:58.267 ********
skipping: [sut] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : Handle credential files - present] ****
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:113
Wednesday 10 July 2024 05:13:53 +0000 (0:00:00.014) 0:01:58.282 ********
skipping: [sut] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : Handle secrets] ***********************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:122
Wednesday 10 July 2024 05:13:53 +0000 (0:00:00.014) 0:01:58.297 ********
included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml for sut => (item=(censored due to no_log))
included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml for sut => (item=(censored due to no_log))
included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml for sut => (item=(censored due to no_log))
TASK [fedora.linux_system_roles.podman : Set variables part 1] *****************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:3
Wednesday 10 July 2024 05:13:53 +0000 (0:00:00.068) 0:01:58.365 ********
ok: [sut] => {
"ansible_facts": {
"__podman_user": "root"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set variables part 2] *****************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:7
Wednesday 10 July 2024 05:13:53 +0000 (0:00:00.020) 0:01:58.386 ********
ok: [sut] => {
"ansible_facts": {
"__podman_rootless": false,
"__podman_xdg_runtime_dir": "/run/user/0"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Manage linger] ************************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:13
Wednesday 10 July 2024 05:13:53 +0000 (0:00:00.022) 0:01:58.409 ********
included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for sut
TASK [fedora.linux_system_roles.podman : Enable linger if needed] **************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12
Wednesday 10 July 2024 05:13:53 +0000 (0:00:00.027) 0:01:58.436 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_rootless | bool",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18
Wednesday 10 July 2024 05:13:53 +0000 (0:00:00.016) 0:01:58.453 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_rootless | bool",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22
Wednesday 10 July 2024 05:13:53 +0000 (0:00:00.016) 0:01:58.469 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_rootless | bool",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Stat XDG_RUNTIME_DIR] *****************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:18
Wednesday 10 July 2024 05:13:53 +0000 (0:00:00.015) 0:01:58.485 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_rootless | bool",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Manage each secret] *******************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:34
Wednesday 10 July 2024 05:13:53 +0000 (0:00:00.015) 0:01:58.501 ********
changed: [sut] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": true
}
TASK [fedora.linux_system_roles.podman : Set variables part 1] *****************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:3
Wednesday 10 July 2024 05:13:54 +0000 (0:00:00.582) 0:01:59.083 ********
ok: [sut] => {
"ansible_facts": {
"__podman_user": "root"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set variables part 2] *****************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:7
Wednesday 10 July 2024 05:13:54 +0000 (0:00:00.021) 0:01:59.104 ********
ok: [sut] => {
"ansible_facts": {
"__podman_rootless": false,
"__podman_xdg_runtime_dir": "/run/user/0"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Manage linger] ************************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:13
Wednesday 10 July 2024 05:13:54 +0000 (0:00:00.024) 0:01:59.129 ********
included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for sut
TASK [fedora.linux_system_roles.podman : Enable linger if needed] **************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12
Wednesday 10 July 2024 05:13:54 +0000 (0:00:00.028) 0:01:59.157 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_rootless | bool",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18
Wednesday 10 July 2024 05:13:54 +0000 (0:00:00.017) 0:01:59.175 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_rootless | bool",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22
Wednesday 10 July 2024 05:13:54 +0000 (0:00:00.017) 0:01:59.192 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_rootless | bool",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Stat XDG_RUNTIME_DIR] *****************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:18
Wednesday 10 July 2024 05:13:54 +0000 (0:00:00.017) 0:01:59.209 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_rootless | bool",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Manage each secret] *******************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:34
Wednesday 10 July 2024 05:13:54 +0000 (0:00:00.016) 0:01:59.225 ********
changed: [sut] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": true
}
TASK [fedora.linux_system_roles.podman : Set variables part 1] *****************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:3
Wednesday 10 July 2024 05:13:54 +0000 (0:00:00.369) 0:01:59.595 ********
ok: [sut] => {
"ansible_facts": {
"__podman_user": "root"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set variables part 2] *****************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:7
Wednesday 10 July 2024 05:13:54 +0000 (0:00:00.020) 0:01:59.616 ********
ok: [sut] => {
"ansible_facts": {
"__podman_rootless": false,
"__podman_xdg_runtime_dir": "/run/user/0"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Manage linger] ************************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:13
Wednesday 10 July 2024 05:13:54 +0000 (0:00:00.024) 0:01:59.641 ********
included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for sut
TASK [fedora.linux_system_roles.podman : Enable linger if needed] **************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12
Wednesday 10 July 2024 05:13:55 +0000 (0:00:00.028) 0:01:59.669 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_rootless | bool",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18
Wednesday 10 July 2024 05:13:55 +0000 (0:00:00.017) 0:01:59.687 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_rootless | bool",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22
Wednesday 10 July 2024 05:13:55 +0000 (0:00:00.016) 0:01:59.703 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_rootless | bool",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Stat XDG_RUNTIME_DIR] *****************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:18
Wednesday 10 July 2024 05:13:55 +0000 (0:00:00.016) 0:01:59.720 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_rootless | bool",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Manage each secret] *******************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:34
Wednesday 10 July 2024 05:13:55 +0000 (0:00:00.016) 0:01:59.736 ********
changed: [sut] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": true
}
TASK [fedora.linux_system_roles.podman : Handle Kubernetes specifications] *****
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:129
Wednesday 10 July 2024 05:13:55 +0000 (0:00:00.379) 0:02:00.116 ********
skipping: [sut] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : Handle Quadlet specifications] ********
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:136
Wednesday 10 July 2024 05:13:55 +0000 (0:00:00.042) 0:02:00.159 ********
included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for sut => (item=(censored due to no_log))
included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for sut => (item=(censored due to no_log))
included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for sut => (item=(censored due to no_log))
included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for sut => (item=(censored due to no_log))
included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for sut => (item=(censored due to no_log))
included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for sut => (item=(censored due to no_log))
TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14
Wednesday 10 July 2024 05:13:55 +0000 (0:00:00.084) 0:02:00.243 ********
ok: [sut] => {
"ansible_facts": {
"__podman_quadlet_file_src": "quadlet-demo.network",
"__podman_quadlet_spec": {},
"__podman_quadlet_str": "[Network]\nSubnet=192.168.30.0/24\nGateway=192.168.30.1\nLabel=app=wordpress",
"__podman_quadlet_template_src": ""
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25
Wednesday 10 July 2024 05:13:55 +0000 (0:00:00.027) 0:02:00.271 ********
ok: [sut] => {
"ansible_facts": {
"__podman_continue_if_pull_fails": false,
"__podman_pull_image": true,
"__podman_state": "created",
"__podman_systemd_unit_scope": "",
"__podman_user": "root"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] *****
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35
Wednesday 10 July 2024 05:13:55 +0000 (0:00:00.023) 0:02:00.294 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_quadlet_file_src",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48
Wednesday 10 July 2024 05:13:55 +0000 (0:00:00.018) 0:02:00.313 ********
ok: [sut] => {
"ansible_facts": {
"__podman_quadlet_name": "quadlet-demo",
"__podman_quadlet_type": "network",
"__podman_rootless": false
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Check user and group information] *****
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57
Wednesday 10 July 2024 05:13:55 +0000 (0:00:00.028) 0:02:00.342 ********
included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for sut
TASK [fedora.linux_system_roles.podman : Get user information] *****************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2
Wednesday 10 July 2024 05:13:55 +0000 (0:00:00.034) 0:02:00.376 ********
skipping: [sut] => {
"changed": false,
"false_condition": "'getent_passwd' not in ansible_facts or __podman_user not in ansible_facts['getent_passwd']",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Fail if user does not exist] **********
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9
Wednesday 10 July 2024 05:13:55 +0000 (0:00:00.021) 0:02:00.397 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not ansible_facts[\"getent_passwd\"][__podman_user]",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set group for podman user] ************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16
Wednesday 10 July 2024 05:13:55 +0000 (0:00:00.020) 0:02:00.418 ********
ok: [sut] => {
"ansible_facts": {
"__podman_group": "0"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Get group information] ****************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:28
Wednesday 10 July 2024 05:13:55 +0000 (0:00:00.027) 0:02:00.446 ********
ok: [sut] => {
"ansible_facts": {
"getent_group": {
"root": [
"x",
"0",
""
]
}
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set group name] ***********************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:35
Wednesday 10 July 2024 05:13:56 +0000 (0:00:00.239) 0:02:00.685 ********
ok: [sut] => {
"ansible_facts": {
"__podman_group_name": "root"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : See if getsubids exists] **************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:39
Wednesday 10 July 2024 05:13:56 +0000 (0:00:00.024) 0:02:00.710 ********
ok: [sut] => {
"changed": false,
"stat": {
"atime": 1720588427.5164855,
"attr_flags": "e",
"attributes": [
"extents"
],
"block_size": 4096,
"blocks": 32,
"charset": "binary",
"checksum": "72ba28a3be19c467e89ab90f5d21f99b605159a7",
"ctime": 1720588377.3007717,
"dev": 51714,
"device_type": 0,
"executable": true,
"exists": true,
"gid": 0,
"gr_name": "root",
"inode": 203603,
"isblk": false,
"ischr": false,
"isdir": false,
"isfifo": false,
"isgid": false,
"islnk": false,
"isreg": true,
"issock": false,
"isuid": false,
"mimetype": "application/x-pie-executable",
"mode": "0755",
"mtime": 1696291200.0,
"nlink": 1,
"path": "/usr/bin/getsubids",
"pw_name": "root",
"readable": true,
"rgrp": true,
"roth": true,
"rusr": true,
"size": 16032,
"uid": 0,
"version": "2092964490",
"wgrp": false,
"woth": false,
"writeable": true,
"wusr": true,
"xgrp": true,
"xoth": true,
"xusr": true
}
}
TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:50
Wednesday 10 July 2024 05:13:56 +0000 (0:00:00.235) 0:02:00.945 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_user not in [\"root\", \"0\"]",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Check group with getsubids] ***********
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:55
Wednesday 10 July 2024 05:13:56 +0000 (0:00:00.017) 0:02:00.963 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_user not in [\"root\", \"0\"]",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ******
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:60
Wednesday 10 July 2024 05:13:56 +0000 (0:00:00.017) 0:02:00.980 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_user not in [\"root\", \"0\"]",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Get subuid file] **********************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:74
Wednesday 10 July 2024 05:13:56 +0000 (0:00:00.016) 0:02:00.997 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_stat_getsubids.stat.exists",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Get subgid file] **********************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:79
Wednesday 10 July 2024 05:13:56 +0000 (0:00:00.041) 0:02:01.038 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_stat_getsubids.stat.exists",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ******
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:84
Wednesday 10 July 2024 05:13:56 +0000 (0:00:00.018) 0:02:01.056 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_stat_getsubids.stat.exists",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ******
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:94
Wednesday 10 July 2024 05:13:56 +0000 (0:00:00.016) 0:02:01.073 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_stat_getsubids.stat.exists",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] *****
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:101
Wednesday 10 July 2024 05:13:56 +0000 (0:00:00.017) 0:02:01.091 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_stat_getsubids.stat.exists",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62
Wednesday 10 July 2024 05:13:56 +0000 (0:00:00.016) 0:02:01.107 ********
ok: [sut] => {
"ansible_facts": {
"__podman_activate_systemd_unit": true,
"__podman_images_found": [],
"__podman_kube_yamls_raw": "",
"__podman_service_name": "quadlet-demo-network.service",
"__podman_systemd_scope": "system",
"__podman_user_home_dir": "/root",
"__podman_xdg_runtime_dir": "/run/user/0"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73
Wednesday 10 July 2024 05:13:56 +0000 (0:00:00.033) 0:02:01.141 ********
ok: [sut] => {
"ansible_facts": {
"__podman_quadlet_path": "/etc/containers/systemd"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Get kube yaml contents] ***************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77
Wednesday 10 July 2024 05:13:56 +0000 (0:00:00.019) 0:02:01.161 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_kube_yamls_raw | length > 0",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87
Wednesday 10 July 2024 05:13:56 +0000 (0:00:00.018) 0:02:01.179 ********
ok: [sut] => {
"ansible_facts": {
"__podman_images": [],
"__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo.network",
"__podman_volumes": []
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 6] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:103
Wednesday 10 July 2024 05:13:56 +0000 (0:00:00.044) 0:02:01.224 ********
ok: [sut] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : Cleanup quadlets] *********************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:110
Wednesday 10 July 2024 05:13:56 +0000 (0:00:00.022) 0:02:01.246 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_state == \"absent\"",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Create and update quadlets] ***********
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:114
Wednesday 10 July 2024 05:13:56 +0000 (0:00:00.017) 0:02:01.263 ********
included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml for sut
TASK [fedora.linux_system_roles.podman : Manage linger] ************************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:2
Wednesday 10 July 2024 05:13:56 +0000 (0:00:00.038) 0:02:01.302 ********
included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for sut
TASK [fedora.linux_system_roles.podman : Enable linger if needed] **************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12
Wednesday 10 July 2024 05:13:56 +0000 (0:00:00.030) 0:02:01.332 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_rootless | bool",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18
Wednesday 10 July 2024 05:13:56 +0000 (0:00:00.017) 0:02:01.350 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_rootless | bool",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22
Wednesday 10 July 2024 05:13:56 +0000 (0:00:00.017) 0:02:01.367 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_rootless | bool",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Create host directories] **************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:7
Wednesday 10 July 2024 05:13:56 +0000 (0:00:00.016) 0:02:01.383 ********
skipping: [sut] => {
"changed": false,
"skipped_reason": "No items in the list"
}
TASK [fedora.linux_system_roles.podman : Ensure container images are present] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:18
Wednesday 10 July 2024 05:13:56 +0000 (0:00:00.042) 0:02:01.425 ********
skipping: [sut] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : Ensure the quadlet directory is present] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:39
Wednesday 10 July 2024 05:13:56 +0000 (0:00:00.018) 0:02:01.443 ********
ok: [sut] => {
"changed": false,
"gid": 0,
"group": "root",
"mode": "0755",
"owner": "root",
"path": "/etc/containers/systemd",
"secontext": "system_u:object_r:etc_t:s0",
"size": 4096,
"state": "directory",
"uid": 0
}
TASK [fedora.linux_system_roles.podman : Ensure quadlet file is copied] ********
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:48
Wednesday 10 July 2024 05:13:57 +0000 (0:00:00.250) 0:02:01.694 ********
changed: [sut] => {
"changed": true,
"checksum": "e57c08d49aff4bae8daab138d913aeddaa8682a0",
"dest": "/etc/containers/systemd/quadlet-demo.network",
"gid": 0,
"group": "root",
"md5sum": "061f3cf318cbd8ab5794bb1173831fb8",
"mode": "0644",
"owner": "root",
"secontext": "system_u:object_r:etc_t:s0",
"size": 74,
"src": "/root/.ansible/tmp/ansible-tmp-1720588437.067682-11949-230791674364504/source",
"state": "file",
"uid": 0
}
TASK [fedora.linux_system_roles.podman : Ensure quadlet file content is present] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:58
Wednesday 10 July 2024 05:13:57 +0000 (0:00:00.652) 0:02:02.347 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_quadlet_file_src",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Ensure quadlet file is present] *******
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:70
Wednesday 10 July 2024 05:13:57 +0000 (0:00:00.020) 0:02:02.368 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_copy_file is skipped",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Reload systemctl] *********************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:82
Wednesday 10 July 2024 05:13:57 +0000 (0:00:00.017) 0:02:02.386 ********
ok: [sut] => {
"changed": false,
"name": null,
"status": {}
}
TASK [fedora.linux_system_roles.podman : Start service] ************************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:110
Wednesday 10 July 2024 05:13:58 +0000 (0:00:00.798) 0:02:03.184 ********
changed: [sut] => {
"changed": true,
"name": "quadlet-demo-network.service",
"state": "started",
"status": {
"AccessSELinuxContext": "system_u:object_r:systemd_unit_file_t:s0",
"ActiveEnterTimestampMonotonic": "0",
"ActiveExitTimestampMonotonic": "0",
"ActiveState": "inactive",
"After": "-.mount systemd-journald.socket basic.target system.slice sysinit.target",
"AllowIsolate": "no",
"AssertResult": "no",
"AssertTimestampMonotonic": "0",
"Before": "shutdown.target",
"BlockIOAccounting": "no",
"BlockIOWeight": "[not set]",
"CPUAccounting": "yes",
"CPUAffinityFromNUMA": "no",
"CPUQuotaPerSecUSec": "infinity",
"CPUQuotaPeriodUSec": "infinity",
"CPUSchedulingPolicy": "0",
"CPUSchedulingPriority": "0",
"CPUSchedulingResetOnFork": "no",
"CPUShares": "[not set]",
"CPUUsageNSec": "[not set]",
"CPUWeight": "[not set]",
"CacheDirectoryMode": "0755",
"CanFreeze": "yes",
"CanIsolate": "no",
"CanReload": "no",
"CanStart": "yes",
"CanStop": "yes",
"CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore",
"CleanResult": "success",
"CollectMode": "inactive",
"ConditionResult": "no",
"ConditionTimestampMonotonic": "0",
"ConfigurationDirectoryMode": "0755",
"Conflicts": "shutdown.target",
"ControlGroupId": "0",
"ControlPID": "0",
"CoredumpFilter": "0x33",
"DefaultDependencies": "yes",
"DefaultMemoryLow": "0",
"DefaultMemoryMin": "0",
"DefaultStartupMemoryLow": "0",
"Delegate": "no",
"Description": "quadlet-demo-network.service",
"DevicePolicy": "auto",
"DropInPaths": "/usr/lib/systemd/system/service.d/10-timeout-abort.conf",
"DynamicUser": "no",
"ExecMainCode": "0",
"ExecMainExitTimestampMonotonic": "0",
"ExecMainPID": "0",
"ExecMainStartTimestampMonotonic": "0",
"ExecMainStatus": "0",
"ExecStart": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman network create --ignore --subnet=192.168.30.0/24 --gateway=192.168.30.1 --label app=wordpress systemd-quadlet-demo ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"ExecStartEx": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman network create --ignore --subnet=192.168.30.0/24 --gateway=192.168.30.1 --label app=wordpress systemd-quadlet-demo ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"ExitType": "main",
"ExtensionImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent",
"FailureAction": "none",
"FileDescriptorStoreMax": "0",
"FileDescriptorStorePreserve": "restart",
"FinalKillSignal": "9",
"FragmentPath": "/run/systemd/generator/quadlet-demo-network.service",
"FreezerState": "running",
"GID": "[not set]",
"GuessMainPID": "yes",
"IOAccounting": "no",
"IOReadBytes": "18446744073709551615",
"IOReadOperations": "18446744073709551615",
"IOSchedulingClass": "2",
"IOSchedulingPriority": "4",
"IOWeight": "[not set]",
"IOWriteBytes": "18446744073709551615",
"IOWriteOperations": "18446744073709551615",
"IPAccounting": "no",
"IPEgressBytes": "[no data]",
"IPEgressPackets": "[no data]",
"IPIngressBytes": "[no data]",
"IPIngressPackets": "[no data]",
"Id": "quadlet-demo-network.service",
"IgnoreOnIsolate": "no",
"IgnoreSIGPIPE": "yes",
"InactiveEnterTimestampMonotonic": "0",
"InactiveExitTimestampMonotonic": "0",
"JobRunningTimeoutUSec": "infinity",
"JobTimeoutAction": "none",
"JobTimeoutUSec": "infinity",
"KeyringMode": "private",
"KillMode": "control-group",
"KillSignal": "15",
"LimitAS": "infinity",
"LimitASSoft": "infinity",
"LimitCORE": "infinity",
"LimitCORESoft": "infinity",
"LimitCPU": "infinity",
"LimitCPUSoft": "infinity",
"LimitDATA": "infinity",
"LimitDATASoft": "infinity",
"LimitFSIZE": "infinity",
"LimitFSIZESoft": "infinity",
"LimitLOCKS": "infinity",
"LimitLOCKSSoft": "infinity",
"LimitMEMLOCK": "8388608",
"LimitMEMLOCKSoft": "8388608",
"LimitMSGQUEUE": "819200",
"LimitMSGQUEUESoft": "819200",
"LimitNICE": "0",
"LimitNICESoft": "0",
"LimitNOFILE": "524288",
"LimitNOFILESoft": "1024",
"LimitNPROC": "14756",
"LimitNPROCSoft": "14756",
"LimitRSS": "infinity",
"LimitRSSSoft": "infinity",
"LimitRTPRIO": "0",
"LimitRTPRIOSoft": "0",
"LimitRTTIME": "infinity",
"LimitRTTIMESoft": "infinity",
"LimitSIGPENDING": "14756",
"LimitSIGPENDINGSoft": "14756",
"LimitSTACK": "infinity",
"LimitSTACKSoft": "8388608",
"LoadState": "loaded",
"LockPersonality": "no",
"LogLevelMax": "-1",
"LogRateLimitBurst": "0",
"LogRateLimitIntervalUSec": "0",
"LogsDirectoryMode": "0755",
"MainPID": "0",
"ManagedOOMMemoryPressure": "auto",
"ManagedOOMMemoryPressureLimit": "0",
"ManagedOOMPreference": "none",
"ManagedOOMSwap": "auto",
"MemoryAccounting": "yes",
"MemoryAvailable": "infinity",
"MemoryCurrent": "[not set]",
"MemoryDenyWriteExecute": "no",
"MemoryHigh": "infinity",
"MemoryKSM": "no",
"MemoryLimit": "infinity",
"MemoryLow": "0",
"MemoryMax": "infinity",
"MemoryMin": "0",
"MemoryPressureThresholdUSec": "200ms",
"MemoryPressureWatch": "auto",
"MemorySwapMax": "infinity",
"MemoryZSwapMax": "infinity",
"MountAPIVFS": "no",
"MountImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent",
"NFileDescriptorStore": "0",
"NRestarts": "0",
"NUMAPolicy": "n/a",
"Names": "quadlet-demo-network.service",
"NeedDaemonReload": "no",
"Nice": "0",
"NoNewPrivileges": "no",
"NonBlocking": "no",
"NotifyAccess": "none",
"OOMPolicy": "stop",
"OOMScoreAdjust": "0",
"OnFailureJobMode": "replace",
"OnSuccessJobMode": "fail",
"Perpetual": "no",
"PrivateDevices": "no",
"PrivateIPC": "no",
"PrivateMounts": "no",
"PrivateNetwork": "no",
"PrivateTmp": "no",
"PrivateUsers": "no",
"ProcSubset": "all",
"ProtectClock": "no",
"ProtectControlGroups": "no",
"ProtectHome": "no",
"ProtectHostname": "no",
"ProtectKernelLogs": "no",
"ProtectKernelModules": "no",
"ProtectKernelTunables": "no",
"ProtectProc": "default",
"ProtectSystem": "no",
"RefuseManualStart": "no",
"RefuseManualStop": "no",
"ReloadResult": "success",
"ReloadSignal": "1",
"RemainAfterExit": "yes",
"RemoveIPC": "no",
"Requires": "-.mount system.slice sysinit.target",
"RequiresMountsFor": "/run/containers",
"Restart": "no",
"RestartKillSignal": "15",
"RestartMaxDelayUSec": "infinity",
"RestartMode": "normal",
"RestartSteps": "0",
"RestartUSec": "100ms",
"RestartUSecNext": "100ms",
"RestrictNamespaces": "no",
"RestrictRealtime": "no",
"RestrictSUIDSGID": "no",
"Result": "success",
"RootDirectoryStartOnly": "no",
"RootEphemeral": "no",
"RootImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent",
"RuntimeDirectoryMode": "0755",
"RuntimeDirectoryPreserve": "no",
"RuntimeMaxUSec": "infinity",
"RuntimeRandomizedExtraUSec": "0",
"SameProcessGroup": "no",
"SecureBits": "0",
"SendSIGHUP": "no",
"SendSIGKILL": "yes",
"Slice": "system.slice",
"StandardError": "inherit",
"StandardInput": "null",
"StandardOutput": "journal",
"StartLimitAction": "none",
"StartLimitBurst": "5",
"StartLimitIntervalUSec": "10s",
"StartupBlockIOWeight": "[not set]",
"StartupCPUShares": "[not set]",
"StartupCPUWeight": "[not set]",
"StartupIOWeight": "[not set]",
"StartupMemoryHigh": "18446744073709551615",
"StartupMemoryLow": "0",
"StartupMemoryMax": "18446744073709551615",
"StartupMemorySwapMax": "18446744073709551615",
"StartupMemoryZSwapMax": "18446744073709551615",
"StateChangeTimestampMonotonic": "0",
"StateDirectoryMode": "0755",
"StatusErrno": "0",
"StopWhenUnneeded": "no",
"SubState": "dead",
"SuccessAction": "none",
"SyslogFacility": "3",
"SyslogIdentifier": "quadlet-demo-network",
"SyslogLevel": "6",
"SyslogLevelPrefix": "yes",
"SyslogPriority": "30",
"SystemCallErrorNumber": "2147483646",
"TTYReset": "no",
"TTYVHangup": "no",
"TTYVTDisallocate": "no",
"TasksAccounting": "yes",
"TasksCurrent": "[not set]",
"TasksMax": "4426",
"TimeoutAbortUSec": "45s",
"TimeoutCleanUSec": "infinity",
"TimeoutStartFailureMode": "terminate",
"TimeoutStartUSec": "infinity",
"TimeoutStopFailureMode": "abort",
"TimeoutStopUSec": "45s",
"TimerSlackNSec": "50000",
"Transient": "no",
"Type": "oneshot",
"UID": "[not set]",
"UMask": "0022",
"UnitFilePreset": "disabled",
"UnitFileState": "generated",
"UtmpMode": "init",
"WatchdogSignal": "6",
"WatchdogTimestampMonotonic": "0",
"WatchdogUSec": "infinity"
}
}
TASK [fedora.linux_system_roles.podman : Restart service] **********************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:125
Wednesday 10 July 2024 05:13:58 +0000 (0:00:00.464) 0:02:03.648 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_service_started is changed",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14
Wednesday 10 July 2024 05:13:59 +0000 (0:00:00.018) 0:02:03.667 ********
ok: [sut] => {
"ansible_facts": {
"__podman_quadlet_file_src": "quadlet-demo-mysql.volume",
"__podman_quadlet_spec": {},
"__podman_quadlet_str": "[Volume]",
"__podman_quadlet_template_src": ""
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25
Wednesday 10 July 2024 05:13:59 +0000 (0:00:00.027) 0:02:03.695 ********
ok: [sut] => {
"ansible_facts": {
"__podman_continue_if_pull_fails": false,
"__podman_pull_image": true,
"__podman_state": "created",
"__podman_systemd_unit_scope": "",
"__podman_user": "root"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] *****
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35
Wednesday 10 July 2024 05:13:59 +0000 (0:00:00.024) 0:02:03.719 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_quadlet_file_src",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48
Wednesday 10 July 2024 05:13:59 +0000 (0:00:00.017) 0:02:03.737 ********
ok: [sut] => {
"ansible_facts": {
"__podman_quadlet_name": "quadlet-demo-mysql",
"__podman_quadlet_type": "volume",
"__podman_rootless": false
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Check user and group information] *****
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57
Wednesday 10 July 2024 05:13:59 +0000 (0:00:00.028) 0:02:03.765 ********
included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for sut
TASK [fedora.linux_system_roles.podman : Get user information] *****************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2
Wednesday 10 July 2024 05:13:59 +0000 (0:00:00.034) 0:02:03.800 ********
skipping: [sut] => {
"changed": false,
"false_condition": "'getent_passwd' not in ansible_facts or __podman_user not in ansible_facts['getent_passwd']",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Fail if user does not exist] **********
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9
Wednesday 10 July 2024 05:13:59 +0000 (0:00:00.020) 0:02:03.821 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not ansible_facts[\"getent_passwd\"][__podman_user]",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set group for podman user] ************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16
Wednesday 10 July 2024 05:13:59 +0000 (0:00:00.019) 0:02:03.841 ********
ok: [sut] => {
"ansible_facts": {
"__podman_group": "0"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Get group information] ****************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:28
Wednesday 10 July 2024 05:13:59 +0000 (0:00:00.026) 0:02:03.867 ********
ok: [sut] => {
"ansible_facts": {
"getent_group": {
"root": [
"x",
"0",
""
]
}
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set group name] ***********************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:35
Wednesday 10 July 2024 05:13:59 +0000 (0:00:00.237) 0:02:04.105 ********
ok: [sut] => {
"ansible_facts": {
"__podman_group_name": "root"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : See if getsubids exists] **************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:39
Wednesday 10 July 2024 05:13:59 +0000 (0:00:00.054) 0:02:04.159 ********
ok: [sut] => {
"changed": false,
"stat": {
"atime": 1720588427.5164855,
"attr_flags": "e",
"attributes": [
"extents"
],
"block_size": 4096,
"blocks": 32,
"charset": "binary",
"checksum": "72ba28a3be19c467e89ab90f5d21f99b605159a7",
"ctime": 1720588377.3007717,
"dev": 51714,
"device_type": 0,
"executable": true,
"exists": true,
"gid": 0,
"gr_name": "root",
"inode": 203603,
"isblk": false,
"ischr": false,
"isdir": false,
"isfifo": false,
"isgid": false,
"islnk": false,
"isreg": true,
"issock": false,
"isuid": false,
"mimetype": "application/x-pie-executable",
"mode": "0755",
"mtime": 1696291200.0,
"nlink": 1,
"path": "/usr/bin/getsubids",
"pw_name": "root",
"readable": true,
"rgrp": true,
"roth": true,
"rusr": true,
"size": 16032,
"uid": 0,
"version": "2092964490",
"wgrp": false,
"woth": false,
"writeable": true,
"wusr": true,
"xgrp": true,
"xoth": true,
"xusr": true
}
}
TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:50
Wednesday 10 July 2024 05:13:59 +0000 (0:00:00.237) 0:02:04.397 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_user not in [\"root\", \"0\"]",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Check group with getsubids] ***********
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:55
Wednesday 10 July 2024 05:13:59 +0000 (0:00:00.017) 0:02:04.415 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_user not in [\"root\", \"0\"]",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ******
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:60
Wednesday 10 July 2024 05:13:59 +0000 (0:00:00.017) 0:02:04.433 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_user not in [\"root\", \"0\"]",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Get subuid file] **********************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:74
Wednesday 10 July 2024 05:13:59 +0000 (0:00:00.016) 0:02:04.450 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_stat_getsubids.stat.exists",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Get subgid file] **********************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:79
Wednesday 10 July 2024 05:13:59 +0000 (0:00:00.017) 0:02:04.467 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_stat_getsubids.stat.exists",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ******
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:84
Wednesday 10 July 2024 05:13:59 +0000 (0:00:00.016) 0:02:04.484 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_stat_getsubids.stat.exists",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ******
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:94
Wednesday 10 July 2024 05:13:59 +0000 (0:00:00.017) 0:02:04.501 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_stat_getsubids.stat.exists",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] *****
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:101
Wednesday 10 July 2024 05:13:59 +0000 (0:00:00.017) 0:02:04.518 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_stat_getsubids.stat.exists",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62
Wednesday 10 July 2024 05:13:59 +0000 (0:00:00.017) 0:02:04.535 ********
ok: [sut] => {
"ansible_facts": {
"__podman_activate_systemd_unit": true,
"__podman_images_found": [],
"__podman_kube_yamls_raw": "",
"__podman_service_name": "quadlet-demo-mysql-volume.service",
"__podman_systemd_scope": "system",
"__podman_user_home_dir": "/root",
"__podman_xdg_runtime_dir": "/run/user/0"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73
Wednesday 10 July 2024 05:13:59 +0000 (0:00:00.033) 0:02:04.569 ********
ok: [sut] => {
"ansible_facts": {
"__podman_quadlet_path": "/etc/containers/systemd"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Get kube yaml contents] ***************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77
Wednesday 10 July 2024 05:13:59 +0000 (0:00:00.019) 0:02:04.588 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_kube_yamls_raw | length > 0",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87
Wednesday 10 July 2024 05:13:59 +0000 (0:00:00.018) 0:02:04.606 ********
ok: [sut] => {
"ansible_facts": {
"__podman_images": [],
"__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo-mysql.volume",
"__podman_volumes": []
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 6] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:103
Wednesday 10 July 2024 05:13:59 +0000 (0:00:00.045) 0:02:04.652 ********
ok: [sut] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : Cleanup quadlets] *********************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:110
Wednesday 10 July 2024 05:14:00 +0000 (0:00:00.023) 0:02:04.675 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_state == \"absent\"",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Create and update quadlets] ***********
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:114
Wednesday 10 July 2024 05:14:00 +0000 (0:00:00.017) 0:02:04.693 ********
included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml for sut
TASK [fedora.linux_system_roles.podman : Manage linger] ************************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:2
Wednesday 10 July 2024 05:14:00 +0000 (0:00:00.037) 0:02:04.731 ********
included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for sut
TASK [fedora.linux_system_roles.podman : Enable linger if needed] **************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12
Wednesday 10 July 2024 05:14:00 +0000 (0:00:00.056) 0:02:04.787 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_rootless | bool",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18
Wednesday 10 July 2024 05:14:00 +0000 (0:00:00.019) 0:02:04.806 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_rootless | bool",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22
Wednesday 10 July 2024 05:14:00 +0000 (0:00:00.017) 0:02:04.823 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_rootless | bool",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Create host directories] **************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:7
Wednesday 10 July 2024 05:14:00 +0000 (0:00:00.017) 0:02:04.841 ********
skipping: [sut] => {
"changed": false,
"skipped_reason": "No items in the list"
}
TASK [fedora.linux_system_roles.podman : Ensure container images are present] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:18
Wednesday 10 July 2024 05:14:00 +0000 (0:00:00.016) 0:02:04.858 ********
skipping: [sut] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : Ensure the quadlet directory is present] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:39
Wednesday 10 July 2024 05:14:00 +0000 (0:00:00.017) 0:02:04.875 ********
ok: [sut] => {
"changed": false,
"gid": 0,
"group": "root",
"mode": "0755",
"owner": "root",
"path": "/etc/containers/systemd",
"secontext": "system_u:object_r:etc_t:s0",
"size": 4096,
"state": "directory",
"uid": 0
}
TASK [fedora.linux_system_roles.podman : Ensure quadlet file is copied] ********
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:48
Wednesday 10 July 2024 05:14:00 +0000 (0:00:00.249) 0:02:05.125 ********
changed: [sut] => {
"changed": true,
"checksum": "585f8cbdf0ec73000f9227dcffbef71e9552ea4a",
"dest": "/etc/containers/systemd/quadlet-demo-mysql.volume",
"gid": 0,
"group": "root",
"md5sum": "5ddd03a022aeb4502d9bc8ce436b4233",
"mode": "0644",
"owner": "root",
"secontext": "system_u:object_r:etc_t:s0",
"size": 9,
"src": "/root/.ansible/tmp/ansible-tmp-1720588440.498636-12001-192902405762867/source",
"state": "file",
"uid": 0
}
TASK [fedora.linux_system_roles.podman : Ensure quadlet file content is present] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:58
Wednesday 10 July 2024 05:14:01 +0000 (0:00:00.588) 0:02:05.713 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_quadlet_file_src",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Ensure quadlet file is present] *******
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:70
Wednesday 10 July 2024 05:14:01 +0000 (0:00:00.020) 0:02:05.734 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_copy_file is skipped",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Reload systemctl] *********************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:82
Wednesday 10 July 2024 05:14:01 +0000 (0:00:00.020) 0:02:05.754 ********
ok: [sut] => {
"changed": false,
"name": null,
"status": {}
}
TASK [fedora.linux_system_roles.podman : Start service] ************************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:110
Wednesday 10 July 2024 05:14:01 +0000 (0:00:00.797) 0:02:06.552 ********
changed: [sut] => {
"changed": true,
"name": "quadlet-demo-mysql-volume.service",
"state": "started",
"status": {
"AccessSELinuxContext": "system_u:object_r:systemd_unit_file_t:s0",
"ActiveEnterTimestampMonotonic": "0",
"ActiveExitTimestampMonotonic": "0",
"ActiveState": "inactive",
"After": "-.mount system.slice systemd-journald.socket sysinit.target basic.target",
"AllowIsolate": "no",
"AssertResult": "no",
"AssertTimestampMonotonic": "0",
"Before": "shutdown.target",
"BlockIOAccounting": "no",
"BlockIOWeight": "[not set]",
"CPUAccounting": "yes",
"CPUAffinityFromNUMA": "no",
"CPUQuotaPerSecUSec": "infinity",
"CPUQuotaPeriodUSec": "infinity",
"CPUSchedulingPolicy": "0",
"CPUSchedulingPriority": "0",
"CPUSchedulingResetOnFork": "no",
"CPUShares": "[not set]",
"CPUUsageNSec": "[not set]",
"CPUWeight": "[not set]",
"CacheDirectoryMode": "0755",
"CanFreeze": "yes",
"CanIsolate": "no",
"CanReload": "no",
"CanStart": "yes",
"CanStop": "yes",
"CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore",
"CleanResult": "success",
"CollectMode": "inactive",
"ConditionResult": "no",
"ConditionTimestampMonotonic": "0",
"ConfigurationDirectoryMode": "0755",
"Conflicts": "shutdown.target",
"ControlGroupId": "0",
"ControlPID": "0",
"CoredumpFilter": "0x33",
"DefaultDependencies": "yes",
"DefaultMemoryLow": "0",
"DefaultMemoryMin": "0",
"DefaultStartupMemoryLow": "0",
"Delegate": "no",
"Description": "quadlet-demo-mysql-volume.service",
"DevicePolicy": "auto",
"DropInPaths": "/usr/lib/systemd/system/service.d/10-timeout-abort.conf",
"DynamicUser": "no",
"ExecMainCode": "0",
"ExecMainExitTimestampMonotonic": "0",
"ExecMainPID": "0",
"ExecMainStartTimestampMonotonic": "0",
"ExecMainStatus": "0",
"ExecStart": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman volume create --ignore systemd-quadlet-demo-mysql ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"ExecStartEx": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman volume create --ignore systemd-quadlet-demo-mysql ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"ExitType": "main",
"ExtensionImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent",
"FailureAction": "none",
"FileDescriptorStoreMax": "0",
"FileDescriptorStorePreserve": "restart",
"FinalKillSignal": "9",
"FragmentPath": "/run/systemd/generator/quadlet-demo-mysql-volume.service",
"FreezerState": "running",
"GID": "[not set]",
"GuessMainPID": "yes",
"IOAccounting": "no",
"IOReadBytes": "18446744073709551615",
"IOReadOperations": "18446744073709551615",
"IOSchedulingClass": "2",
"IOSchedulingPriority": "4",
"IOWeight": "[not set]",
"IOWriteBytes": "18446744073709551615",
"IOWriteOperations": "18446744073709551615",
"IPAccounting": "no",
"IPEgressBytes": "[no data]",
"IPEgressPackets": "[no data]",
"IPIngressBytes": "[no data]",
"IPIngressPackets": "[no data]",
"Id": "quadlet-demo-mysql-volume.service",
"IgnoreOnIsolate": "no",
"IgnoreSIGPIPE": "yes",
"InactiveEnterTimestampMonotonic": "0",
"InactiveExitTimestampMonotonic": "0",
"JobRunningTimeoutUSec": "infinity",
"JobTimeoutAction": "none",
"JobTimeoutUSec": "infinity",
"KeyringMode": "private",
"KillMode": "control-group",
"KillSignal": "15",
"LimitAS": "infinity",
"LimitASSoft": "infinity",
"LimitCORE": "infinity",
"LimitCORESoft": "infinity",
"LimitCPU": "infinity",
"LimitCPUSoft": "infinity",
"LimitDATA": "infinity",
"LimitDATASoft": "infinity",
"LimitFSIZE": "infinity",
"LimitFSIZESoft": "infinity",
"LimitLOCKS": "infinity",
"LimitLOCKSSoft": "infinity",
"LimitMEMLOCK": "8388608",
"LimitMEMLOCKSoft": "8388608",
"LimitMSGQUEUE": "819200",
"LimitMSGQUEUESoft": "819200",
"LimitNICE": "0",
"LimitNICESoft": "0",
"LimitNOFILE": "524288",
"LimitNOFILESoft": "1024",
"LimitNPROC": "14756",
"LimitNPROCSoft": "14756",
"LimitRSS": "infinity",
"LimitRSSSoft": "infinity",
"LimitRTPRIO": "0",
"LimitRTPRIOSoft": "0",
"LimitRTTIME": "infinity",
"LimitRTTIMESoft": "infinity",
"LimitSIGPENDING": "14756",
"LimitSIGPENDINGSoft": "14756",
"LimitSTACK": "infinity",
"LimitSTACKSoft": "8388608",
"LoadState": "loaded",
"LockPersonality": "no",
"LogLevelMax": "-1",
"LogRateLimitBurst": "0",
"LogRateLimitIntervalUSec": "0",
"LogsDirectoryMode": "0755",
"MainPID": "0",
"ManagedOOMMemoryPressure": "auto",
"ManagedOOMMemoryPressureLimit": "0",
"ManagedOOMPreference": "none",
"ManagedOOMSwap": "auto",
"MemoryAccounting": "yes",
"MemoryAvailable": "infinity",
"MemoryCurrent": "[not set]",
"MemoryDenyWriteExecute": "no",
"MemoryHigh": "infinity",
"MemoryKSM": "no",
"MemoryLimit": "infinity",
"MemoryLow": "0",
"MemoryMax": "infinity",
"MemoryMin": "0",
"MemoryPressureThresholdUSec": "200ms",
"MemoryPressureWatch": "auto",
"MemorySwapMax": "infinity",
"MemoryZSwapMax": "infinity",
"MountAPIVFS": "no",
"MountImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent",
"NFileDescriptorStore": "0",
"NRestarts": "0",
"NUMAPolicy": "n/a",
"Names": "quadlet-demo-mysql-volume.service",
"NeedDaemonReload": "no",
"Nice": "0",
"NoNewPrivileges": "no",
"NonBlocking": "no",
"NotifyAccess": "none",
"OOMPolicy": "stop",
"OOMScoreAdjust": "0",
"OnFailureJobMode": "replace",
"OnSuccessJobMode": "fail",
"Perpetual": "no",
"PrivateDevices": "no",
"PrivateIPC": "no",
"PrivateMounts": "no",
"PrivateNetwork": "no",
"PrivateTmp": "no",
"PrivateUsers": "no",
"ProcSubset": "all",
"ProtectClock": "no",
"ProtectControlGroups": "no",
"ProtectHome": "no",
"ProtectHostname": "no",
"ProtectKernelLogs": "no",
"ProtectKernelModules": "no",
"ProtectKernelTunables": "no",
"ProtectProc": "default",
"ProtectSystem": "no",
"RefuseManualStart": "no",
"RefuseManualStop": "no",
"ReloadResult": "success",
"ReloadSignal": "1",
"RemainAfterExit": "yes",
"RemoveIPC": "no",
"Requires": "sysinit.target -.mount system.slice",
"RequiresMountsFor": "/run/containers",
"Restart": "no",
"RestartKillSignal": "15",
"RestartMaxDelayUSec": "infinity",
"RestartMode": "normal",
"RestartSteps": "0",
"RestartUSec": "100ms",
"RestartUSecNext": "100ms",
"RestrictNamespaces": "no",
"RestrictRealtime": "no",
"RestrictSUIDSGID": "no",
"Result": "success",
"RootDirectoryStartOnly": "no",
"RootEphemeral": "no",
"RootImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent",
"RuntimeDirectoryMode": "0755",
"RuntimeDirectoryPreserve": "no",
"RuntimeMaxUSec": "infinity",
"RuntimeRandomizedExtraUSec": "0",
"SameProcessGroup": "no",
"SecureBits": "0",
"SendSIGHUP": "no",
"SendSIGKILL": "yes",
"Slice": "system.slice",
"StandardError": "inherit",
"StandardInput": "null",
"StandardOutput": "journal",
"StartLimitAction": "none",
"StartLimitBurst": "5",
"StartLimitIntervalUSec": "10s",
"StartupBlockIOWeight": "[not set]",
"StartupCPUShares": "[not set]",
"StartupCPUWeight": "[not set]",
"StartupIOWeight": "[not set]",
"StartupMemoryHigh": "18446744073709551615",
"StartupMemoryLow": "0",
"StartupMemoryMax": "18446744073709551615",
"StartupMemorySwapMax": "18446744073709551615",
"StartupMemoryZSwapMax": "18446744073709551615",
"StateChangeTimestampMonotonic": "0",
"StateDirectoryMode": "0755",
"StatusErrno": "0",
"StopWhenUnneeded": "no",
"SubState": "dead",
"SuccessAction": "none",
"SyslogFacility": "3",
"SyslogIdentifier": "quadlet-demo-mysql-volume",
"SyslogLevel": "6",
"SyslogLevelPrefix": "yes",
"SyslogPriority": "30",
"SystemCallErrorNumber": "2147483646",
"TTYReset": "no",
"TTYVHangup": "no",
"TTYVTDisallocate": "no",
"TasksAccounting": "yes",
"TasksCurrent": "[not set]",
"TasksMax": "4426",
"TimeoutAbortUSec": "45s",
"TimeoutCleanUSec": "infinity",
"TimeoutStartFailureMode": "terminate",
"TimeoutStartUSec": "infinity",
"TimeoutStopFailureMode": "abort",
"TimeoutStopUSec": "45s",
"TimerSlackNSec": "50000",
"Transient": "no",
"Type": "oneshot",
"UID": "[not set]",
"UMask": "0022",
"UnitFilePreset": "disabled",
"UnitFileState": "generated",
"UtmpMode": "init",
"WatchdogSignal": "6",
"WatchdogTimestampMonotonic": "0",
"WatchdogUSec": "infinity"
}
}
TASK [fedora.linux_system_roles.podman : Restart service] **********************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:125
Wednesday 10 July 2024 05:14:02 +0000 (0:00:00.462) 0:02:07.014 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_service_started is changed",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14
Wednesday 10 July 2024 05:14:02 +0000 (0:00:00.021) 0:02:07.035 ********
ok: [sut] => {
"ansible_facts": {
"__podman_quadlet_file_src": "",
"__podman_quadlet_spec": {},
"__podman_quadlet_str": "[Install]\nWantedBy=default.target\n\n[Container]\nImage=quay.io/linux-system-roles/mysql:5.6\nContainerName=quadlet-demo-mysql\nVolume=quadlet-demo-mysql.volume:/var/lib/mysql\nVolume=/tmp/quadlet_demo:/var/lib/quadlet_demo:Z\nNetwork=quadlet-demo.network\nSecret=mysql-root-password-container,type=env,target=MYSQL_ROOT_PASSWORD\nHealthCmd=/usr/bin/true\nHealthOnFailure=kill\n",
"__podman_quadlet_template_src": "quadlet-demo-mysql.container.j2"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25
Wednesday 10 July 2024 05:14:02 +0000 (0:00:00.058) 0:02:07.094 ********
ok: [sut] => {
"ansible_facts": {
"__podman_continue_if_pull_fails": false,
"__podman_pull_image": true,
"__podman_state": "created",
"__podman_systemd_unit_scope": "",
"__podman_user": "root"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] *****
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35
Wednesday 10 July 2024 05:14:02 +0000 (0:00:00.024) 0:02:07.118 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_quadlet_str",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48
Wednesday 10 July 2024 05:14:02 +0000 (0:00:00.019) 0:02:07.138 ********
ok: [sut] => {
"ansible_facts": {
"__podman_quadlet_name": "quadlet-demo-mysql",
"__podman_quadlet_type": "container",
"__podman_rootless": false
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Check user and group information] *****
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57
Wednesday 10 July 2024 05:14:02 +0000 (0:00:00.028) 0:02:07.166 ********
included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for sut
TASK [fedora.linux_system_roles.podman : Get user information] *****************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2
Wednesday 10 July 2024 05:14:02 +0000 (0:00:00.062) 0:02:07.229 ********
skipping: [sut] => {
"changed": false,
"false_condition": "'getent_passwd' not in ansible_facts or __podman_user not in ansible_facts['getent_passwd']",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Fail if user does not exist] **********
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9
Wednesday 10 July 2024 05:14:02 +0000 (0:00:00.022) 0:02:07.252 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not ansible_facts[\"getent_passwd\"][__podman_user]",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set group for podman user] ************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16
Wednesday 10 July 2024 05:14:02 +0000 (0:00:00.021) 0:02:07.273 ********
ok: [sut] => {
"ansible_facts": {
"__podman_group": "0"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Get group information] ****************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:28
Wednesday 10 July 2024 05:14:02 +0000 (0:00:00.026) 0:02:07.299 ********
ok: [sut] => {
"ansible_facts": {
"getent_group": {
"root": [
"x",
"0",
""
]
}
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set group name] ***********************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:35
Wednesday 10 July 2024 05:14:02 +0000 (0:00:00.239) 0:02:07.539 ********
ok: [sut] => {
"ansible_facts": {
"__podman_group_name": "root"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : See if getsubids exists] **************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:39
Wednesday 10 July 2024 05:14:02 +0000 (0:00:00.026) 0:02:07.565 ********
ok: [sut] => {
"changed": false,
"stat": {
"atime": 1720588427.5164855,
"attr_flags": "e",
"attributes": [
"extents"
],
"block_size": 4096,
"blocks": 32,
"charset": "binary",
"checksum": "72ba28a3be19c467e89ab90f5d21f99b605159a7",
"ctime": 1720588377.3007717,
"dev": 51714,
"device_type": 0,
"executable": true,
"exists": true,
"gid": 0,
"gr_name": "root",
"inode": 203603,
"isblk": false,
"ischr": false,
"isdir": false,
"isfifo": false,
"isgid": false,
"islnk": false,
"isreg": true,
"issock": false,
"isuid": false,
"mimetype": "application/x-pie-executable",
"mode": "0755",
"mtime": 1696291200.0,
"nlink": 1,
"path": "/usr/bin/getsubids",
"pw_name": "root",
"readable": true,
"rgrp": true,
"roth": true,
"rusr": true,
"size": 16032,
"uid": 0,
"version": "2092964490",
"wgrp": false,
"woth": false,
"writeable": true,
"wusr": true,
"xgrp": true,
"xoth": true,
"xusr": true
}
}
TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:50
Wednesday 10 July 2024 05:14:03 +0000 (0:00:00.234) 0:02:07.800 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_user not in [\"root\", \"0\"]",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Check group with getsubids] ***********
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:55
Wednesday 10 July 2024 05:14:03 +0000 (0:00:00.018) 0:02:07.818 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_user not in [\"root\", \"0\"]",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ******
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:60
Wednesday 10 July 2024 05:14:03 +0000 (0:00:00.017) 0:02:07.836 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_user not in [\"root\", \"0\"]",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Get subuid file] **********************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:74
Wednesday 10 July 2024 05:14:03 +0000 (0:00:00.017) 0:02:07.853 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_stat_getsubids.stat.exists",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Get subgid file] **********************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:79
Wednesday 10 July 2024 05:14:03 +0000 (0:00:00.017) 0:02:07.870 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_stat_getsubids.stat.exists",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ******
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:84
Wednesday 10 July 2024 05:14:03 +0000 (0:00:00.017) 0:02:07.888 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_stat_getsubids.stat.exists",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ******
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:94
Wednesday 10 July 2024 05:14:03 +0000 (0:00:00.017) 0:02:07.905 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_stat_getsubids.stat.exists",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] *****
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:101
Wednesday 10 July 2024 05:14:03 +0000 (0:00:00.016) 0:02:07.922 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_stat_getsubids.stat.exists",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62
Wednesday 10 July 2024 05:14:03 +0000 (0:00:00.017) 0:02:07.940 ********
ok: [sut] => {
"ansible_facts": {
"__podman_activate_systemd_unit": true,
"__podman_images_found": [
"quay.io/linux-system-roles/mysql:5.6"
],
"__podman_kube_yamls_raw": "",
"__podman_service_name": "quadlet-demo-mysql.service",
"__podman_systemd_scope": "system",
"__podman_user_home_dir": "/root",
"__podman_xdg_runtime_dir": "/run/user/0"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73
Wednesday 10 July 2024 05:14:03 +0000 (0:00:00.033) 0:02:07.973 ********
ok: [sut] => {
"ansible_facts": {
"__podman_quadlet_path": "/etc/containers/systemd"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Get kube yaml contents] ***************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77
Wednesday 10 July 2024 05:14:03 +0000 (0:00:00.020) 0:02:07.994 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_kube_yamls_raw | length > 0",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87
Wednesday 10 July 2024 05:14:03 +0000 (0:00:00.018) 0:02:08.013 ********
ok: [sut] => {
"ansible_facts": {
"__podman_images": [
"quay.io/linux-system-roles/mysql:5.6"
],
"__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo-mysql.container",
"__podman_volumes": [
"/tmp/quadlet_demo"
]
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 6] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:103
Wednesday 10 July 2024 05:14:03 +0000 (0:00:00.044) 0:02:08.057 ********
ok: [sut] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : Cleanup quadlets] *********************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:110
Wednesday 10 July 2024 05:14:03 +0000 (0:00:00.051) 0:02:08.109 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_state == \"absent\"",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Create and update quadlets] ***********
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:114
Wednesday 10 July 2024 05:14:03 +0000 (0:00:00.017) 0:02:08.127 ********
included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml for sut
TASK [fedora.linux_system_roles.podman : Manage linger] ************************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:2
Wednesday 10 July 2024 05:14:03 +0000 (0:00:00.039) 0:02:08.167 ********
included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for sut
TASK [fedora.linux_system_roles.podman : Enable linger if needed] **************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12
Wednesday 10 July 2024 05:14:03 +0000 (0:00:00.030) 0:02:08.197 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_rootless | bool",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18
Wednesday 10 July 2024 05:14:03 +0000 (0:00:00.017) 0:02:08.215 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_rootless | bool",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22
Wednesday 10 July 2024 05:14:03 +0000 (0:00:00.017) 0:02:08.232 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_rootless | bool",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Create host directories] **************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:7
Wednesday 10 July 2024 05:14:03 +0000 (0:00:00.016) 0:02:08.249 ********
changed: [sut] => (item=/tmp/quadlet_demo) => {
"ansible_loop_var": "item",
"changed": true,
"gid": 0,
"group": "root",
"item": "/tmp/quadlet_demo",
"mode": "0777",
"owner": "root",
"path": "/tmp/quadlet_demo",
"secontext": "unconfined_u:object_r:user_tmp_t:s0",
"size": 40,
"state": "directory",
"uid": 0
}
TASK [fedora.linux_system_roles.podman : Ensure container images are present] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:18
Wednesday 10 July 2024 05:14:03 +0000 (0:00:00.252) 0:02:08.502 ********
changed: [sut] => (item=None) => {
"attempts": 1,
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": true
}
changed: [sut] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": true
}
TASK [fedora.linux_system_roles.podman : Ensure the quadlet directory is present] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:39
Wednesday 10 July 2024 05:14:10 +0000 (0:00:06.857) 0:02:15.359 ********
ok: [sut] => {
"changed": false,
"gid": 0,
"group": "root",
"mode": "0755",
"owner": "root",
"path": "/etc/containers/systemd",
"secontext": "system_u:object_r:etc_t:s0",
"size": 4096,
"state": "directory",
"uid": 0
}
TASK [fedora.linux_system_roles.podman : Ensure quadlet file is copied] ********
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:48
Wednesday 10 July 2024 05:14:10 +0000 (0:00:00.245) 0:02:15.604 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_quadlet_file_src | length > 0",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Ensure quadlet file content is present] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:58
Wednesday 10 July 2024 05:14:10 +0000 (0:00:00.019) 0:02:15.624 ********
changed: [sut] => {
"changed": true,
"checksum": "7bf9312d57274d4dabbd222b01717e2cc0cfa0ed",
"dest": "/etc/containers/systemd/quadlet-demo-mysql.container",
"gid": 0,
"group": "root",
"md5sum": "f1cd3875deb45b806389825bb4b5ad4a",
"mode": "0644",
"owner": "root",
"secontext": "system_u:object_r:etc_t:s0",
"size": 367,
"src": "/root/.ansible/tmp/ansible-tmp-1720588450.9974556-12056-158306006965184/source",
"state": "file",
"uid": 0
}
TASK [fedora.linux_system_roles.podman : Ensure quadlet file is present] *******
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:70
Wednesday 10 July 2024 05:14:11 +0000 (0:00:00.584) 0:02:16.208 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_copy_content is skipped",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Reload systemctl] *********************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:82
Wednesday 10 July 2024 05:14:11 +0000 (0:00:00.018) 0:02:16.227 ********
ok: [sut] => {
"changed": false,
"name": null,
"status": {}
}
TASK [fedora.linux_system_roles.podman : Start service] ************************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:110
Wednesday 10 July 2024 05:14:12 +0000 (0:00:00.812) 0:02:17.039 ********
changed: [sut] => {
"changed": true,
"name": "quadlet-demo-mysql.service",
"state": "started",
"status": {
"AccessSELinuxContext": "system_u:object_r:systemd_unit_file_t:s0",
"ActiveEnterTimestampMonotonic": "0",
"ActiveExitTimestampMonotonic": "0",
"ActiveState": "inactive",
"After": "sysinit.target basic.target quadlet-demo-network.service tmp.mount quadlet-demo-mysql-volume.service -.mount system.slice systemd-journald.socket",
"AllowIsolate": "no",
"AssertResult": "no",
"AssertTimestampMonotonic": "0",
"Before": "multi-user.target shutdown.target",
"BlockIOAccounting": "no",
"BlockIOWeight": "[not set]",
"CPUAccounting": "yes",
"CPUAffinityFromNUMA": "no",
"CPUQuotaPerSecUSec": "infinity",
"CPUQuotaPeriodUSec": "infinity",
"CPUSchedulingPolicy": "0",
"CPUSchedulingPriority": "0",
"CPUSchedulingResetOnFork": "no",
"CPUShares": "[not set]",
"CPUUsageNSec": "[not set]",
"CPUWeight": "[not set]",
"CacheDirectoryMode": "0755",
"CanFreeze": "yes",
"CanIsolate": "no",
"CanReload": "no",
"CanStart": "yes",
"CanStop": "yes",
"CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore",
"CleanResult": "success",
"CollectMode": "inactive",
"ConditionResult": "no",
"ConditionTimestampMonotonic": "0",
"ConfigurationDirectoryMode": "0755",
"Conflicts": "shutdown.target",
"ControlGroupId": "0",
"ControlPID": "0",
"CoredumpFilter": "0x33",
"DefaultDependencies": "yes",
"DefaultMemoryLow": "0",
"DefaultMemoryMin": "0",
"DefaultStartupMemoryLow": "0",
"Delegate": "yes",
"DelegateControllers": "cpu cpuset io memory pids",
"Description": "quadlet-demo-mysql.service",
"DevicePolicy": "auto",
"DropInPaths": "/usr/lib/systemd/system/service.d/10-timeout-abort.conf",
"DynamicUser": "no",
"Environment": "PODMAN_SYSTEMD_UNIT=quadlet-demo-mysql.service",
"ExecMainCode": "0",
"ExecMainExitTimestampMonotonic": "0",
"ExecMainPID": "0",
"ExecMainStartTimestampMonotonic": "0",
"ExecMainStatus": "0",
"ExecStart": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman run --name=quadlet-demo-mysql --cidfile=/run/quadlet-demo-mysql.cid --replace --rm --cgroups=split --network=systemd-quadlet-demo --sdnotify=conmon -d -v systemd-quadlet-demo-mysql:/var/lib/mysql -v /tmp/quadlet_demo:/var/lib/quadlet_demo:Z --secret mysql-root-password-container,type=env,target=MYSQL_ROOT_PASSWORD --health-cmd /usr/bin/true --health-on-failure kill quay.io/linux-system-roles/mysql:5.6 ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"ExecStartEx": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman run --name=quadlet-demo-mysql --cidfile=/run/quadlet-demo-mysql.cid --replace --rm --cgroups=split --network=systemd-quadlet-demo --sdnotify=conmon -d -v systemd-quadlet-demo-mysql:/var/lib/mysql -v /tmp/quadlet_demo:/var/lib/quadlet_demo:Z --secret mysql-root-password-container,type=env,target=MYSQL_ROOT_PASSWORD --health-cmd /usr/bin/true --health-on-failure kill quay.io/linux-system-roles/mysql:5.6 ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"ExecStop": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman rm -v -f -i --cidfile=/run/quadlet-demo-mysql.cid ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"ExecStopEx": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman rm -v -f -i --cidfile=/run/quadlet-demo-mysql.cid ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"ExecStopPost": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman rm -v -f -i --cidfile=/run/quadlet-demo-mysql.cid ; ignore_errors=yes ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"ExecStopPostEx": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman rm -v -f -i --cidfile=/run/quadlet-demo-mysql.cid ; flags=ignore-failure ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"ExitType": "main",
"ExtensionImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent",
"FailureAction": "none",
"FileDescriptorStoreMax": "0",
"FileDescriptorStorePreserve": "restart",
"FinalKillSignal": "9",
"FragmentPath": "/run/systemd/generator/quadlet-demo-mysql.service",
"FreezerState": "running",
"GID": "[not set]",
"GuessMainPID": "yes",
"IOAccounting": "no",
"IOReadBytes": "18446744073709551615",
"IOReadOperations": "18446744073709551615",
"IOSchedulingClass": "2",
"IOSchedulingPriority": "4",
"IOWeight": "[not set]",
"IOWriteBytes": "18446744073709551615",
"IOWriteOperations": "18446744073709551615",
"IPAccounting": "no",
"IPEgressBytes": "[no data]",
"IPEgressPackets": "[no data]",
"IPIngressBytes": "[no data]",
"IPIngressPackets": "[no data]",
"Id": "quadlet-demo-mysql.service",
"IgnoreOnIsolate": "no",
"IgnoreSIGPIPE": "yes",
"InactiveEnterTimestampMonotonic": "0",
"InactiveExitTimestampMonotonic": "0",
"JobRunningTimeoutUSec": "infinity",
"JobTimeoutAction": "none",
"JobTimeoutUSec": "infinity",
"KeyringMode": "private",
"KillMode": "mixed",
"KillSignal": "15",
"LimitAS": "infinity",
"LimitASSoft": "infinity",
"LimitCORE": "infinity",
"LimitCORESoft": "infinity",
"LimitCPU": "infinity",
"LimitCPUSoft": "infinity",
"LimitDATA": "infinity",
"LimitDATASoft": "infinity",
"LimitFSIZE": "infinity",
"LimitFSIZESoft": "infinity",
"LimitLOCKS": "infinity",
"LimitLOCKSSoft": "infinity",
"LimitMEMLOCK": "8388608",
"LimitMEMLOCKSoft": "8388608",
"LimitMSGQUEUE": "819200",
"LimitMSGQUEUESoft": "819200",
"LimitNICE": "0",
"LimitNICESoft": "0",
"LimitNOFILE": "524288",
"LimitNOFILESoft": "1024",
"LimitNPROC": "14756",
"LimitNPROCSoft": "14756",
"LimitRSS": "infinity",
"LimitRSSSoft": "infinity",
"LimitRTPRIO": "0",
"LimitRTPRIOSoft": "0",
"LimitRTTIME": "infinity",
"LimitRTTIMESoft": "infinity",
"LimitSIGPENDING": "14756",
"LimitSIGPENDINGSoft": "14756",
"LimitSTACK": "infinity",
"LimitSTACKSoft": "8388608",
"LoadState": "loaded",
"LockPersonality": "no",
"LogLevelMax": "-1",
"LogRateLimitBurst": "0",
"LogRateLimitIntervalUSec": "0",
"LogsDirectoryMode": "0755",
"MainPID": "0",
"ManagedOOMMemoryPressure": "auto",
"ManagedOOMMemoryPressureLimit": "0",
"ManagedOOMPreference": "none",
"ManagedOOMSwap": "auto",
"MemoryAccounting": "yes",
"MemoryAvailable": "infinity",
"MemoryCurrent": "[not set]",
"MemoryDenyWriteExecute": "no",
"MemoryHigh": "infinity",
"MemoryKSM": "no",
"MemoryLimit": "infinity",
"MemoryLow": "0",
"MemoryMax": "infinity",
"MemoryMin": "0",
"MemoryPressureThresholdUSec": "200ms",
"MemoryPressureWatch": "auto",
"MemorySwapMax": "infinity",
"MemoryZSwapMax": "infinity",
"MountAPIVFS": "no",
"MountImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent",
"NFileDescriptorStore": "0",
"NRestarts": "0",
"NUMAPolicy": "n/a",
"Names": "quadlet-demo-mysql.service",
"NeedDaemonReload": "no",
"Nice": "0",
"NoNewPrivileges": "no",
"NonBlocking": "no",
"NotifyAccess": "all",
"OOMPolicy": "continue",
"OOMScoreAdjust": "0",
"OnFailureJobMode": "replace",
"OnSuccessJobMode": "fail",
"Perpetual": "no",
"PrivateDevices": "no",
"PrivateIPC": "no",
"PrivateMounts": "no",
"PrivateNetwork": "no",
"PrivateTmp": "no",
"PrivateUsers": "no",
"ProcSubset": "all",
"ProtectClock": "no",
"ProtectControlGroups": "no",
"ProtectHome": "no",
"ProtectHostname": "no",
"ProtectKernelLogs": "no",
"ProtectKernelModules": "no",
"ProtectKernelTunables": "no",
"ProtectProc": "default",
"ProtectSystem": "no",
"RefuseManualStart": "no",
"RefuseManualStop": "no",
"ReloadResult": "success",
"ReloadSignal": "1",
"RemainAfterExit": "no",
"RemoveIPC": "no",
"Requires": "system.slice -.mount tmp.mount sysinit.target quadlet-demo-network.service quadlet-demo-mysql-volume.service",
"RequiresMountsFor": "/tmp/quadlet_demo /run/containers",
"Restart": "no",
"RestartKillSignal": "15",
"RestartMaxDelayUSec": "infinity",
"RestartMode": "normal",
"RestartSteps": "0",
"RestartUSec": "100ms",
"RestartUSecNext": "100ms",
"RestrictNamespaces": "no",
"RestrictRealtime": "no",
"RestrictSUIDSGID": "no",
"Result": "success",
"RootDirectoryStartOnly": "no",
"RootEphemeral": "no",
"RootImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent",
"RuntimeDirectoryMode": "0755",
"RuntimeDirectoryPreserve": "no",
"RuntimeMaxUSec": "infinity",
"RuntimeRandomizedExtraUSec": "0",
"SameProcessGroup": "no",
"SecureBits": "0",
"SendSIGHUP": "no",
"SendSIGKILL": "yes",
"Slice": "system.slice",
"SourcePath": "/etc/containers/systemd/quadlet-demo-mysql.container",
"StandardError": "inherit",
"StandardInput": "null",
"StandardOutput": "journal",
"StartLimitAction": "none",
"StartLimitBurst": "5",
"StartLimitIntervalUSec": "10s",
"StartupBlockIOWeight": "[not set]",
"StartupCPUShares": "[not set]",
"StartupCPUWeight": "[not set]",
"StartupIOWeight": "[not set]",
"StartupMemoryHigh": "18446744073709551615",
"StartupMemoryLow": "0",
"StartupMemoryMax": "18446744073709551615",
"StartupMemorySwapMax": "18446744073709551615",
"StartupMemoryZSwapMax": "18446744073709551615",
"StateChangeTimestampMonotonic": "0",
"StateDirectoryMode": "0755",
"StatusErrno": "0",
"StopWhenUnneeded": "no",
"SubState": "dead",
"SuccessAction": "none",
"SyslogFacility": "3",
"SyslogIdentifier": "quadlet-demo-mysql",
"SyslogLevel": "6",
"SyslogLevelPrefix": "yes",
"SyslogPriority": "30",
"SystemCallErrorNumber": "2147483646",
"TTYReset": "no",
"TTYVHangup": "no",
"TTYVTDisallocate": "no",
"TasksAccounting": "yes",
"TasksCurrent": "[not set]",
"TasksMax": "4426",
"TimeoutAbortUSec": "45s",
"TimeoutCleanUSec": "infinity",
"TimeoutStartFailureMode": "terminate",
"TimeoutStartUSec": "45s",
"TimeoutStopFailureMode": "abort",
"TimeoutStopUSec": "45s",
"TimerSlackNSec": "50000",
"Transient": "no",
"Type": "notify",
"UID": "[not set]",
"UMask": "0022",
"UnitFilePreset": "disabled",
"UnitFileState": "generated",
"UtmpMode": "init",
"WantedBy": "multi-user.target",
"WatchdogSignal": "6",
"WatchdogTimestampMonotonic": "0",
"WatchdogUSec": "infinity"
}
}
TASK [fedora.linux_system_roles.podman : Restart service] **********************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:125
Wednesday 10 July 2024 05:14:13 +0000 (0:00:00.907) 0:02:17.946 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_service_started is changed",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14
Wednesday 10 July 2024 05:14:13 +0000 (0:00:00.019) 0:02:17.965 ********
ok: [sut] => {
"ansible_facts": {
"__podman_quadlet_file_src": "envoy-proxy-configmap.yml",
"__podman_quadlet_spec": {},
"__podman_quadlet_str": "---\napiVersion: v1\nkind: ConfigMap\nmetadata:\n name: envoy-proxy-config\ndata:\n envoy.yaml: |\n admin:\n address:\n socket_address:\n address: 0.0.0.0\n port_value: 9901\n\n static_resources:\n listeners:\n - name: listener_0\n address:\n socket_address:\n address: 0.0.0.0\n port_value: 8080\n filter_chains:\n - filters:\n - name: envoy.filters.network.http_connection_manager\n typed_config:\n \"@type\": type.googleapis.com/envoy.extensions.filters.network.http_connection_manager.v3.HttpConnectionManager\n stat_prefix: ingress_http\n codec_type: AUTO\n route_config:\n name: local_route\n virtual_hosts:\n - name: local_service\n domains: [\"*\"]\n routes:\n - match:\n prefix: \"/\"\n route:\n cluster: backend\n http_filters:\n - name: envoy.filters.http.router\n typed_config:\n \"@type\": type.googleapis.com/envoy.extensions.filters.http.router.v3.Router\n transport_socket:\n name: envoy.transport_sockets.tls\n typed_config:\n \"@type\": type.googleapis.com/envoy.extensions.transport_sockets.tls.v3.DownstreamTlsContext\n common_tls_context:\n tls_certificates:\n - certificate_chain:\n filename: /etc/envoy-certificates/certificate.pem\n private_key:\n filename: /etc/envoy-certificates/certificate.key\n clusters:\n - name: backend\n connect_timeout: 5s\n type: STATIC\n dns_refresh_rate: 1800s\n lb_policy: ROUND_ROBIN\n load_assignment:\n cluster_name: backend\n endpoints:\n - lb_endpoints:\n - endpoint:\n address:\n socket_address:\n address: 127.0.0.1\n port_value: 80",
"__podman_quadlet_template_src": ""
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25
Wednesday 10 July 2024 05:14:13 +0000 (0:00:00.025) 0:02:17.991 ********
ok: [sut] => {
"ansible_facts": {
"__podman_continue_if_pull_fails": false,
"__podman_pull_image": true,
"__podman_state": "created",
"__podman_systemd_unit_scope": "",
"__podman_user": "root"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] *****
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35
Wednesday 10 July 2024 05:14:13 +0000 (0:00:00.052) 0:02:18.044 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_quadlet_file_src",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48
Wednesday 10 July 2024 05:14:13 +0000 (0:00:00.019) 0:02:18.064 ********
ok: [sut] => {
"ansible_facts": {
"__podman_quadlet_name": "envoy-proxy-configmap",
"__podman_quadlet_type": "yml",
"__podman_rootless": false
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Check user and group information] *****
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57
Wednesday 10 July 2024 05:14:13 +0000 (0:00:00.029) 0:02:18.094 ********
included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for sut
TASK [fedora.linux_system_roles.podman : Get user information] *****************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2
Wednesday 10 July 2024 05:14:13 +0000 (0:00:00.035) 0:02:18.129 ********
skipping: [sut] => {
"changed": false,
"false_condition": "'getent_passwd' not in ansible_facts or __podman_user not in ansible_facts['getent_passwd']",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Fail if user does not exist] **********
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9
Wednesday 10 July 2024 05:14:13 +0000 (0:00:00.020) 0:02:18.150 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not ansible_facts[\"getent_passwd\"][__podman_user]",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set group for podman user] ************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16
Wednesday 10 July 2024 05:14:13 +0000 (0:00:00.020) 0:02:18.170 ********
ok: [sut] => {
"ansible_facts": {
"__podman_group": "0"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Get group information] ****************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:28
Wednesday 10 July 2024 05:14:13 +0000 (0:00:00.027) 0:02:18.197 ********
ok: [sut] => {
"ansible_facts": {
"getent_group": {
"root": [
"x",
"0",
""
]
}
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set group name] ***********************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:35
Wednesday 10 July 2024 05:14:13 +0000 (0:00:00.267) 0:02:18.465 ********
ok: [sut] => {
"ansible_facts": {
"__podman_group_name": "root"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : See if getsubids exists] **************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:39
Wednesday 10 July 2024 05:14:13 +0000 (0:00:00.025) 0:02:18.490 ********
ok: [sut] => {
"changed": false,
"stat": {
"atime": 1720588427.5164855,
"attr_flags": "e",
"attributes": [
"extents"
],
"block_size": 4096,
"blocks": 32,
"charset": "binary",
"checksum": "72ba28a3be19c467e89ab90f5d21f99b605159a7",
"ctime": 1720588377.3007717,
"dev": 51714,
"device_type": 0,
"executable": true,
"exists": true,
"gid": 0,
"gr_name": "root",
"inode": 203603,
"isblk": false,
"ischr": false,
"isdir": false,
"isfifo": false,
"isgid": false,
"islnk": false,
"isreg": true,
"issock": false,
"isuid": false,
"mimetype": "application/x-pie-executable",
"mode": "0755",
"mtime": 1696291200.0,
"nlink": 1,
"path": "/usr/bin/getsubids",
"pw_name": "root",
"readable": true,
"rgrp": true,
"roth": true,
"rusr": true,
"size": 16032,
"uid": 0,
"version": "2092964490",
"wgrp": false,
"woth": false,
"writeable": true,
"wusr": true,
"xgrp": true,
"xoth": true,
"xusr": true
}
}
TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:50
Wednesday 10 July 2024 05:14:14 +0000 (0:00:00.235) 0:02:18.725 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_user not in [\"root\", \"0\"]",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Check group with getsubids] ***********
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:55
Wednesday 10 July 2024 05:14:14 +0000 (0:00:00.017) 0:02:18.743 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_user not in [\"root\", \"0\"]",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ******
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:60
Wednesday 10 July 2024 05:14:14 +0000 (0:00:00.017) 0:02:18.761 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_user not in [\"root\", \"0\"]",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Get subuid file] **********************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:74
Wednesday 10 July 2024 05:14:14 +0000 (0:00:00.017) 0:02:18.778 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_stat_getsubids.stat.exists",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Get subgid file] **********************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:79
Wednesday 10 July 2024 05:14:14 +0000 (0:00:00.017) 0:02:18.795 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_stat_getsubids.stat.exists",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ******
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:84
Wednesday 10 July 2024 05:14:14 +0000 (0:00:00.017) 0:02:18.813 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_stat_getsubids.stat.exists",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ******
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:94
Wednesday 10 July 2024 05:14:14 +0000 (0:00:00.016) 0:02:18.829 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_stat_getsubids.stat.exists",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] *****
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:101
Wednesday 10 July 2024 05:14:14 +0000 (0:00:00.016) 0:02:18.846 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_stat_getsubids.stat.exists",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62
Wednesday 10 July 2024 05:14:14 +0000 (0:00:00.016) 0:02:18.863 ********
ok: [sut] => {
"ansible_facts": {
"__podman_activate_systemd_unit": true,
"__podman_images_found": [],
"__podman_kube_yamls_raw": "",
"__podman_service_name": "",
"__podman_systemd_scope": "system",
"__podman_user_home_dir": "/root",
"__podman_xdg_runtime_dir": "/run/user/0"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73
Wednesday 10 July 2024 05:14:14 +0000 (0:00:00.062) 0:02:18.925 ********
ok: [sut] => {
"ansible_facts": {
"__podman_quadlet_path": "/etc/containers/systemd"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Get kube yaml contents] ***************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77
Wednesday 10 July 2024 05:14:14 +0000 (0:00:00.020) 0:02:18.946 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_kube_yamls_raw | length > 0",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87
Wednesday 10 July 2024 05:14:14 +0000 (0:00:00.019) 0:02:18.965 ********
ok: [sut] => {
"ansible_facts": {
"__podman_images": [],
"__podman_quadlet_file": "/etc/containers/systemd/envoy-proxy-configmap.yml",
"__podman_volumes": []
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 6] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:103
Wednesday 10 July 2024 05:14:14 +0000 (0:00:00.045) 0:02:19.011 ********
ok: [sut] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : Cleanup quadlets] *********************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:110
Wednesday 10 July 2024 05:14:14 +0000 (0:00:00.022) 0:02:19.034 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_state == \"absent\"",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Create and update quadlets] ***********
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:114
Wednesday 10 July 2024 05:14:14 +0000 (0:00:00.017) 0:02:19.052 ********
included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml for sut
TASK [fedora.linux_system_roles.podman : Manage linger] ************************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:2
Wednesday 10 July 2024 05:14:14 +0000 (0:00:00.038) 0:02:19.090 ********
included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for sut
TASK [fedora.linux_system_roles.podman : Enable linger if needed] **************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12
Wednesday 10 July 2024 05:14:14 +0000 (0:00:00.030) 0:02:19.121 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_rootless | bool",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18
Wednesday 10 July 2024 05:14:14 +0000 (0:00:00.017) 0:02:19.139 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_rootless | bool",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22
Wednesday 10 July 2024 05:14:14 +0000 (0:00:00.016) 0:02:19.156 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_rootless | bool",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Create host directories] **************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:7
Wednesday 10 July 2024 05:14:14 +0000 (0:00:00.017) 0:02:19.173 ********
skipping: [sut] => {
"changed": false,
"skipped_reason": "No items in the list"
}
TASK [fedora.linux_system_roles.podman : Ensure container images are present] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:18
Wednesday 10 July 2024 05:14:14 +0000 (0:00:00.015) 0:02:19.188 ********
skipping: [sut] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : Ensure the quadlet directory is present] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:39
Wednesday 10 July 2024 05:14:14 +0000 (0:00:00.016) 0:02:19.205 ********
ok: [sut] => {
"changed": false,
"gid": 0,
"group": "root",
"mode": "0755",
"owner": "root",
"path": "/etc/containers/systemd",
"secontext": "system_u:object_r:etc_t:s0",
"size": 4096,
"state": "directory",
"uid": 0
}
TASK [fedora.linux_system_roles.podman : Ensure quadlet file is copied] ********
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:48
Wednesday 10 July 2024 05:14:14 +0000 (0:00:00.256) 0:02:19.461 ********
changed: [sut] => {
"changed": true,
"checksum": "d681c7d56f912150d041873e880818b22a90c188",
"dest": "/etc/containers/systemd/envoy-proxy-configmap.yml",
"gid": 0,
"group": "root",
"md5sum": "aec75d972c231aac004e1338934544cf",
"mode": "0644",
"owner": "root",
"secontext": "system_u:object_r:etc_t:s0",
"size": 2102,
"src": "/root/.ansible/tmp/ansible-tmp-1720588454.8339243-12107-16527747628217/source",
"state": "file",
"uid": 0
}
TASK [fedora.linux_system_roles.podman : Ensure quadlet file content is present] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:58
Wednesday 10 July 2024 05:14:15 +0000 (0:00:00.578) 0:02:20.040 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_quadlet_file_src",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Ensure quadlet file is present] *******
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:70
Wednesday 10 July 2024 05:14:15 +0000 (0:00:00.020) 0:02:20.060 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_copy_file is skipped",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Reload systemctl] *********************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:82
Wednesday 10 July 2024 05:14:15 +0000 (0:00:00.017) 0:02:20.078 ********
ok: [sut] => {
"changed": false,
"name": null,
"status": {}
}
TASK [fedora.linux_system_roles.podman : Start service] ************************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:110
Wednesday 10 July 2024 05:14:16 +0000 (0:00:00.910) 0:02:20.988 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_service_name | length > 0",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Restart service] **********************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:125
Wednesday 10 July 2024 05:14:16 +0000 (0:00:00.022) 0:02:21.010 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_service_name | length > 0",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14
Wednesday 10 July 2024 05:14:16 +0000 (0:00:00.020) 0:02:21.031 ********
ok: [sut] => {
"ansible_facts": {
"__podman_quadlet_file_src": "",
"__podman_quadlet_spec": {},
"__podman_quadlet_str": "---\napiVersion: v1\nkind: PersistentVolumeClaim\nmetadata:\n name: wp-pv-claim\n labels:\n app: wordpress\nspec:\n accessModes:\n - ReadWriteOnce\n resources:\n requests:\n storage: 20Gi\n---\napiVersion: v1\nkind: Pod\nmetadata:\n name: quadlet-demo\nspec:\n containers:\n - name: wordpress\n image: quay.io/linux-system-roles/wordpress:4.8-apache\n env:\n - name: WORDPRESS_DB_HOST\n value: quadlet-demo-mysql\n - name: WORDPRESS_DB_PASSWORD\n valueFrom:\n secretKeyRef:\n name: mysql-root-password-kube\n key: password\n volumeMounts:\n - name: wordpress-persistent-storage\n mountPath: /var/www/html\n resources:\n requests:\n memory: \"64Mi\"\n cpu: \"250m\"\n limits:\n memory: \"128Mi\"\n cpu: \"500m\"\n - name: envoy\n image: quay.io/linux-system-roles/envoyproxy:v1.25.0\n volumeMounts:\n - name: config-volume\n mountPath: /etc/envoy\n - name: certificates\n mountPath: /etc/envoy-certificates\n env:\n - name: ENVOY_UID\n value: \"0\"\n resources:\n requests:\n memory: \"64Mi\"\n cpu: \"250m\"\n limits:\n memory: \"128Mi\"\n cpu: \"500m\"\n volumes:\n - name: config-volume\n configMap:\n name: envoy-proxy-config\n - name: certificates\n secret:\n secretName: envoy-certificates\n - name: wordpress-persistent-storage\n persistentVolumeClaim:\n claimName: wp-pv-claim\n - name: www # not used - for testing hostpath\n hostPath:\n path: /tmp/httpd3\n - name: create # not used - for testing hostpath\n hostPath:\n path: /tmp/httpd3-create\n",
"__podman_quadlet_template_src": "quadlet-demo.yml.j2"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25
Wednesday 10 July 2024 05:14:16 +0000 (0:00:00.057) 0:02:21.088 ********
ok: [sut] => {
"ansible_facts": {
"__podman_continue_if_pull_fails": false,
"__podman_pull_image": true,
"__podman_state": "created",
"__podman_systemd_unit_scope": "",
"__podman_user": "root"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] *****
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35
Wednesday 10 July 2024 05:14:16 +0000 (0:00:00.024) 0:02:21.112 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_quadlet_str",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48
Wednesday 10 July 2024 05:14:16 +0000 (0:00:00.019) 0:02:21.132 ********
ok: [sut] => {
"ansible_facts": {
"__podman_quadlet_name": "quadlet-demo",
"__podman_quadlet_type": "yml",
"__podman_rootless": false
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Check user and group information] *****
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57
Wednesday 10 July 2024 05:14:16 +0000 (0:00:00.028) 0:02:21.160 ********
included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for sut
TASK [fedora.linux_system_roles.podman : Get user information] *****************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2
Wednesday 10 July 2024 05:14:16 +0000 (0:00:00.034) 0:02:21.195 ********
skipping: [sut] => {
"changed": false,
"false_condition": "'getent_passwd' not in ansible_facts or __podman_user not in ansible_facts['getent_passwd']",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Fail if user does not exist] **********
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9
Wednesday 10 July 2024 05:14:16 +0000 (0:00:00.021) 0:02:21.216 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not ansible_facts[\"getent_passwd\"][__podman_user]",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set group for podman user] ************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16
Wednesday 10 July 2024 05:14:16 +0000 (0:00:00.021) 0:02:21.237 ********
ok: [sut] => {
"ansible_facts": {
"__podman_group": "0"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Get group information] ****************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:28
Wednesday 10 July 2024 05:14:16 +0000 (0:00:00.026) 0:02:21.264 ********
ok: [sut] => {
"ansible_facts": {
"getent_group": {
"root": [
"x",
"0",
""
]
}
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set group name] ***********************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:35
Wednesday 10 July 2024 05:14:16 +0000 (0:00:00.391) 0:02:21.656 ********
ok: [sut] => {
"ansible_facts": {
"__podman_group_name": "root"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : See if getsubids exists] **************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:39
Wednesday 10 July 2024 05:14:17 +0000 (0:00:00.026) 0:02:21.682 ********
ok: [sut] => {
"changed": false,
"stat": {
"atime": 1720588427.5164855,
"attr_flags": "e",
"attributes": [
"extents"
],
"block_size": 4096,
"blocks": 32,
"charset": "binary",
"checksum": "72ba28a3be19c467e89ab90f5d21f99b605159a7",
"ctime": 1720588377.3007717,
"dev": 51714,
"device_type": 0,
"executable": true,
"exists": true,
"gid": 0,
"gr_name": "root",
"inode": 203603,
"isblk": false,
"ischr": false,
"isdir": false,
"isfifo": false,
"isgid": false,
"islnk": false,
"isreg": true,
"issock": false,
"isuid": false,
"mimetype": "application/x-pie-executable",
"mode": "0755",
"mtime": 1696291200.0,
"nlink": 1,
"path": "/usr/bin/getsubids",
"pw_name": "root",
"readable": true,
"rgrp": true,
"roth": true,
"rusr": true,
"size": 16032,
"uid": 0,
"version": "2092964490",
"wgrp": false,
"woth": false,
"writeable": true,
"wusr": true,
"xgrp": true,
"xoth": true,
"xusr": true
}
}
TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:50
Wednesday 10 July 2024 05:14:17 +0000 (0:00:00.290) 0:02:21.973 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_user not in [\"root\", \"0\"]",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Check group with getsubids] ***********
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:55
Wednesday 10 July 2024 05:14:17 +0000 (0:00:00.018) 0:02:21.991 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_user not in [\"root\", \"0\"]",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ******
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:60
Wednesday 10 July 2024 05:14:17 +0000 (0:00:00.017) 0:02:22.008 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_user not in [\"root\", \"0\"]",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Get subuid file] **********************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:74
Wednesday 10 July 2024 05:14:17 +0000 (0:00:00.017) 0:02:22.026 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_stat_getsubids.stat.exists",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Get subgid file] **********************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:79
Wednesday 10 July 2024 05:14:17 +0000 (0:00:00.017) 0:02:22.044 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_stat_getsubids.stat.exists",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ******
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:84
Wednesday 10 July 2024 05:14:17 +0000 (0:00:00.046) 0:02:22.090 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_stat_getsubids.stat.exists",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ******
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:94
Wednesday 10 July 2024 05:14:17 +0000 (0:00:00.018) 0:02:22.109 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_stat_getsubids.stat.exists",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] *****
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:101
Wednesday 10 July 2024 05:14:17 +0000 (0:00:00.017) 0:02:22.127 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_stat_getsubids.stat.exists",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62
Wednesday 10 July 2024 05:14:17 +0000 (0:00:00.018) 0:02:22.145 ********
ok: [sut] => {
"ansible_facts": {
"__podman_activate_systemd_unit": true,
"__podman_images_found": [],
"__podman_kube_yamls_raw": "",
"__podman_service_name": "",
"__podman_systemd_scope": "system",
"__podman_user_home_dir": "/root",
"__podman_xdg_runtime_dir": "/run/user/0"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73
Wednesday 10 July 2024 05:14:17 +0000 (0:00:00.033) 0:02:22.178 ********
ok: [sut] => {
"ansible_facts": {
"__podman_quadlet_path": "/etc/containers/systemd"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Get kube yaml contents] ***************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77
Wednesday 10 July 2024 05:14:17 +0000 (0:00:00.019) 0:02:22.198 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_kube_yamls_raw | length > 0",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87
Wednesday 10 July 2024 05:14:17 +0000 (0:00:00.018) 0:02:22.217 ********
ok: [sut] => {
"ansible_facts": {
"__podman_images": [],
"__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo.yml",
"__podman_volumes": []
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 6] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:103
Wednesday 10 July 2024 05:14:17 +0000 (0:00:00.044) 0:02:22.262 ********
ok: [sut] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : Cleanup quadlets] *********************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:110
Wednesday 10 July 2024 05:14:17 +0000 (0:00:00.023) 0:02:22.285 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_state == \"absent\"",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Create and update quadlets] ***********
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:114
Wednesday 10 July 2024 05:14:17 +0000 (0:00:00.016) 0:02:22.302 ********
included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml for sut
TASK [fedora.linux_system_roles.podman : Manage linger] ************************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:2
Wednesday 10 July 2024 05:14:17 +0000 (0:00:00.039) 0:02:22.342 ********
included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for sut
TASK [fedora.linux_system_roles.podman : Enable linger if needed] **************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12
Wednesday 10 July 2024 05:14:17 +0000 (0:00:00.030) 0:02:22.372 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_rootless | bool",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18
Wednesday 10 July 2024 05:14:17 +0000 (0:00:00.017) 0:02:22.389 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_rootless | bool",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22
Wednesday 10 July 2024 05:14:17 +0000 (0:00:00.017) 0:02:22.406 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_rootless | bool",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Create host directories] **************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:7
Wednesday 10 July 2024 05:14:17 +0000 (0:00:00.017) 0:02:22.424 ********
skipping: [sut] => {
"changed": false,
"skipped_reason": "No items in the list"
}
TASK [fedora.linux_system_roles.podman : Ensure container images are present] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:18
Wednesday 10 July 2024 05:14:17 +0000 (0:00:00.014) 0:02:22.439 ********
skipping: [sut] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : Ensure the quadlet directory is present] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:39
Wednesday 10 July 2024 05:14:17 +0000 (0:00:00.046) 0:02:22.485 ********
ok: [sut] => {
"changed": false,
"gid": 0,
"group": "root",
"mode": "0755",
"owner": "root",
"path": "/etc/containers/systemd",
"secontext": "system_u:object_r:etc_t:s0",
"size": 4096,
"state": "directory",
"uid": 0
}
TASK [fedora.linux_system_roles.podman : Ensure quadlet file is copied] ********
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:48
Wednesday 10 July 2024 05:14:18 +0000 (0:00:00.343) 0:02:22.828 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_quadlet_file_src | length > 0",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Ensure quadlet file content is present] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:58
Wednesday 10 July 2024 05:14:18 +0000 (0:00:00.019) 0:02:22.848 ********
changed: [sut] => {
"changed": true,
"checksum": "998dccde0483b1654327a46ddd89cbaa47650370",
"dest": "/etc/containers/systemd/quadlet-demo.yml",
"gid": 0,
"group": "root",
"md5sum": "fd890594adfc24339cb9cdc5e7b19a66",
"mode": "0644",
"owner": "root",
"secontext": "system_u:object_r:etc_t:s0",
"size": 1605,
"src": "/root/.ansible/tmp/ansible-tmp-1720588458.2299404-12159-197295334066146/source",
"state": "file",
"uid": 0
}
TASK [fedora.linux_system_roles.podman : Ensure quadlet file is present] *******
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:70
Wednesday 10 July 2024 05:14:19 +0000 (0:00:00.922) 0:02:23.771 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_copy_content is skipped",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Reload systemctl] *********************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:82
Wednesday 10 July 2024 05:14:19 +0000 (0:00:00.018) 0:02:23.789 ********
ok: [sut] => {
"changed": false,
"name": null,
"status": {}
}
TASK [fedora.linux_system_roles.podman : Start service] ************************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:110
Wednesday 10 July 2024 05:14:20 +0000 (0:00:01.279) 0:02:25.068 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_service_name | length > 0",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Restart service] **********************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:125
Wednesday 10 July 2024 05:14:20 +0000 (0:00:00.021) 0:02:25.089 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_service_name | length > 0",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14
Wednesday 10 July 2024 05:14:20 +0000 (0:00:00.021) 0:02:25.111 ********
ok: [sut] => {
"ansible_facts": {
"__podman_quadlet_file_src": "quadlet-demo.kube",
"__podman_quadlet_spec": {},
"__podman_quadlet_str": "[Install]\nWantedBy=default.target\n\n[Unit]\nRequires=quadlet-demo-mysql.service\nAfter=quadlet-demo-mysql.service\n\n[Kube]\n# Point to the yaml file in the same directory\nYaml=quadlet-demo.yml\n# Use the quadlet-demo network\nNetwork=quadlet-demo.network\n# Publish the envoy proxy data port\nPublishPort=8000:8080\n# Publish the envoy proxy admin port\nPublishPort=9000:9901\n# Use the envoy proxy config map in the same directory\nConfigMap=envoy-proxy-configmap.yml",
"__podman_quadlet_template_src": ""
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25
Wednesday 10 July 2024 05:14:20 +0000 (0:00:00.026) 0:02:25.137 ********
ok: [sut] => {
"ansible_facts": {
"__podman_continue_if_pull_fails": false,
"__podman_pull_image": true,
"__podman_state": "created",
"__podman_systemd_unit_scope": "",
"__podman_user": "root"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] *****
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35
Wednesday 10 July 2024 05:14:20 +0000 (0:00:00.024) 0:02:25.162 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_quadlet_file_src",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48
Wednesday 10 July 2024 05:14:20 +0000 (0:00:00.018) 0:02:25.181 ********
ok: [sut] => {
"ansible_facts": {
"__podman_quadlet_name": "quadlet-demo",
"__podman_quadlet_type": "kube",
"__podman_rootless": false
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Check user and group information] *****
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57
Wednesday 10 July 2024 05:14:20 +0000 (0:00:00.028) 0:02:25.209 ********
included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for sut
TASK [fedora.linux_system_roles.podman : Get user information] *****************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2
Wednesday 10 July 2024 05:14:20 +0000 (0:00:00.035) 0:02:25.244 ********
skipping: [sut] => {
"changed": false,
"false_condition": "'getent_passwd' not in ansible_facts or __podman_user not in ansible_facts['getent_passwd']",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Fail if user does not exist] **********
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9
Wednesday 10 July 2024 05:14:20 +0000 (0:00:00.020) 0:02:25.265 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not ansible_facts[\"getent_passwd\"][__podman_user]",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set group for podman user] ************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16
Wednesday 10 July 2024 05:14:20 +0000 (0:00:00.020) 0:02:25.285 ********
ok: [sut] => {
"ansible_facts": {
"__podman_group": "0"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Get group information] ****************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:28
Wednesday 10 July 2024 05:14:20 +0000 (0:00:00.026) 0:02:25.312 ********
ok: [sut] => {
"ansible_facts": {
"getent_group": {
"root": [
"x",
"0",
""
]
}
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set group name] ***********************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:35
Wednesday 10 July 2024 05:14:21 +0000 (0:00:00.362) 0:02:25.674 ********
ok: [sut] => {
"ansible_facts": {
"__podman_group_name": "root"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : See if getsubids exists] **************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:39
Wednesday 10 July 2024 05:14:21 +0000 (0:00:00.026) 0:02:25.700 ********
ok: [sut] => {
"changed": false,
"stat": {
"atime": 1720588427.5164855,
"attr_flags": "e",
"attributes": [
"extents"
],
"block_size": 4096,
"blocks": 32,
"charset": "binary",
"checksum": "72ba28a3be19c467e89ab90f5d21f99b605159a7",
"ctime": 1720588377.3007717,
"dev": 51714,
"device_type": 0,
"executable": true,
"exists": true,
"gid": 0,
"gr_name": "root",
"inode": 203603,
"isblk": false,
"ischr": false,
"isdir": false,
"isfifo": false,
"isgid": false,
"islnk": false,
"isreg": true,
"issock": false,
"isuid": false,
"mimetype": "application/x-pie-executable",
"mode": "0755",
"mtime": 1696291200.0,
"nlink": 1,
"path": "/usr/bin/getsubids",
"pw_name": "root",
"readable": true,
"rgrp": true,
"roth": true,
"rusr": true,
"size": 16032,
"uid": 0,
"version": "2092964490",
"wgrp": false,
"woth": false,
"writeable": true,
"wusr": true,
"xgrp": true,
"xoth": true,
"xusr": true
}
}
TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:50
Wednesday 10 July 2024 05:14:21 +0000 (0:00:00.399) 0:02:26.100 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_user not in [\"root\", \"0\"]",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Check group with getsubids] ***********
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:55
Wednesday 10 July 2024 05:14:21 +0000 (0:00:00.018) 0:02:26.118 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_user not in [\"root\", \"0\"]",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ******
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:60
Wednesday 10 July 2024 05:14:21 +0000 (0:00:00.018) 0:02:26.136 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_user not in [\"root\", \"0\"]",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Get subuid file] **********************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:74
Wednesday 10 July 2024 05:14:21 +0000 (0:00:00.017) 0:02:26.153 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_stat_getsubids.stat.exists",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Get subgid file] **********************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:79
Wednesday 10 July 2024 05:14:21 +0000 (0:00:00.017) 0:02:26.171 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_stat_getsubids.stat.exists",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ******
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:84
Wednesday 10 July 2024 05:14:21 +0000 (0:00:00.017) 0:02:26.188 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_stat_getsubids.stat.exists",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ******
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:94
Wednesday 10 July 2024 05:14:21 +0000 (0:00:00.017) 0:02:26.205 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_stat_getsubids.stat.exists",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] *****
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:101
Wednesday 10 July 2024 05:14:21 +0000 (0:00:00.017) 0:02:26.223 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_stat_getsubids.stat.exists",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62
Wednesday 10 July 2024 05:14:21 +0000 (0:00:00.016) 0:02:26.240 ********
ok: [sut] => {
"ansible_facts": {
"__podman_activate_systemd_unit": true,
"__podman_images_found": [],
"__podman_kube_yamls_raw": [
"quadlet-demo.yml"
],
"__podman_service_name": "quadlet-demo.service",
"__podman_systemd_scope": "system",
"__podman_user_home_dir": "/root",
"__podman_xdg_runtime_dir": "/run/user/0"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73
Wednesday 10 July 2024 05:14:21 +0000 (0:00:00.033) 0:02:26.274 ********
ok: [sut] => {
"ansible_facts": {
"__podman_quadlet_path": "/etc/containers/systemd"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Get kube yaml contents] ***************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77
Wednesday 10 July 2024 05:14:21 +0000 (0:00:00.019) 0:02:26.293 ********
ok: [sut] => {
"changed": false,
"content": "LS0tCmFwaVZlcnNpb246IHYxCmtpbmQ6IFBlcnNpc3RlbnRWb2x1bWVDbGFpbQptZXRhZGF0YToKICBuYW1lOiB3cC1wdi1jbGFpbQogIGxhYmVsczoKICAgIGFwcDogd29yZHByZXNzCnNwZWM6CiAgYWNjZXNzTW9kZXM6CiAgLSBSZWFkV3JpdGVPbmNlCiAgcmVzb3VyY2VzOgogICAgcmVxdWVzdHM6CiAgICAgIHN0b3JhZ2U6IDIwR2kKLS0tCmFwaVZlcnNpb246IHYxCmtpbmQ6IFBvZAptZXRhZGF0YToKICBuYW1lOiBxdWFkbGV0LWRlbW8Kc3BlYzoKICBjb250YWluZXJzOgogIC0gbmFtZTogd29yZHByZXNzCiAgICBpbWFnZTogcXVheS5pby9saW51eC1zeXN0ZW0tcm9sZXMvd29yZHByZXNzOjQuOC1hcGFjaGUKICAgIGVudjoKICAgIC0gbmFtZTogV09SRFBSRVNTX0RCX0hPU1QKICAgICAgdmFsdWU6IHF1YWRsZXQtZGVtby1teXNxbAogICAgLSBuYW1lOiBXT1JEUFJFU1NfREJfUEFTU1dPUkQKICAgICAgdmFsdWVGcm9tOgogICAgICAgIHNlY3JldEtleVJlZjoKICAgICAgICAgIG5hbWU6IG15c3FsLXJvb3QtcGFzc3dvcmQta3ViZQogICAgICAgICAga2V5OiBwYXNzd29yZAogICAgdm9sdW1lTW91bnRzOgogICAgLSBuYW1lOiB3b3JkcHJlc3MtcGVyc2lzdGVudC1zdG9yYWdlCiAgICAgIG1vdW50UGF0aDogL3Zhci93d3cvaHRtbAogICAgcmVzb3VyY2VzOgogICAgICByZXF1ZXN0czoKICAgICAgICBtZW1vcnk6ICI2NE1pIgogICAgICAgIGNwdTogIjI1MG0iCiAgICAgIGxpbWl0czoKICAgICAgICBtZW1vcnk6ICIxMjhNaSIKICAgICAgICBjcHU6ICI1MDBtIgogIC0gbmFtZTogZW52b3kKICAgIGltYWdlOiBxdWF5LmlvL2xpbnV4LXN5c3RlbS1yb2xlcy9lbnZveXByb3h5OnYxLjI1LjAKICAgIHZvbHVtZU1vdW50czoKICAgIC0gbmFtZTogY29uZmlnLXZvbHVtZQogICAgICBtb3VudFBhdGg6IC9ldGMvZW52b3kKICAgIC0gbmFtZTogY2VydGlmaWNhdGVzCiAgICAgIG1vdW50UGF0aDogL2V0Yy9lbnZveS1jZXJ0aWZpY2F0ZXMKICAgIGVudjoKICAgIC0gbmFtZTogRU5WT1lfVUlECiAgICAgIHZhbHVlOiAiMCIKICAgIHJlc291cmNlczoKICAgICAgcmVxdWVzdHM6CiAgICAgICAgbWVtb3J5OiAiNjRNaSIKICAgICAgICBjcHU6ICIyNTBtIgogICAgICBsaW1pdHM6CiAgICAgICAgbWVtb3J5OiAiMTI4TWkiCiAgICAgICAgY3B1OiAiNTAwbSIKICB2b2x1bWVzOgogIC0gbmFtZTogY29uZmlnLXZvbHVtZQogICAgY29uZmlnTWFwOgogICAgICBuYW1lOiBlbnZveS1wcm94eS1jb25maWcKICAtIG5hbWU6IGNlcnRpZmljYXRlcwogICAgc2VjcmV0OgogICAgICBzZWNyZXROYW1lOiBlbnZveS1jZXJ0aWZpY2F0ZXMKICAtIG5hbWU6IHdvcmRwcmVzcy1wZXJzaXN0ZW50LXN0b3JhZ2UKICAgIHBlcnNpc3RlbnRWb2x1bWVDbGFpbToKICAgICAgY2xhaW1OYW1lOiB3cC1wdi1jbGFpbQogIC0gbmFtZTogd3d3ICAjIG5vdCB1c2VkIC0gZm9yIHRlc3RpbmcgaG9zdHBhdGgKICAgIGhvc3RQYXRoOgogICAgICBwYXRoOiAvdG1wL2h0dHBkMwogIC0gbmFtZTogY3JlYXRlICAjIG5vdCB1c2VkIC0gZm9yIHRlc3RpbmcgaG9zdHBhdGgKICAgIGhvc3RQYXRoOgogICAgICBwYXRoOiAvdG1wL2h0dHBkMy1jcmVhdGUK",
"encoding": "base64",
"source": "/etc/containers/systemd/quadlet-demo.yml"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87
Wednesday 10 July 2024 05:14:21 +0000 (0:00:00.230) 0:02:26.524 ********
ok: [sut] => {
"ansible_facts": {
"__podman_images": [
"quay.io/linux-system-roles/wordpress:4.8-apache",
"quay.io/linux-system-roles/envoyproxy:v1.25.0"
],
"__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo.kube",
"__podman_volumes": [
"/tmp/httpd3",
"/tmp/httpd3-create"
]
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 6] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:103
Wednesday 10 July 2024 05:14:21 +0000 (0:00:00.054) 0:02:26.579 ********
ok: [sut] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : Cleanup quadlets] *********************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:110
Wednesday 10 July 2024 05:14:21 +0000 (0:00:00.022) 0:02:26.601 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_state == \"absent\"",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Create and update quadlets] ***********
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:114
Wednesday 10 July 2024 05:14:21 +0000 (0:00:00.018) 0:02:26.620 ********
included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml for sut
TASK [fedora.linux_system_roles.podman : Manage linger] ************************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:2
Wednesday 10 July 2024 05:14:21 +0000 (0:00:00.039) 0:02:26.659 ********
included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for sut
TASK [fedora.linux_system_roles.podman : Enable linger if needed] **************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12
Wednesday 10 July 2024 05:14:22 +0000 (0:00:00.030) 0:02:26.689 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_rootless | bool",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18
Wednesday 10 July 2024 05:14:22 +0000 (0:00:00.047) 0:02:26.736 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_rootless | bool",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22
Wednesday 10 July 2024 05:14:22 +0000 (0:00:00.018) 0:02:26.755 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_rootless | bool",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Create host directories] **************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:7
Wednesday 10 July 2024 05:14:22 +0000 (0:00:00.017) 0:02:26.772 ********
changed: [sut] => (item=/tmp/httpd3) => {
"ansible_loop_var": "item",
"changed": true,
"gid": 0,
"group": "root",
"item": "/tmp/httpd3",
"mode": "0755",
"owner": "root",
"path": "/tmp/httpd3",
"secontext": "unconfined_u:object_r:user_tmp_t:s0",
"size": 40,
"state": "directory",
"uid": 0
}
changed: [sut] => (item=/tmp/httpd3-create) => {
"ansible_loop_var": "item",
"changed": true,
"gid": 0,
"group": "root",
"item": "/tmp/httpd3-create",
"mode": "0755",
"owner": "root",
"path": "/tmp/httpd3-create",
"secontext": "unconfined_u:object_r:user_tmp_t:s0",
"size": 40,
"state": "directory",
"uid": 0
}
TASK [fedora.linux_system_roles.podman : Ensure container images are present] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:18
Wednesday 10 July 2024 05:14:22 +0000 (0:00:00.479) 0:02:27.252 ********
changed: [sut] => (item=None) => {
"attempts": 1,
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": true
}
changed: [sut] => (item=None) => {
"attempts": 1,
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": true
}
changed: [sut] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": true
}
TASK [fedora.linux_system_roles.podman : Ensure the quadlet directory is present] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:39
Wednesday 10 July 2024 05:14:40 +0000 (0:00:18.019) 0:02:45.271 ********
ok: [sut] => {
"changed": false,
"gid": 0,
"group": "root",
"mode": "0755",
"owner": "root",
"path": "/etc/containers/systemd",
"secontext": "system_u:object_r:etc_t:s0",
"size": 4096,
"state": "directory",
"uid": 0
}
TASK [fedora.linux_system_roles.podman : Ensure quadlet file is copied] ********
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:48
Wednesday 10 July 2024 05:14:40 +0000 (0:00:00.247) 0:02:45.519 ********
changed: [sut] => {
"changed": true,
"checksum": "7a5c73a5d935a42431c87bcdbeb8a04ed0909dc7",
"dest": "/etc/containers/systemd/quadlet-demo.kube",
"gid": 0,
"group": "root",
"md5sum": "da53c88f92b68b0487aa209f795b6bb3",
"mode": "0644",
"owner": "root",
"secontext": "system_u:object_r:etc_t:s0",
"size": 456,
"src": "/root/.ansible/tmp/ansible-tmp-1720588480.89213-12214-115703090519746/source",
"state": "file",
"uid": 0
}
TASK [fedora.linux_system_roles.podman : Ensure quadlet file content is present] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:58
Wednesday 10 July 2024 05:14:41 +0000 (0:00:00.584) 0:02:46.103 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_quadlet_file_src",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Ensure quadlet file is present] *******
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:70
Wednesday 10 July 2024 05:14:41 +0000 (0:00:00.020) 0:02:46.124 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_copy_file is skipped",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Reload systemctl] *********************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:82
Wednesday 10 July 2024 05:14:41 +0000 (0:00:00.017) 0:02:46.142 ********
ok: [sut] => {
"changed": false,
"name": null,
"status": {}
}
TASK [fedora.linux_system_roles.podman : Start service] ************************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:110
Wednesday 10 July 2024 05:14:42 +0000 (0:00:00.812) 0:02:46.954 ********
changed: [sut] => {
"changed": true,
"name": "quadlet-demo.service",
"state": "started",
"status": {
"AccessSELinuxContext": "system_u:object_r:systemd_unit_file_t:s0",
"ActiveEnterTimestampMonotonic": "0",
"ActiveExitTimestampMonotonic": "0",
"ActiveState": "inactive",
"After": "-.mount quadlet-demo-mysql.service basic.target quadlet-demo-network.service sysinit.target systemd-journald.socket system.slice",
"AllowIsolate": "no",
"AssertResult": "no",
"AssertTimestampMonotonic": "0",
"Before": "shutdown.target multi-user.target",
"BlockIOAccounting": "no",
"BlockIOWeight": "[not set]",
"CPUAccounting": "yes",
"CPUAffinityFromNUMA": "no",
"CPUQuotaPerSecUSec": "infinity",
"CPUQuotaPeriodUSec": "infinity",
"CPUSchedulingPolicy": "0",
"CPUSchedulingPriority": "0",
"CPUSchedulingResetOnFork": "no",
"CPUShares": "[not set]",
"CPUUsageNSec": "[not set]",
"CPUWeight": "[not set]",
"CacheDirectoryMode": "0755",
"CanFreeze": "yes",
"CanIsolate": "no",
"CanReload": "no",
"CanStart": "yes",
"CanStop": "yes",
"CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore",
"CleanResult": "success",
"CollectMode": "inactive",
"ConditionResult": "no",
"ConditionTimestampMonotonic": "0",
"ConfigurationDirectoryMode": "0755",
"Conflicts": "shutdown.target",
"ControlGroupId": "0",
"ControlPID": "0",
"CoredumpFilter": "0x33",
"DefaultDependencies": "yes",
"DefaultMemoryLow": "0",
"DefaultMemoryMin": "0",
"DefaultStartupMemoryLow": "0",
"Delegate": "no",
"Description": "quadlet-demo.service",
"DevicePolicy": "auto",
"DropInPaths": "/usr/lib/systemd/system/service.d/10-timeout-abort.conf",
"DynamicUser": "no",
"Environment": "PODMAN_SYSTEMD_UNIT=quadlet-demo.service",
"ExecMainCode": "0",
"ExecMainExitTimestampMonotonic": "0",
"ExecMainPID": "0",
"ExecMainStartTimestampMonotonic": "0",
"ExecMainStatus": "0",
"ExecStart": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman kube play --replace --service-container=true --network=systemd-quadlet-demo --configmap /etc/containers/systemd/envoy-proxy-configmap.yml --publish 8000:8080 --publish 9000:9901 /etc/containers/systemd/quadlet-demo.yml ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"ExecStartEx": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman kube play --replace --service-container=true --network=systemd-quadlet-demo --configmap /etc/containers/systemd/envoy-proxy-configmap.yml --publish 8000:8080 --publish 9000:9901 /etc/containers/systemd/quadlet-demo.yml ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"ExecStopPost": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman kube down /etc/containers/systemd/quadlet-demo.yml ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"ExecStopPostEx": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman kube down /etc/containers/systemd/quadlet-demo.yml ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"ExitType": "main",
"ExtensionImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent",
"FailureAction": "none",
"FileDescriptorStoreMax": "0",
"FileDescriptorStorePreserve": "restart",
"FinalKillSignal": "9",
"FragmentPath": "/run/systemd/generator/quadlet-demo.service",
"FreezerState": "running",
"GID": "[not set]",
"GuessMainPID": "yes",
"IOAccounting": "no",
"IOReadBytes": "18446744073709551615",
"IOReadOperations": "18446744073709551615",
"IOSchedulingClass": "2",
"IOSchedulingPriority": "4",
"IOWeight": "[not set]",
"IOWriteBytes": "18446744073709551615",
"IOWriteOperations": "18446744073709551615",
"IPAccounting": "no",
"IPEgressBytes": "[no data]",
"IPEgressPackets": "[no data]",
"IPIngressBytes": "[no data]",
"IPIngressPackets": "[no data]",
"Id": "quadlet-demo.service",
"IgnoreOnIsolate": "no",
"IgnoreSIGPIPE": "yes",
"InactiveEnterTimestampMonotonic": "0",
"InactiveExitTimestampMonotonic": "0",
"JobRunningTimeoutUSec": "infinity",
"JobTimeoutAction": "none",
"JobTimeoutUSec": "infinity",
"KeyringMode": "private",
"KillMode": "mixed",
"KillSignal": "15",
"LimitAS": "infinity",
"LimitASSoft": "infinity",
"LimitCORE": "infinity",
"LimitCORESoft": "infinity",
"LimitCPU": "infinity",
"LimitCPUSoft": "infinity",
"LimitDATA": "infinity",
"LimitDATASoft": "infinity",
"LimitFSIZE": "infinity",
"LimitFSIZESoft": "infinity",
"LimitLOCKS": "infinity",
"LimitLOCKSSoft": "infinity",
"LimitMEMLOCK": "8388608",
"LimitMEMLOCKSoft": "8388608",
"LimitMSGQUEUE": "819200",
"LimitMSGQUEUESoft": "819200",
"LimitNICE": "0",
"LimitNICESoft": "0",
"LimitNOFILE": "524288",
"LimitNOFILESoft": "1024",
"LimitNPROC": "14756",
"LimitNPROCSoft": "14756",
"LimitRSS": "infinity",
"LimitRSSSoft": "infinity",
"LimitRTPRIO": "0",
"LimitRTPRIOSoft": "0",
"LimitRTTIME": "infinity",
"LimitRTTIMESoft": "infinity",
"LimitSIGPENDING": "14756",
"LimitSIGPENDINGSoft": "14756",
"LimitSTACK": "infinity",
"LimitSTACKSoft": "8388608",
"LoadState": "loaded",
"LockPersonality": "no",
"LogLevelMax": "-1",
"LogRateLimitBurst": "0",
"LogRateLimitIntervalUSec": "0",
"LogsDirectoryMode": "0755",
"MainPID": "0",
"ManagedOOMMemoryPressure": "auto",
"ManagedOOMMemoryPressureLimit": "0",
"ManagedOOMPreference": "none",
"ManagedOOMSwap": "auto",
"MemoryAccounting": "yes",
"MemoryAvailable": "infinity",
"MemoryCurrent": "[not set]",
"MemoryDenyWriteExecute": "no",
"MemoryHigh": "infinity",
"MemoryKSM": "no",
"MemoryLimit": "infinity",
"MemoryLow": "0",
"MemoryMax": "infinity",
"MemoryMin": "0",
"MemoryPressureThresholdUSec": "200ms",
"MemoryPressureWatch": "auto",
"MemorySwapMax": "infinity",
"MemoryZSwapMax": "infinity",
"MountAPIVFS": "no",
"MountImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent",
"NFileDescriptorStore": "0",
"NRestarts": "0",
"NUMAPolicy": "n/a",
"Names": "quadlet-demo.service",
"NeedDaemonReload": "no",
"Nice": "0",
"NoNewPrivileges": "no",
"NonBlocking": "no",
"NotifyAccess": "all",
"OOMPolicy": "stop",
"OOMScoreAdjust": "0",
"OnFailureJobMode": "replace",
"OnSuccessJobMode": "fail",
"Perpetual": "no",
"PrivateDevices": "no",
"PrivateIPC": "no",
"PrivateMounts": "no",
"PrivateNetwork": "no",
"PrivateTmp": "no",
"PrivateUsers": "no",
"ProcSubset": "all",
"ProtectClock": "no",
"ProtectControlGroups": "no",
"ProtectHome": "no",
"ProtectHostname": "no",
"ProtectKernelLogs": "no",
"ProtectKernelModules": "no",
"ProtectKernelTunables": "no",
"ProtectProc": "default",
"ProtectSystem": "no",
"RefuseManualStart": "no",
"RefuseManualStop": "no",
"ReloadResult": "success",
"ReloadSignal": "1",
"RemainAfterExit": "no",
"RemoveIPC": "no",
"Requires": "sysinit.target system.slice -.mount quadlet-demo-mysql.service quadlet-demo-network.service",
"RequiresMountsFor": "/run/containers",
"Restart": "no",
"RestartKillSignal": "15",
"RestartMaxDelayUSec": "infinity",
"RestartMode": "normal",
"RestartSteps": "0",
"RestartUSec": "100ms",
"RestartUSecNext": "100ms",
"RestrictNamespaces": "no",
"RestrictRealtime": "no",
"RestrictSUIDSGID": "no",
"Result": "success",
"RootDirectoryStartOnly": "no",
"RootEphemeral": "no",
"RootImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent",
"RuntimeDirectoryMode": "0755",
"RuntimeDirectoryPreserve": "no",
"RuntimeMaxUSec": "infinity",
"RuntimeRandomizedExtraUSec": "0",
"SameProcessGroup": "no",
"SecureBits": "0",
"SendSIGHUP": "no",
"SendSIGKILL": "yes",
"Slice": "system.slice",
"SourcePath": "/etc/containers/systemd/quadlet-demo.kube",
"StandardError": "inherit",
"StandardInput": "null",
"StandardOutput": "journal",
"StartLimitAction": "none",
"StartLimitBurst": "5",
"StartLimitIntervalUSec": "10s",
"StartupBlockIOWeight": "[not set]",
"StartupCPUShares": "[not set]",
"StartupCPUWeight": "[not set]",
"StartupIOWeight": "[not set]",
"StartupMemoryHigh": "18446744073709551615",
"StartupMemoryLow": "0",
"StartupMemoryMax": "18446744073709551615",
"StartupMemorySwapMax": "18446744073709551615",
"StartupMemoryZSwapMax": "18446744073709551615",
"StateChangeTimestampMonotonic": "0",
"StateDirectoryMode": "0755",
"StatusErrno": "0",
"StopWhenUnneeded": "no",
"SubState": "dead",
"SuccessAction": "none",
"SyslogFacility": "3",
"SyslogIdentifier": "quadlet-demo",
"SyslogLevel": "6",
"SyslogLevelPrefix": "yes",
"SyslogPriority": "30",
"SystemCallErrorNumber": "2147483646",
"TTYReset": "no",
"TTYVHangup": "no",
"TTYVTDisallocate": "no",
"TasksAccounting": "yes",
"TasksCurrent": "[not set]",
"TasksMax": "4426",
"TimeoutAbortUSec": "45s",
"TimeoutCleanUSec": "infinity",
"TimeoutStartFailureMode": "terminate",
"TimeoutStartUSec": "45s",
"TimeoutStopFailureMode": "abort",
"TimeoutStopUSec": "45s",
"TimerSlackNSec": "50000",
"Transient": "no",
"Type": "notify",
"UID": "[not set]",
"UMask": "0022",
"UnitFilePreset": "disabled",
"UnitFileState": "generated",
"UtmpMode": "init",
"WantedBy": "multi-user.target",
"WatchdogSignal": "6",
"WatchdogTimestampMonotonic": "0",
"WatchdogUSec": "infinity"
}
}
TASK [fedora.linux_system_roles.podman : Restart service] **********************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:125
Wednesday 10 July 2024 05:14:43 +0000 (0:00:01.335) 0:02:48.290 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_service_started is changed",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Cancel linger] ************************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:143
Wednesday 10 July 2024 05:14:43 +0000 (0:00:00.019) 0:02:48.309 ********
skipping: [sut] => {
"changed": false,
"skipped_reason": "No items in the list"
}
TASK [fedora.linux_system_roles.podman : Handle credential files - absent] *****
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:149
Wednesday 10 July 2024 05:14:43 +0000 (0:00:00.015) 0:02:48.325 ********
skipping: [sut] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : Handle certs.d files - absent] ********
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:158
Wednesday 10 July 2024 05:14:43 +0000 (0:00:00.015) 0:02:48.341 ********
skipping: [sut] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [Check quadlet files] *****************************************************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/tests/tests_quadlet_demo.yml:84
Wednesday 10 July 2024 05:14:43 +0000 (0:00:00.025) 0:02:48.366 ********
ok: [sut] => {
"changed": false,
"cmd": [
"ls",
"-alrtF",
"/etc/containers/systemd"
],
"delta": "0:00:00.006798",
"end": "2024-07-10 05:14:43.956628",
"rc": 0,
"start": "2024-07-10 05:14:43.949830"
}
STDOUT:
total 32
drwxr-xr-x. 8 root root 4096 Jul 10 05:13 ../
-rw-r--r--. 1 root root 74 Jul 10 05:13 quadlet-demo.network
-rw-r--r--. 1 root root 9 Jul 10 05:14 quadlet-demo-mysql.volume
-rw-r--r--. 1 root root 367 Jul 10 05:14 quadlet-demo-mysql.container
-rw-r--r--. 1 root root 2102 Jul 10 05:14 envoy-proxy-configmap.yml
-rw-r--r--. 1 root root 1605 Jul 10 05:14 quadlet-demo.yml
-rw-r--r--. 1 root root 456 Jul 10 05:14 quadlet-demo.kube
drwxr-xr-x. 2 root root 4096 Jul 10 05:14 ./
TASK [Check containers] ********************************************************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/tests/tests_quadlet_demo.yml:88
Wednesday 10 July 2024 05:14:43 +0000 (0:00:00.297) 0:02:48.663 ********
ok: [sut] => {
"changed": false,
"cmd": [
"podman",
"ps",
"-a"
],
"delta": "0:00:00.063239",
"end": "2024-07-10 05:14:44.294651",
"failed_when_result": false,
"rc": 0,
"start": "2024-07-10 05:14:44.231412"
}
STDOUT:
CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES
21b1326f5785 quay.io/linux-system-roles/mysql:5.6 mysqld 31 seconds ago Up 31 seconds (starting) quadlet-demo-mysql
9de2222e9853 localhost/podman-pause:4.9.4-1711445992 1 second ago Up 1 second a96f3a51b8d1-service
ee4bc05257fc localhost/podman-pause:4.9.4-1711445992 1 second ago Up 1 second 0.0.0.0:8000->8080/tcp, 0.0.0.0:9000->9901/tcp da66443b0d18-infra
af5eaff28e9f quay.io/linux-system-roles/wordpress:4.8-apache apache2-foregroun... 1 second ago Up 1 second 0.0.0.0:8000->8080/tcp, 0.0.0.0:9000->9901/tcp quadlet-demo-wordpress
dce8c2f6a9a9 quay.io/linux-system-roles/envoyproxy:v1.25.0 envoy -c /etc/env... 1 second ago Up 1 second 0.0.0.0:8000->8080/tcp, 0.0.0.0:9000->9901/tcp quadlet-demo-envoy
TASK [Check pods] **************************************************************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/tests/tests_quadlet_demo.yml:93
Wednesday 10 July 2024 05:14:44 +0000 (0:00:00.330) 0:02:48.994 ********
ok: [sut] => {
"changed": false,
"cmd": [
"podman",
"pod",
"ps",
"--ctr-ids",
"--ctr-names",
"--ctr-status"
],
"delta": "0:00:00.046308",
"end": "2024-07-10 05:14:44.616406",
"failed_when_result": false,
"rc": 0,
"start": "2024-07-10 05:14:44.570098"
}
STDOUT:
POD ID NAME STATUS CREATED INFRA ID IDS NAMES STATUS
da66443b0d18 quadlet-demo Running 1 second ago ee4bc05257fc ee4bc05257fc,af5eaff28e9f,dce8c2f6a9a9 da66443b0d18-infra,quadlet-demo-wordpress,quadlet-demo-envoy running,running,running
TASK [Check systemd] ***********************************************************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/tests/tests_quadlet_demo.yml:98
Wednesday 10 July 2024 05:14:44 +0000 (0:00:00.319) 0:02:49.314 ********
ok: [sut] => {
"changed": false,
"cmd": "set -euo pipefail; systemctl list-units | grep quadlet",
"delta": "0:00:00.014096",
"end": "2024-07-10 05:14:44.900233",
"failed_when_result": false,
"rc": 0,
"start": "2024-07-10 05:14:44.886137"
}
STDOUT:
quadlet-demo-mysql-volume.service loaded active exited quadlet-demo-mysql-volume.service
quadlet-demo-mysql.service loaded active running quadlet-demo-mysql.service
quadlet-demo-network.service loaded active exited quadlet-demo-network.service
quadlet-demo.service loaded active running quadlet-demo.service
TASK [Check web] ***************************************************************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/tests/tests_quadlet_demo.yml:104
Wednesday 10 July 2024 05:14:44 +0000 (0:00:00.315) 0:02:49.630 ********
changed: [sut] => {
"attempts": 1,
"changed": true,
"checksum_dest": null,
"checksum_src": "d49966cc24f102ead5187c35b5a35440048247ef",
"dest": "/run/out",
"elapsed": 0,
"gid": 0,
"group": "root",
"md5sum": "a56cfe4431509ef44591163d2a219089",
"mode": "0600",
"owner": "root",
"secontext": "system_u:object_r:var_run_t:s0",
"size": 11666,
"src": "/root/.ansible/tmp/ansible-moduletmp-1720588485.3553376-550v36b3/tmp_gz0ldug",
"state": "file",
"status_code": 200,
"uid": 0,
"url": "https://localhost:8000"
}
MSG:
OK (unknown bytes)
TASK [Show web] ****************************************************************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/tests/tests_quadlet_demo.yml:115
Wednesday 10 July 2024 05:14:46 +0000 (0:00:01.112) 0:02:50.742 ********
ok: [sut] => {
"changed": false,
"cmd": [
"cat",
"/run/out"
],
"delta": "0:00:00.003218",
"end": "2024-07-10 05:14:46.277916",
"rc": 0,
"start": "2024-07-10 05:14:46.274698"
}
STDOUT:
WordPress › Installation
WordPress
TASK [Error] *******************************************************************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/tests/tests_quadlet_demo.yml:120
Wednesday 10 July 2024 05:14:46 +0000 (0:00:00.233) 0:02:50.976 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__web_status is failed",
"skip_reason": "Conditional result was False"
}
TASK [Check] *******************************************************************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/tests/tests_quadlet_demo.yml:131
Wednesday 10 July 2024 05:14:46 +0000 (0:00:00.017) 0:02:50.994 ********
ok: [sut] => {
"changed": false,
"cmd": [
"podman",
"ps",
"-a"
],
"delta": "0:00:00.039210",
"end": "2024-07-10 05:14:46.563898",
"rc": 0,
"start": "2024-07-10 05:14:46.524688"
}
STDOUT:
CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES
21b1326f5785 quay.io/linux-system-roles/mysql:5.6 mysqld 33 seconds ago Up 33 seconds (starting) quadlet-demo-mysql
9de2222e9853 localhost/podman-pause:4.9.4-1711445992 3 seconds ago Up 3 seconds a96f3a51b8d1-service
ee4bc05257fc localhost/podman-pause:4.9.4-1711445992 3 seconds ago Up 3 seconds 0.0.0.0:8000->8080/tcp, 0.0.0.0:9000->9901/tcp da66443b0d18-infra
af5eaff28e9f quay.io/linux-system-roles/wordpress:4.8-apache apache2-foregroun... 3 seconds ago Up 3 seconds 0.0.0.0:8000->8080/tcp, 0.0.0.0:9000->9901/tcp quadlet-demo-wordpress
dce8c2f6a9a9 quay.io/linux-system-roles/envoyproxy:v1.25.0 envoy -c /etc/env... 3 seconds ago Up 3 seconds 0.0.0.0:8000->8080/tcp, 0.0.0.0:9000->9901/tcp quadlet-demo-envoy
TASK [Check pods] **************************************************************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/tests/tests_quadlet_demo.yml:135
Wednesday 10 July 2024 05:14:46 +0000 (0:00:00.267) 0:02:51.261 ********
ok: [sut] => {
"changed": false,
"cmd": [
"podman",
"pod",
"ps",
"--ctr-ids",
"--ctr-names",
"--ctr-status"
],
"delta": "0:00:00.040209",
"end": "2024-07-10 05:14:46.830078",
"failed_when_result": false,
"rc": 0,
"start": "2024-07-10 05:14:46.789869"
}
STDOUT:
POD ID NAME STATUS CREATED INFRA ID IDS NAMES STATUS
da66443b0d18 quadlet-demo Running 3 seconds ago ee4bc05257fc ee4bc05257fc,af5eaff28e9f,dce8c2f6a9a9 da66443b0d18-infra,quadlet-demo-wordpress,quadlet-demo-envoy running,running,running
TASK [Check systemd] ***********************************************************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/tests/tests_quadlet_demo.yml:140
Wednesday 10 July 2024 05:14:46 +0000 (0:00:00.269) 0:02:51.530 ********
ok: [sut] => {
"changed": false,
"cmd": "set -euo pipefail; systemctl list-units --all | grep quadlet",
"delta": "0:00:00.014630",
"end": "2024-07-10 05:14:47.077049",
"failed_when_result": false,
"rc": 0,
"start": "2024-07-10 05:14:47.062419"
}
STDOUT:
quadlet-demo-mysql-volume.service loaded active exited quadlet-demo-mysql-volume.service
quadlet-demo-mysql.service loaded active running quadlet-demo-mysql.service
quadlet-demo-network.service loaded active exited quadlet-demo-network.service
quadlet-demo.service loaded active running quadlet-demo.service
TASK [LS] **********************************************************************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/tests/tests_quadlet_demo.yml:148
Wednesday 10 July 2024 05:14:47 +0000 (0:00:00.245) 0:02:51.775 ********
ok: [sut] => {
"changed": false,
"cmd": [
"ls",
"-alrtF",
"/etc/systemd/system"
],
"delta": "0:00:00.004008",
"end": "2024-07-10 05:14:47.311045",
"failed_when_result": false,
"rc": 0,
"start": "2024-07-10 05:14:47.307037"
}
STDOUT:
total 56
lrwxrwxrwx. 1 root root 48 Jun 20 10:26 dbus-org.freedesktop.resolve1.service -> /usr/lib/systemd/system/systemd-resolved.service
lrwxrwxrwx. 1 root root 44 Jun 20 10:26 dbus-org.freedesktop.oom1.service -> /usr/lib/systemd/system/systemd-oomd.service
drwxr-xr-x. 2 root root 4096 Jun 20 10:26 systemd-journald.service.wants/
lrwxrwxrwx. 1 root root 43 Jun 20 10:26 dbus.service -> /usr/lib/systemd/system/dbus-broker.service
lrwxrwxrwx. 1 root root 37 Jun 20 10:26 ctrl-alt-del.target -> /usr/lib/systemd/system/reboot.target
drwxr-xr-x. 2 root root 4096 Jun 20 10:26 getty.target.wants/
drwxr-xr-x. 5 root root 4096 Jun 20 10:26 ../
lrwxrwxrwx. 1 root root 45 Jun 20 10:26 dbus-org.freedesktop.home1.service -> /usr/lib/systemd/system/systemd-homed.service
drwxr-xr-x. 2 root root 4096 Jun 20 10:26 systemd-homed.service.wants/
lrwxrwxrwx. 1 root root 41 Jun 20 10:26 dbus-org.bluez.service -> /usr/lib/systemd/system/bluetooth.service
drwxr-xr-x. 2 root root 4096 Jun 20 10:26 bluetooth.target.wants/
drwxr-xr-x. 2 root root 4096 Jun 20 10:26 graphical.target.wants/
drwxr-xr-x. 2 root root 4096 Jun 20 10:27 timers.target.wants/
lrwxrwxrwx. 1 root root 57 Jun 20 10:27 dbus-org.freedesktop.nm-dispatcher.service -> /usr/lib/systemd/system/NetworkManager-dispatcher.service
drwxr-xr-x. 2 root root 4096 Jun 20 10:27 network-online.target.wants/
lrwxrwxrwx. 1 root root 41 Jun 20 10:31 default.target -> /usr/lib/systemd/system/multi-user.target
drwxr-xr-x. 2 root root 4096 Jun 20 10:45 remote-fs.target.wants/
drwxr-xr-x. 2 root root 4096 Jun 20 10:46 cloud-init.target.wants/
drwxr-xr-x. 2 root root 4096 Jun 20 10:47 sockets.target.wants/
drwxr-xr-x. 2 root root 4096 Jul 10 05:02 sysinit.target.wants/
drwxr-xr-x. 2 root root 4096 Jul 10 05:13 multi-user.target.wants/
lrwxrwxrwx. 1 root root 41 Jul 10 05:13 dbus-org.fedoraproject.FirewallD1.service -> /usr/lib/systemd/system/firewalld.service
drwxr-xr-x. 14 root root 4096 Jul 10 05:13 ./
TASK [Cleanup] *****************************************************************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/tests/tests_quadlet_demo.yml:155
Wednesday 10 July 2024 05:14:47 +0000 (0:00:00.235) 0:02:52.010 ********
TASK [fedora.linux_system_roles.podman : Set platform/version specific variables] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:3
Wednesday 10 July 2024 05:14:47 +0000 (0:00:00.042) 0:02:52.053 ********
included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml for sut
TASK [fedora.linux_system_roles.podman : Ensure ansible_facts used by role] ****
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:3
Wednesday 10 July 2024 05:14:47 +0000 (0:00:00.032) 0:02:52.086 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_required_facts | difference(ansible_facts.keys() | list) | length > 0",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Check if system is ostree] ************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:11
Wednesday 10 July 2024 05:14:47 +0000 (0:00:00.023) 0:02:52.109 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_is_ostree is defined",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set flag to indicate system is ostree] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:16
Wednesday 10 July 2024 05:14:47 +0000 (0:00:00.018) 0:02:52.128 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_is_ostree is defined",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set platform/version specific variables] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:20
Wednesday 10 July 2024 05:14:47 +0000 (0:00:00.018) 0:02:52.146 ********
ok: [sut] => (item=RedHat.yml) => {
"ansible_facts": {
"__podman_packages": [
"podman",
"shadow-utils-subid"
]
},
"ansible_included_var_files": [
"/WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/vars/RedHat.yml"
],
"ansible_loop_var": "item",
"changed": false,
"item": "RedHat.yml"
}
ok: [sut] => (item=Fedora.yml) => {
"ansible_facts": {
"__podman_packages": [
"iptables-nft",
"podman",
"shadow-utils-subid"
]
},
"ansible_included_var_files": [
"/WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/vars/Fedora.yml"
],
"ansible_loop_var": "item",
"changed": false,
"item": "Fedora.yml"
}
skipping: [sut] => (item=Fedora_39.yml) => {
"ansible_loop_var": "item",
"changed": false,
"false_condition": "__vars_file is file",
"item": "Fedora_39.yml",
"skip_reason": "Conditional result was False"
}
skipping: [sut] => (item=Fedora_39.yml) => {
"ansible_loop_var": "item",
"changed": false,
"false_condition": "__vars_file is file",
"item": "Fedora_39.yml",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Gather the package facts] *************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:6
Wednesday 10 July 2024 05:14:47 +0000 (0:00:00.041) 0:02:52.188 ********
ok: [sut] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : Enable copr if requested] *************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:10
Wednesday 10 July 2024 05:14:48 +0000 (0:00:00.841) 0:02:53.029 ********
skipping: [sut] => {
"changed": false,
"false_condition": "podman_use_copr | d(false)",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Ensure required packages are installed] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:14
Wednesday 10 July 2024 05:14:48 +0000 (0:00:00.020) 0:02:53.049 ********
skipping: [sut] => {
"changed": false,
"false_condition": "(__podman_packages | difference(ansible_facts.packages))",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Get podman version] *******************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:22
Wednesday 10 July 2024 05:14:48 +0000 (0:00:00.022) 0:02:53.072 ********
ok: [sut] => {
"changed": false,
"cmd": [
"podman",
"--version"
],
"delta": "0:00:00.029344",
"end": "2024-07-10 05:14:48.632335",
"rc": 0,
"start": "2024-07-10 05:14:48.602991"
}
STDOUT:
podman version 4.9.4
TASK [fedora.linux_system_roles.podman : Set podman version] *******************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:28
Wednesday 10 July 2024 05:14:48 +0000 (0:00:00.259) 0:02:53.331 ********
ok: [sut] => {
"ansible_facts": {
"podman_version": "4.9.4"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Podman package version must be 4.2 or later] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:32
Wednesday 10 July 2024 05:14:48 +0000 (0:00:00.022) 0:02:53.354 ********
skipping: [sut] => {
"changed": false,
"false_condition": "podman_version is version(\"4.2\", \"<\")",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Podman package version must be 4.4 or later for quadlet, secrets] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:39
Wednesday 10 July 2024 05:14:48 +0000 (0:00:00.018) 0:02:53.372 ********
skipping: [sut] => {
"changed": false,
"false_condition": "podman_version is version(\"4.4\", \"<\")",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Podman package version must be 4.4 or later for quadlet, secrets] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:49
Wednesday 10 July 2024 05:14:48 +0000 (0:00:00.023) 0:02:53.395 ********
META: end_host conditional evaluated to False, continuing execution for sut
skipping: [sut] => {
"skip_reason": "end_host conditional evaluated to False, continuing execution for sut"
}
MSG:
end_host conditional evaluated to false, continuing execution for sut
TASK [fedora.linux_system_roles.podman : Check user and group information] *****
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:56
Wednesday 10 July 2024 05:14:48 +0000 (0:00:00.025) 0:02:53.421 ********
included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for sut
TASK [fedora.linux_system_roles.podman : Get user information] *****************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2
Wednesday 10 July 2024 05:14:48 +0000 (0:00:00.040) 0:02:53.461 ********
skipping: [sut] => {
"changed": false,
"false_condition": "'getent_passwd' not in ansible_facts or __podman_user not in ansible_facts['getent_passwd']",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Fail if user does not exist] **********
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9
Wednesday 10 July 2024 05:14:48 +0000 (0:00:00.023) 0:02:53.484 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not ansible_facts[\"getent_passwd\"][__podman_user]",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set group for podman user] ************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16
Wednesday 10 July 2024 05:14:48 +0000 (0:00:00.022) 0:02:53.507 ********
ok: [sut] => {
"ansible_facts": {
"__podman_group": "0"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Get group information] ****************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:28
Wednesday 10 July 2024 05:14:48 +0000 (0:00:00.027) 0:02:53.534 ********
ok: [sut] => {
"ansible_facts": {
"getent_group": {
"root": [
"x",
"0",
""
]
}
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set group name] ***********************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:35
Wednesday 10 July 2024 05:14:49 +0000 (0:00:00.284) 0:02:53.819 ********
ok: [sut] => {
"ansible_facts": {
"__podman_group_name": "root"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : See if getsubids exists] **************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:39
Wednesday 10 July 2024 05:14:49 +0000 (0:00:00.026) 0:02:53.845 ********
ok: [sut] => {
"changed": false,
"stat": {
"atime": 1720588427.5164855,
"attr_flags": "e",
"attributes": [
"extents"
],
"block_size": 4096,
"blocks": 32,
"charset": "binary",
"checksum": "72ba28a3be19c467e89ab90f5d21f99b605159a7",
"ctime": 1720588377.3007717,
"dev": 51714,
"device_type": 0,
"executable": true,
"exists": true,
"gid": 0,
"gr_name": "root",
"inode": 203603,
"isblk": false,
"ischr": false,
"isdir": false,
"isfifo": false,
"isgid": false,
"islnk": false,
"isreg": true,
"issock": false,
"isuid": false,
"mimetype": "application/x-pie-executable",
"mode": "0755",
"mtime": 1696291200.0,
"nlink": 1,
"path": "/usr/bin/getsubids",
"pw_name": "root",
"readable": true,
"rgrp": true,
"roth": true,
"rusr": true,
"size": 16032,
"uid": 0,
"version": "2092964490",
"wgrp": false,
"woth": false,
"writeable": true,
"wusr": true,
"xgrp": true,
"xoth": true,
"xusr": true
}
}
TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:50
Wednesday 10 July 2024 05:14:49 +0000 (0:00:00.239) 0:02:54.084 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_user not in [\"root\", \"0\"]",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Check group with getsubids] ***********
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:55
Wednesday 10 July 2024 05:14:49 +0000 (0:00:00.019) 0:02:54.103 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_user not in [\"root\", \"0\"]",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ******
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:60
Wednesday 10 July 2024 05:14:49 +0000 (0:00:00.052) 0:02:54.156 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_user not in [\"root\", \"0\"]",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Get subuid file] **********************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:74
Wednesday 10 July 2024 05:14:49 +0000 (0:00:00.020) 0:02:54.176 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_stat_getsubids.stat.exists",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Get subgid file] **********************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:79
Wednesday 10 July 2024 05:14:49 +0000 (0:00:00.017) 0:02:54.194 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_stat_getsubids.stat.exists",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ******
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:84
Wednesday 10 July 2024 05:14:49 +0000 (0:00:00.018) 0:02:54.212 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_stat_getsubids.stat.exists",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ******
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:94
Wednesday 10 July 2024 05:14:49 +0000 (0:00:00.017) 0:02:54.230 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_stat_getsubids.stat.exists",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] *****
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:101
Wednesday 10 July 2024 05:14:49 +0000 (0:00:00.018) 0:02:54.248 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_stat_getsubids.stat.exists",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set config file paths] ****************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:62
Wednesday 10 July 2024 05:14:49 +0000 (0:00:00.018) 0:02:54.266 ********
ok: [sut] => {
"ansible_facts": {
"__podman_container_conf_file": "/etc/containers/containers.conf.d/50-systemroles.conf",
"__podman_policy_json_file": "/etc/containers/policy.json",
"__podman_registries_conf_file": "/etc/containers/registries.conf.d/50-systemroles.conf",
"__podman_storage_conf_file": "/etc/containers/storage.conf"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Handle container.conf.d] **************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:71
Wednesday 10 July 2024 05:14:49 +0000 (0:00:00.024) 0:02:54.291 ********
included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_container_conf_d.yml for sut
TASK [fedora.linux_system_roles.podman : Ensure containers.d exists] ***********
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_container_conf_d.yml:5
Wednesday 10 July 2024 05:14:49 +0000 (0:00:00.037) 0:02:54.328 ********
skipping: [sut] => {
"changed": false,
"false_condition": "podman_containers_conf | length > 0",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Update container config file] *********
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_container_conf_d.yml:13
Wednesday 10 July 2024 05:14:49 +0000 (0:00:00.018) 0:02:54.347 ********
skipping: [sut] => {
"changed": false,
"false_condition": "podman_containers_conf | length > 0",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Handle registries.conf.d] *************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:74
Wednesday 10 July 2024 05:14:49 +0000 (0:00:00.019) 0:02:54.367 ********
included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_registries_conf_d.yml for sut
TASK [fedora.linux_system_roles.podman : Ensure registries.d exists] ***********
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_registries_conf_d.yml:5
Wednesday 10 July 2024 05:14:49 +0000 (0:00:00.037) 0:02:54.404 ********
skipping: [sut] => {
"changed": false,
"false_condition": "podman_registries_conf | length > 0",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Update registries config file] ********
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_registries_conf_d.yml:13
Wednesday 10 July 2024 05:14:49 +0000 (0:00:00.019) 0:02:54.423 ********
skipping: [sut] => {
"changed": false,
"false_condition": "podman_registries_conf | length > 0",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Handle storage.conf] ******************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:77
Wednesday 10 July 2024 05:14:49 +0000 (0:00:00.018) 0:02:54.442 ********
included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_storage_conf.yml for sut
TASK [fedora.linux_system_roles.podman : Ensure storage.conf parent dir exists] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_storage_conf.yml:5
Wednesday 10 July 2024 05:14:49 +0000 (0:00:00.038) 0:02:54.480 ********
skipping: [sut] => {
"changed": false,
"false_condition": "podman_storage_conf | length > 0",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Update storage config file] ***********
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_storage_conf.yml:13
Wednesday 10 July 2024 05:14:49 +0000 (0:00:00.048) 0:02:54.528 ********
skipping: [sut] => {
"changed": false,
"false_condition": "podman_storage_conf | length > 0",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Handle policy.json] *******************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:80
Wednesday 10 July 2024 05:14:49 +0000 (0:00:00.020) 0:02:54.549 ********
included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml for sut
TASK [fedora.linux_system_roles.podman : Ensure policy.json parent dir exists] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:6
Wednesday 10 July 2024 05:14:49 +0000 (0:00:00.040) 0:02:54.590 ********
skipping: [sut] => {
"changed": false,
"false_condition": "podman_policy_json | length > 0",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Stat the policy.json file] ************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:14
Wednesday 10 July 2024 05:14:49 +0000 (0:00:00.020) 0:02:54.611 ********
skipping: [sut] => {
"changed": false,
"false_condition": "podman_policy_json | length > 0",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Get the existing policy.json] *********
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:19
Wednesday 10 July 2024 05:14:49 +0000 (0:00:00.019) 0:02:54.630 ********
skipping: [sut] => {
"changed": false,
"false_condition": "podman_policy_json | length > 0",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Write new policy.json file] ***********
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:25
Wednesday 10 July 2024 05:14:49 +0000 (0:00:00.019) 0:02:54.650 ********
skipping: [sut] => {
"changed": false,
"false_condition": "podman_policy_json | length > 0",
"skip_reason": "Conditional result was False"
}
TASK [Manage firewall for specified ports] *************************************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:86
Wednesday 10 July 2024 05:14:50 +0000 (0:00:00.018) 0:02:54.668 ********
TASK [fedora.linux_system_roles.firewall : Setup firewalld] ********************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:2
Wednesday 10 July 2024 05:14:50 +0000 (0:00:00.065) 0:02:54.734 ********
included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml for sut
TASK [fedora.linux_system_roles.firewall : Ensure ansible_facts used by role] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:2
Wednesday 10 July 2024 05:14:50 +0000 (0:00:00.032) 0:02:54.766 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__firewall_required_facts | difference(ansible_facts.keys() | list) | length > 0",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.firewall : Check if system is ostree] **********
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:10
Wednesday 10 July 2024 05:14:50 +0000 (0:00:00.025) 0:02:54.791 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __firewall_is_ostree is defined",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.firewall : Set flag to indicate system is ostree] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:15
Wednesday 10 July 2024 05:14:50 +0000 (0:00:00.019) 0:02:54.810 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __firewall_is_ostree is defined",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.firewall : Install firewalld] ******************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:19
Wednesday 10 July 2024 05:14:50 +0000 (0:00:00.018) 0:02:54.829 ********
ok: [sut] => {
"changed": false,
"rc": 0,
"results": []
}
MSG:
Nothing to do
TASK [fedora.linux_system_roles.firewall : Collect service facts] **************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:5
Wednesday 10 July 2024 05:14:52 +0000 (0:00:02.286) 0:02:57.116 ********
skipping: [sut] => {
"changed": false,
"false_condition": "firewall_disable_conflicting_services | bool",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.firewall : Attempt to stop and disable conflicting services] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:9
Wednesday 10 July 2024 05:14:52 +0000 (0:00:00.018) 0:02:57.134 ********
skipping: [sut] => (item=nftables) => {
"ansible_loop_var": "item",
"changed": false,
"false_condition": "firewall_disable_conflicting_services | bool",
"item": "nftables",
"skip_reason": "Conditional result was False"
}
skipping: [sut] => (item=iptables) => {
"ansible_loop_var": "item",
"changed": false,
"false_condition": "firewall_disable_conflicting_services | bool",
"item": "iptables",
"skip_reason": "Conditional result was False"
}
skipping: [sut] => (item=ufw) => {
"ansible_loop_var": "item",
"changed": false,
"false_condition": "firewall_disable_conflicting_services | bool",
"item": "ufw",
"skip_reason": "Conditional result was False"
}
skipping: [sut] => {
"changed": false
}
MSG:
All items skipped
TASK [fedora.linux_system_roles.firewall : Unmask firewalld service] ***********
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:22
Wednesday 10 July 2024 05:14:52 +0000 (0:00:00.055) 0:02:57.189 ********
ok: [sut] => {
"changed": false,
"name": "firewalld",
"status": {
"AccessSELinuxContext": "system_u:object_r:firewalld_unit_file_t:s0",
"ActiveEnterTimestamp": "Wed 2024-07-10 05:13:52 UTC",
"ActiveEnterTimestampMonotonic": "952448512",
"ActiveExitTimestampMonotonic": "0",
"ActiveState": "active",
"After": "basic.target polkit.service dbus.socket system.slice dbus-broker.service sysinit.target",
"AllowIsolate": "no",
"AssertResult": "yes",
"AssertTimestamp": "Wed 2024-07-10 05:13:51 UTC",
"AssertTimestampMonotonic": "952190941",
"Before": "shutdown.target network-pre.target multi-user.target",
"BlockIOAccounting": "no",
"BlockIOWeight": "[not set]",
"BusName": "org.fedoraproject.FirewallD1",
"CPUAccounting": "yes",
"CPUAffinityFromNUMA": "no",
"CPUQuotaPerSecUSec": "infinity",
"CPUQuotaPeriodUSec": "infinity",
"CPUSchedulingPolicy": "0",
"CPUSchedulingPriority": "0",
"CPUSchedulingResetOnFork": "no",
"CPUShares": "[not set]",
"CPUUsageNSec": "589336000",
"CPUWeight": "[not set]",
"CacheDirectoryMode": "0755",
"CanFreeze": "yes",
"CanIsolate": "no",
"CanReload": "yes",
"CanStart": "yes",
"CanStop": "yes",
"CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore",
"CleanResult": "success",
"CollectMode": "inactive",
"ConditionResult": "yes",
"ConditionTimestamp": "Wed 2024-07-10 05:13:51 UTC",
"ConditionTimestampMonotonic": "952190937",
"ConfigurationDirectoryMode": "0755",
"Conflicts": "shutdown.target iptables.service nftables.service ebtables.service ip6tables.service ipset.service",
"ControlGroup": "/system.slice/firewalld.service",
"ControlGroupId": "6620",
"ControlPID": "0",
"CoredumpFilter": "0x33",
"DefaultDependencies": "yes",
"DefaultMemoryLow": "0",
"DefaultMemoryMin": "0",
"DefaultStartupMemoryLow": "0",
"Delegate": "no",
"Description": "firewalld - dynamic firewall daemon",
"DevicePolicy": "auto",
"Documentation": "\"man:firewalld(1)\"",
"DropInPaths": "/usr/lib/systemd/system/service.d/10-timeout-abort.conf",
"DynamicUser": "no",
"EffectiveCPUs": "0-1",
"EffectiveMemoryNodes": "0",
"EnvironmentFiles": "/etc/sysconfig/firewalld (ignore_errors=yes)",
"ExecMainCode": "0",
"ExecMainExitTimestampMonotonic": "0",
"ExecMainPID": "6763",
"ExecMainStartTimestamp": "Wed 2024-07-10 05:13:51 UTC",
"ExecMainStartTimestampMonotonic": "952204109",
"ExecMainStatus": "0",
"ExecReload": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"ExecReloadEx": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"ExecStart": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"ExecStartEx": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"ExitType": "main",
"ExtensionImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent",
"FailureAction": "none",
"FileDescriptorStoreMax": "0",
"FileDescriptorStorePreserve": "restart",
"FinalKillSignal": "9",
"FragmentPath": "/usr/lib/systemd/system/firewalld.service",
"FreezerState": "running",
"GID": "[not set]",
"GuessMainPID": "yes",
"IOAccounting": "no",
"IOReadBytes": "18446744073709551615",
"IOReadOperations": "18446744073709551615",
"IOSchedulingClass": "2",
"IOSchedulingPriority": "4",
"IOWeight": "[not set]",
"IOWriteBytes": "18446744073709551615",
"IOWriteOperations": "18446744073709551615",
"IPAccounting": "no",
"IPEgressBytes": "[no data]",
"IPEgressPackets": "[no data]",
"IPIngressBytes": "[no data]",
"IPIngressPackets": "[no data]",
"Id": "firewalld.service",
"IgnoreOnIsolate": "no",
"IgnoreSIGPIPE": "yes",
"InactiveEnterTimestampMonotonic": "0",
"InactiveExitTimestamp": "Wed 2024-07-10 05:13:51 UTC",
"InactiveExitTimestampMonotonic": "952204469",
"InvocationID": "6eb0bd9cbf764a82a46a19747bb71bc4",
"JobRunningTimeoutUSec": "infinity",
"JobTimeoutAction": "none",
"JobTimeoutUSec": "infinity",
"KeyringMode": "private",
"KillMode": "mixed",
"KillSignal": "15",
"LimitAS": "infinity",
"LimitASSoft": "infinity",
"LimitCORE": "infinity",
"LimitCORESoft": "infinity",
"LimitCPU": "infinity",
"LimitCPUSoft": "infinity",
"LimitDATA": "infinity",
"LimitDATASoft": "infinity",
"LimitFSIZE": "infinity",
"LimitFSIZESoft": "infinity",
"LimitLOCKS": "infinity",
"LimitLOCKSSoft": "infinity",
"LimitMEMLOCK": "8388608",
"LimitMEMLOCKSoft": "8388608",
"LimitMSGQUEUE": "819200",
"LimitMSGQUEUESoft": "819200",
"LimitNICE": "0",
"LimitNICESoft": "0",
"LimitNOFILE": "524288",
"LimitNOFILESoft": "1024",
"LimitNPROC": "14756",
"LimitNPROCSoft": "14756",
"LimitRSS": "infinity",
"LimitRSSSoft": "infinity",
"LimitRTPRIO": "0",
"LimitRTPRIOSoft": "0",
"LimitRTTIME": "infinity",
"LimitRTTIMESoft": "infinity",
"LimitSIGPENDING": "14756",
"LimitSIGPENDINGSoft": "14756",
"LimitSTACK": "infinity",
"LimitSTACKSoft": "8388608",
"LoadState": "loaded",
"LockPersonality": "no",
"LogLevelMax": "-1",
"LogRateLimitBurst": "0",
"LogRateLimitIntervalUSec": "0",
"LogsDirectoryMode": "0755",
"MainPID": "6763",
"ManagedOOMMemoryPressure": "auto",
"ManagedOOMMemoryPressureLimit": "0",
"ManagedOOMPreference": "none",
"ManagedOOMSwap": "auto",
"MemoryAccounting": "yes",
"MemoryAvailable": "infinity",
"MemoryCurrent": "34697216",
"MemoryDenyWriteExecute": "no",
"MemoryHigh": "infinity",
"MemoryKSM": "no",
"MemoryLimit": "infinity",
"MemoryLow": "0",
"MemoryMax": "infinity",
"MemoryMin": "0",
"MemoryPressureThresholdUSec": "200ms",
"MemoryPressureWatch": "auto",
"MemorySwapMax": "infinity",
"MemoryZSwapMax": "infinity",
"MountAPIVFS": "no",
"MountImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent",
"NFileDescriptorStore": "0",
"NRestarts": "0",
"NUMAPolicy": "n/a",
"Names": "firewalld.service dbus-org.fedoraproject.FirewallD1.service",
"NeedDaemonReload": "no",
"Nice": "0",
"NoNewPrivileges": "no",
"NonBlocking": "no",
"NotifyAccess": "none",
"OOMPolicy": "stop",
"OOMScoreAdjust": "0",
"OnFailureJobMode": "replace",
"OnSuccessJobMode": "fail",
"Perpetual": "no",
"PrivateDevices": "no",
"PrivateIPC": "no",
"PrivateMounts": "no",
"PrivateNetwork": "no",
"PrivateTmp": "no",
"PrivateUsers": "no",
"ProcSubset": "all",
"ProtectClock": "no",
"ProtectControlGroups": "no",
"ProtectHome": "no",
"ProtectHostname": "no",
"ProtectKernelLogs": "no",
"ProtectKernelModules": "no",
"ProtectKernelTunables": "no",
"ProtectProc": "default",
"ProtectSystem": "no",
"RefuseManualStart": "no",
"RefuseManualStop": "no",
"ReloadResult": "success",
"ReloadSignal": "1",
"RemainAfterExit": "no",
"RemoveIPC": "no",
"Requires": "sysinit.target system.slice dbus.socket",
"Restart": "no",
"RestartKillSignal": "15",
"RestartMaxDelayUSec": "infinity",
"RestartMode": "normal",
"RestartSteps": "0",
"RestartUSec": "100ms",
"RestartUSecNext": "100ms",
"RestrictNamespaces": "no",
"RestrictRealtime": "no",
"RestrictSUIDSGID": "no",
"Result": "success",
"RootDirectoryStartOnly": "no",
"RootEphemeral": "no",
"RootImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent",
"RuntimeDirectoryMode": "0755",
"RuntimeDirectoryPreserve": "no",
"RuntimeMaxUSec": "infinity",
"RuntimeRandomizedExtraUSec": "0",
"SameProcessGroup": "no",
"SecureBits": "0",
"SendSIGHUP": "no",
"SendSIGKILL": "yes",
"Slice": "system.slice",
"StandardError": "null",
"StandardInput": "null",
"StandardOutput": "null",
"StartLimitAction": "none",
"StartLimitBurst": "5",
"StartLimitIntervalUSec": "10s",
"StartupBlockIOWeight": "[not set]",
"StartupCPUShares": "[not set]",
"StartupCPUWeight": "[not set]",
"StartupIOWeight": "[not set]",
"StartupMemoryHigh": "18446744073709551615",
"StartupMemoryLow": "0",
"StartupMemoryMax": "18446744073709551615",
"StartupMemorySwapMax": "18446744073709551615",
"StartupMemoryZSwapMax": "18446744073709551615",
"StateChangeTimestamp": "Wed 2024-07-10 05:14:42 UTC",
"StateChangeTimestampMonotonic": "1002588722",
"StateDirectoryMode": "0755",
"StatusErrno": "0",
"StopWhenUnneeded": "no",
"SubState": "running",
"SuccessAction": "none",
"SyslogFacility": "3",
"SyslogLevel": "6",
"SyslogLevelPrefix": "yes",
"SyslogPriority": "30",
"SystemCallErrorNumber": "2147483646",
"TTYReset": "no",
"TTYVHangup": "no",
"TTYVTDisallocate": "no",
"TasksAccounting": "yes",
"TasksCurrent": "2",
"TasksMax": "4426",
"TimeoutAbortUSec": "45s",
"TimeoutCleanUSec": "infinity",
"TimeoutStartFailureMode": "terminate",
"TimeoutStartUSec": "45s",
"TimeoutStopFailureMode": "abort",
"TimeoutStopUSec": "45s",
"TimerSlackNSec": "50000",
"Transient": "no",
"Type": "dbus",
"UID": "[not set]",
"UMask": "0022",
"UnitFilePreset": "enabled",
"UnitFileState": "enabled",
"UtmpMode": "init",
"WantedBy": "multi-user.target",
"Wants": "network-pre.target",
"WatchdogSignal": "6",
"WatchdogTimestampMonotonic": "0",
"WatchdogUSec": "0"
}
}
TASK [fedora.linux_system_roles.firewall : Enable and start firewalld service] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:28
Wednesday 10 July 2024 05:14:52 +0000 (0:00:00.414) 0:02:57.604 ********
ok: [sut] => {
"changed": false,
"enabled": true,
"name": "firewalld",
"state": "started",
"status": {
"AccessSELinuxContext": "system_u:object_r:firewalld_unit_file_t:s0",
"ActiveEnterTimestamp": "Wed 2024-07-10 05:13:52 UTC",
"ActiveEnterTimestampMonotonic": "952448512",
"ActiveExitTimestampMonotonic": "0",
"ActiveState": "active",
"After": "basic.target polkit.service dbus.socket system.slice dbus-broker.service sysinit.target",
"AllowIsolate": "no",
"AssertResult": "yes",
"AssertTimestamp": "Wed 2024-07-10 05:13:51 UTC",
"AssertTimestampMonotonic": "952190941",
"Before": "shutdown.target network-pre.target multi-user.target",
"BlockIOAccounting": "no",
"BlockIOWeight": "[not set]",
"BusName": "org.fedoraproject.FirewallD1",
"CPUAccounting": "yes",
"CPUAffinityFromNUMA": "no",
"CPUQuotaPerSecUSec": "infinity",
"CPUQuotaPeriodUSec": "infinity",
"CPUSchedulingPolicy": "0",
"CPUSchedulingPriority": "0",
"CPUSchedulingResetOnFork": "no",
"CPUShares": "[not set]",
"CPUUsageNSec": "589336000",
"CPUWeight": "[not set]",
"CacheDirectoryMode": "0755",
"CanFreeze": "yes",
"CanIsolate": "no",
"CanReload": "yes",
"CanStart": "yes",
"CanStop": "yes",
"CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore",
"CleanResult": "success",
"CollectMode": "inactive",
"ConditionResult": "yes",
"ConditionTimestamp": "Wed 2024-07-10 05:13:51 UTC",
"ConditionTimestampMonotonic": "952190937",
"ConfigurationDirectoryMode": "0755",
"Conflicts": "shutdown.target iptables.service nftables.service ebtables.service ip6tables.service ipset.service",
"ControlGroup": "/system.slice/firewalld.service",
"ControlGroupId": "6620",
"ControlPID": "0",
"CoredumpFilter": "0x33",
"DefaultDependencies": "yes",
"DefaultMemoryLow": "0",
"DefaultMemoryMin": "0",
"DefaultStartupMemoryLow": "0",
"Delegate": "no",
"Description": "firewalld - dynamic firewall daemon",
"DevicePolicy": "auto",
"Documentation": "\"man:firewalld(1)\"",
"DropInPaths": "/usr/lib/systemd/system/service.d/10-timeout-abort.conf",
"DynamicUser": "no",
"EffectiveCPUs": "0-1",
"EffectiveMemoryNodes": "0",
"EnvironmentFiles": "/etc/sysconfig/firewalld (ignore_errors=yes)",
"ExecMainCode": "0",
"ExecMainExitTimestampMonotonic": "0",
"ExecMainPID": "6763",
"ExecMainStartTimestamp": "Wed 2024-07-10 05:13:51 UTC",
"ExecMainStartTimestampMonotonic": "952204109",
"ExecMainStatus": "0",
"ExecReload": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"ExecReloadEx": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"ExecStart": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"ExecStartEx": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"ExitType": "main",
"ExtensionImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent",
"FailureAction": "none",
"FileDescriptorStoreMax": "0",
"FileDescriptorStorePreserve": "restart",
"FinalKillSignal": "9",
"FragmentPath": "/usr/lib/systemd/system/firewalld.service",
"FreezerState": "running",
"GID": "[not set]",
"GuessMainPID": "yes",
"IOAccounting": "no",
"IOReadBytes": "18446744073709551615",
"IOReadOperations": "18446744073709551615",
"IOSchedulingClass": "2",
"IOSchedulingPriority": "4",
"IOWeight": "[not set]",
"IOWriteBytes": "18446744073709551615",
"IOWriteOperations": "18446744073709551615",
"IPAccounting": "no",
"IPEgressBytes": "[no data]",
"IPEgressPackets": "[no data]",
"IPIngressBytes": "[no data]",
"IPIngressPackets": "[no data]",
"Id": "firewalld.service",
"IgnoreOnIsolate": "no",
"IgnoreSIGPIPE": "yes",
"InactiveEnterTimestampMonotonic": "0",
"InactiveExitTimestamp": "Wed 2024-07-10 05:13:51 UTC",
"InactiveExitTimestampMonotonic": "952204469",
"InvocationID": "6eb0bd9cbf764a82a46a19747bb71bc4",
"JobRunningTimeoutUSec": "infinity",
"JobTimeoutAction": "none",
"JobTimeoutUSec": "infinity",
"KeyringMode": "private",
"KillMode": "mixed",
"KillSignal": "15",
"LimitAS": "infinity",
"LimitASSoft": "infinity",
"LimitCORE": "infinity",
"LimitCORESoft": "infinity",
"LimitCPU": "infinity",
"LimitCPUSoft": "infinity",
"LimitDATA": "infinity",
"LimitDATASoft": "infinity",
"LimitFSIZE": "infinity",
"LimitFSIZESoft": "infinity",
"LimitLOCKS": "infinity",
"LimitLOCKSSoft": "infinity",
"LimitMEMLOCK": "8388608",
"LimitMEMLOCKSoft": "8388608",
"LimitMSGQUEUE": "819200",
"LimitMSGQUEUESoft": "819200",
"LimitNICE": "0",
"LimitNICESoft": "0",
"LimitNOFILE": "524288",
"LimitNOFILESoft": "1024",
"LimitNPROC": "14756",
"LimitNPROCSoft": "14756",
"LimitRSS": "infinity",
"LimitRSSSoft": "infinity",
"LimitRTPRIO": "0",
"LimitRTPRIOSoft": "0",
"LimitRTTIME": "infinity",
"LimitRTTIMESoft": "infinity",
"LimitSIGPENDING": "14756",
"LimitSIGPENDINGSoft": "14756",
"LimitSTACK": "infinity",
"LimitSTACKSoft": "8388608",
"LoadState": "loaded",
"LockPersonality": "no",
"LogLevelMax": "-1",
"LogRateLimitBurst": "0",
"LogRateLimitIntervalUSec": "0",
"LogsDirectoryMode": "0755",
"MainPID": "6763",
"ManagedOOMMemoryPressure": "auto",
"ManagedOOMMemoryPressureLimit": "0",
"ManagedOOMPreference": "none",
"ManagedOOMSwap": "auto",
"MemoryAccounting": "yes",
"MemoryAvailable": "infinity",
"MemoryCurrent": "34697216",
"MemoryDenyWriteExecute": "no",
"MemoryHigh": "infinity",
"MemoryKSM": "no",
"MemoryLimit": "infinity",
"MemoryLow": "0",
"MemoryMax": "infinity",
"MemoryMin": "0",
"MemoryPressureThresholdUSec": "200ms",
"MemoryPressureWatch": "auto",
"MemorySwapMax": "infinity",
"MemoryZSwapMax": "infinity",
"MountAPIVFS": "no",
"MountImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent",
"NFileDescriptorStore": "0",
"NRestarts": "0",
"NUMAPolicy": "n/a",
"Names": "firewalld.service dbus-org.fedoraproject.FirewallD1.service",
"NeedDaemonReload": "no",
"Nice": "0",
"NoNewPrivileges": "no",
"NonBlocking": "no",
"NotifyAccess": "none",
"OOMPolicy": "stop",
"OOMScoreAdjust": "0",
"OnFailureJobMode": "replace",
"OnSuccessJobMode": "fail",
"Perpetual": "no",
"PrivateDevices": "no",
"PrivateIPC": "no",
"PrivateMounts": "no",
"PrivateNetwork": "no",
"PrivateTmp": "no",
"PrivateUsers": "no",
"ProcSubset": "all",
"ProtectClock": "no",
"ProtectControlGroups": "no",
"ProtectHome": "no",
"ProtectHostname": "no",
"ProtectKernelLogs": "no",
"ProtectKernelModules": "no",
"ProtectKernelTunables": "no",
"ProtectProc": "default",
"ProtectSystem": "no",
"RefuseManualStart": "no",
"RefuseManualStop": "no",
"ReloadResult": "success",
"ReloadSignal": "1",
"RemainAfterExit": "no",
"RemoveIPC": "no",
"Requires": "sysinit.target system.slice dbus.socket",
"Restart": "no",
"RestartKillSignal": "15",
"RestartMaxDelayUSec": "infinity",
"RestartMode": "normal",
"RestartSteps": "0",
"RestartUSec": "100ms",
"RestartUSecNext": "100ms",
"RestrictNamespaces": "no",
"RestrictRealtime": "no",
"RestrictSUIDSGID": "no",
"Result": "success",
"RootDirectoryStartOnly": "no",
"RootEphemeral": "no",
"RootImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent",
"RuntimeDirectoryMode": "0755",
"RuntimeDirectoryPreserve": "no",
"RuntimeMaxUSec": "infinity",
"RuntimeRandomizedExtraUSec": "0",
"SameProcessGroup": "no",
"SecureBits": "0",
"SendSIGHUP": "no",
"SendSIGKILL": "yes",
"Slice": "system.slice",
"StandardError": "null",
"StandardInput": "null",
"StandardOutput": "null",
"StartLimitAction": "none",
"StartLimitBurst": "5",
"StartLimitIntervalUSec": "10s",
"StartupBlockIOWeight": "[not set]",
"StartupCPUShares": "[not set]",
"StartupCPUWeight": "[not set]",
"StartupIOWeight": "[not set]",
"StartupMemoryHigh": "18446744073709551615",
"StartupMemoryLow": "0",
"StartupMemoryMax": "18446744073709551615",
"StartupMemorySwapMax": "18446744073709551615",
"StartupMemoryZSwapMax": "18446744073709551615",
"StateChangeTimestamp": "Wed 2024-07-10 05:14:42 UTC",
"StateChangeTimestampMonotonic": "1002588722",
"StateDirectoryMode": "0755",
"StatusErrno": "0",
"StopWhenUnneeded": "no",
"SubState": "running",
"SuccessAction": "none",
"SyslogFacility": "3",
"SyslogLevel": "6",
"SyslogLevelPrefix": "yes",
"SyslogPriority": "30",
"SystemCallErrorNumber": "2147483646",
"TTYReset": "no",
"TTYVHangup": "no",
"TTYVTDisallocate": "no",
"TasksAccounting": "yes",
"TasksCurrent": "2",
"TasksMax": "4426",
"TimeoutAbortUSec": "45s",
"TimeoutCleanUSec": "infinity",
"TimeoutStartFailureMode": "terminate",
"TimeoutStartUSec": "45s",
"TimeoutStopFailureMode": "abort",
"TimeoutStopUSec": "45s",
"TimerSlackNSec": "50000",
"Transient": "no",
"Type": "dbus",
"UID": "[not set]",
"UMask": "0022",
"UnitFilePreset": "enabled",
"UnitFileState": "enabled",
"UtmpMode": "init",
"WantedBy": "multi-user.target",
"Wants": "network-pre.target",
"WatchdogSignal": "6",
"WatchdogTimestampMonotonic": "0",
"WatchdogUSec": "0"
}
}
TASK [fedora.linux_system_roles.firewall : Check if previous replaced is defined] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:34
Wednesday 10 July 2024 05:14:53 +0000 (0:00:00.415) 0:02:58.020 ********
ok: [sut] => {
"ansible_facts": {
"__firewall_previous_replaced": false,
"__firewall_python_cmd": "/usr/bin/python3",
"__firewall_report_changed": true
},
"changed": false
}
TASK [fedora.linux_system_roles.firewall : Get config files, checksums before and remove] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:43
Wednesday 10 July 2024 05:14:53 +0000 (0:00:00.026) 0:02:58.047 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__firewall_previous_replaced | bool",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.firewall : Tell firewall module it is able to report changed] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:55
Wednesday 10 July 2024 05:14:53 +0000 (0:00:00.018) 0:02:58.065 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__firewall_previous_replaced | bool",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.firewall : Configure firewall] *****************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:71
Wednesday 10 July 2024 05:14:53 +0000 (0:00:00.018) 0:02:58.084 ********
ok: [sut] => (item={'port': '8000/tcp', 'state': 'enabled'}) => {
"__firewall_changed": false,
"ansible_loop_var": "item",
"changed": false,
"item": {
"port": "8000/tcp",
"state": "enabled"
}
}
ok: [sut] => (item={'port': '9000/tcp', 'state': 'enabled'}) => {
"__firewall_changed": false,
"ansible_loop_var": "item",
"changed": false,
"item": {
"port": "9000/tcp",
"state": "enabled"
}
}
TASK [fedora.linux_system_roles.firewall : Gather firewall config information] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:120
Wednesday 10 July 2024 05:14:54 +0000 (0:00:00.753) 0:02:58.838 ********
skipping: [sut] => (item={'port': '8000/tcp', 'state': 'enabled'}) => {
"ansible_loop_var": "item",
"changed": false,
"false_condition": "firewall | length == 1",
"item": {
"port": "8000/tcp",
"state": "enabled"
},
"skip_reason": "Conditional result was False"
}
skipping: [sut] => (item={'port': '9000/tcp', 'state': 'enabled'}) => {
"ansible_loop_var": "item",
"changed": false,
"false_condition": "firewall | length == 1",
"item": {
"port": "9000/tcp",
"state": "enabled"
},
"skip_reason": "Conditional result was False"
}
skipping: [sut] => {
"changed": false
}
MSG:
All items skipped
TASK [fedora.linux_system_roles.firewall : Update firewalld_config fact] *******
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:130
Wednesday 10 July 2024 05:14:54 +0000 (0:00:00.030) 0:02:58.868 ********
skipping: [sut] => {
"changed": false,
"false_condition": "firewall | length == 1",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.firewall : Gather firewall config if no arguments] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:139
Wednesday 10 July 2024 05:14:54 +0000 (0:00:00.021) 0:02:58.889 ********
skipping: [sut] => {
"changed": false,
"false_condition": "firewall == None or firewall | length == 0",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.firewall : Update firewalld_config fact] *******
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:144
Wednesday 10 July 2024 05:14:54 +0000 (0:00:00.020) 0:02:58.909 ********
skipping: [sut] => {
"changed": false,
"false_condition": "firewall == None or firewall | length == 0",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.firewall : Get config files, checksums after] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:153
Wednesday 10 July 2024 05:14:54 +0000 (0:00:00.019) 0:02:58.929 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__firewall_previous_replaced | bool",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.firewall : Calculate what has changed] *********
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:163
Wednesday 10 July 2024 05:14:54 +0000 (0:00:00.017) 0:02:58.947 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__firewall_previous_replaced | bool",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.firewall : Show diffs] *************************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:169
Wednesday 10 July 2024 05:14:54 +0000 (0:00:00.018) 0:02:58.965 ********
skipping: [sut] => {
"false_condition": "__firewall_previous_replaced | bool"
}
TASK [Manage selinux for specified ports] **************************************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:93
Wednesday 10 July 2024 05:14:54 +0000 (0:00:00.028) 0:02:58.994 ********
skipping: [sut] => {
"changed": false,
"false_condition": "podman_selinux_ports | length > 0",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Keep track of users that need to cancel linger] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:100
Wednesday 10 July 2024 05:14:54 +0000 (0:00:00.019) 0:02:59.013 ********
ok: [sut] => {
"ansible_facts": {
"__podman_cancel_user_linger": []
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Handle certs.d files - present] *******
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:104
Wednesday 10 July 2024 05:14:54 +0000 (0:00:00.018) 0:02:59.031 ********
skipping: [sut] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : Handle credential files - present] ****
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:113
Wednesday 10 July 2024 05:14:54 +0000 (0:00:00.016) 0:02:59.048 ********
skipping: [sut] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : Handle secrets] ***********************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:122
Wednesday 10 July 2024 05:14:54 +0000 (0:00:00.015) 0:02:59.063 ********
included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml for sut => (item=(censored due to no_log))
included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml for sut => (item=(censored due to no_log))
included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml for sut => (item=(censored due to no_log))
TASK [fedora.linux_system_roles.podman : Set variables part 1] *****************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:3
Wednesday 10 July 2024 05:14:54 +0000 (0:00:00.087) 0:02:59.151 ********
ok: [sut] => {
"ansible_facts": {
"__podman_user": "root"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set variables part 2] *****************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:7
Wednesday 10 July 2024 05:14:54 +0000 (0:00:00.023) 0:02:59.175 ********
ok: [sut] => {
"ansible_facts": {
"__podman_rootless": false,
"__podman_xdg_runtime_dir": "/run/user/0"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Manage linger] ************************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:13
Wednesday 10 July 2024 05:14:54 +0000 (0:00:00.026) 0:02:59.201 ********
included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for sut
TASK [fedora.linux_system_roles.podman : Enable linger if needed] **************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12
Wednesday 10 July 2024 05:14:54 +0000 (0:00:00.031) 0:02:59.232 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_rootless | bool",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18
Wednesday 10 July 2024 05:14:54 +0000 (0:00:00.019) 0:02:59.252 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_rootless | bool",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22
Wednesday 10 July 2024 05:14:54 +0000 (0:00:00.018) 0:02:59.270 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_rootless | bool",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Stat XDG_RUNTIME_DIR] *****************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:18
Wednesday 10 July 2024 05:14:54 +0000 (0:00:00.018) 0:02:59.288 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_rootless | bool",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Manage each secret] *******************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:34
Wednesday 10 July 2024 05:14:54 +0000 (0:00:00.018) 0:02:59.307 ********
changed: [sut] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": true
}
TASK [fedora.linux_system_roles.podman : Set variables part 1] *****************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:3
Wednesday 10 July 2024 05:14:54 +0000 (0:00:00.282) 0:02:59.590 ********
ok: [sut] => {
"ansible_facts": {
"__podman_user": "root"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set variables part 2] *****************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:7
Wednesday 10 July 2024 05:14:54 +0000 (0:00:00.022) 0:02:59.613 ********
ok: [sut] => {
"ansible_facts": {
"__podman_rootless": false,
"__podman_xdg_runtime_dir": "/run/user/0"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Manage linger] ************************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:13
Wednesday 10 July 2024 05:14:54 +0000 (0:00:00.025) 0:02:59.638 ********
included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for sut
TASK [fedora.linux_system_roles.podman : Enable linger if needed] **************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12
Wednesday 10 July 2024 05:14:55 +0000 (0:00:00.031) 0:02:59.669 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_rootless | bool",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18
Wednesday 10 July 2024 05:14:55 +0000 (0:00:00.018) 0:02:59.688 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_rootless | bool",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22
Wednesday 10 July 2024 05:14:55 +0000 (0:00:00.018) 0:02:59.706 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_rootless | bool",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Stat XDG_RUNTIME_DIR] *****************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:18
Wednesday 10 July 2024 05:14:55 +0000 (0:00:00.017) 0:02:59.724 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_rootless | bool",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Manage each secret] *******************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:34
Wednesday 10 July 2024 05:14:55 +0000 (0:00:00.049) 0:02:59.773 ********
changed: [sut] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": true
}
TASK [fedora.linux_system_roles.podman : Set variables part 1] *****************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:3
Wednesday 10 July 2024 05:14:55 +0000 (0:00:00.284) 0:03:00.057 ********
ok: [sut] => {
"ansible_facts": {
"__podman_user": "root"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set variables part 2] *****************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:7
Wednesday 10 July 2024 05:14:55 +0000 (0:00:00.023) 0:03:00.081 ********
ok: [sut] => {
"ansible_facts": {
"__podman_rootless": false,
"__podman_xdg_runtime_dir": "/run/user/0"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Manage linger] ************************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:13
Wednesday 10 July 2024 05:14:55 +0000 (0:00:00.026) 0:03:00.107 ********
included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for sut
TASK [fedora.linux_system_roles.podman : Enable linger if needed] **************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12
Wednesday 10 July 2024 05:14:55 +0000 (0:00:00.030) 0:03:00.138 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_rootless | bool",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18
Wednesday 10 July 2024 05:14:55 +0000 (0:00:00.018) 0:03:00.157 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_rootless | bool",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22
Wednesday 10 July 2024 05:14:55 +0000 (0:00:00.018) 0:03:00.175 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_rootless | bool",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Stat XDG_RUNTIME_DIR] *****************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:18
Wednesday 10 July 2024 05:14:55 +0000 (0:00:00.018) 0:03:00.194 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_rootless | bool",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Manage each secret] *******************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:34
Wednesday 10 July 2024 05:14:55 +0000 (0:00:00.017) 0:03:00.211 ********
changed: [sut] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": true
}
TASK [fedora.linux_system_roles.podman : Handle Kubernetes specifications] *****
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:129
Wednesday 10 July 2024 05:14:55 +0000 (0:00:00.282) 0:03:00.494 ********
skipping: [sut] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : Handle Quadlet specifications] ********
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:136
Wednesday 10 July 2024 05:14:55 +0000 (0:00:00.017) 0:03:00.511 ********
included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for sut => (item=(censored due to no_log))
included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for sut => (item=(censored due to no_log))
included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for sut => (item=(censored due to no_log))
included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for sut => (item=(censored due to no_log))
included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for sut => (item=(censored due to no_log))
included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for sut => (item=(censored due to no_log))
TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14
Wednesday 10 July 2024 05:14:55 +0000 (0:00:00.090) 0:03:00.601 ********
ok: [sut] => {
"ansible_facts": {
"__podman_quadlet_file_src": "quadlet-demo.kube",
"__podman_quadlet_spec": {},
"__podman_quadlet_str": "[Install]\nWantedBy=default.target\n\n[Unit]\nRequires=quadlet-demo-mysql.service\nAfter=quadlet-demo-mysql.service\n\n[Kube]\n# Point to the yaml file in the same directory\nYaml=quadlet-demo.yml\n# Use the quadlet-demo network\nNetwork=quadlet-demo.network\n# Publish the envoy proxy data port\nPublishPort=8000:8080\n# Publish the envoy proxy admin port\nPublishPort=9000:9901\n# Use the envoy proxy config map in the same directory\nConfigMap=envoy-proxy-configmap.yml",
"__podman_quadlet_template_src": ""
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25
Wednesday 10 July 2024 05:14:55 +0000 (0:00:00.030) 0:03:00.631 ********
ok: [sut] => {
"ansible_facts": {
"__podman_continue_if_pull_fails": false,
"__podman_pull_image": true,
"__podman_state": "absent",
"__podman_systemd_unit_scope": "",
"__podman_user": "root"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] *****
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35
Wednesday 10 July 2024 05:14:55 +0000 (0:00:00.025) 0:03:00.657 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_quadlet_file_src",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48
Wednesday 10 July 2024 05:14:56 +0000 (0:00:00.019) 0:03:00.676 ********
ok: [sut] => {
"ansible_facts": {
"__podman_quadlet_name": "quadlet-demo",
"__podman_quadlet_type": "kube",
"__podman_rootless": false
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Check user and group information] *****
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57
Wednesday 10 July 2024 05:14:56 +0000 (0:00:00.065) 0:03:00.741 ********
included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for sut
TASK [fedora.linux_system_roles.podman : Get user information] *****************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2
Wednesday 10 July 2024 05:14:56 +0000 (0:00:00.037) 0:03:00.779 ********
skipping: [sut] => {
"changed": false,
"false_condition": "'getent_passwd' not in ansible_facts or __podman_user not in ansible_facts['getent_passwd']",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Fail if user does not exist] **********
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9
Wednesday 10 July 2024 05:14:56 +0000 (0:00:00.023) 0:03:00.803 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not ansible_facts[\"getent_passwd\"][__podman_user]",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set group for podman user] ************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16
Wednesday 10 July 2024 05:14:56 +0000 (0:00:00.022) 0:03:00.825 ********
ok: [sut] => {
"ansible_facts": {
"__podman_group": "0"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Get group information] ****************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:28
Wednesday 10 July 2024 05:14:56 +0000 (0:00:00.029) 0:03:00.854 ********
ok: [sut] => {
"ansible_facts": {
"getent_group": {
"root": [
"x",
"0",
""
]
}
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set group name] ***********************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:35
Wednesday 10 July 2024 05:14:56 +0000 (0:00:00.241) 0:03:01.095 ********
ok: [sut] => {
"ansible_facts": {
"__podman_group_name": "root"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : See if getsubids exists] **************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:39
Wednesday 10 July 2024 05:14:56 +0000 (0:00:00.029) 0:03:01.124 ********
ok: [sut] => {
"changed": false,
"stat": {
"atime": 1720588427.5164855,
"attr_flags": "e",
"attributes": [
"extents"
],
"block_size": 4096,
"blocks": 32,
"charset": "binary",
"checksum": "72ba28a3be19c467e89ab90f5d21f99b605159a7",
"ctime": 1720588377.3007717,
"dev": 51714,
"device_type": 0,
"executable": true,
"exists": true,
"gid": 0,
"gr_name": "root",
"inode": 203603,
"isblk": false,
"ischr": false,
"isdir": false,
"isfifo": false,
"isgid": false,
"islnk": false,
"isreg": true,
"issock": false,
"isuid": false,
"mimetype": "application/x-pie-executable",
"mode": "0755",
"mtime": 1696291200.0,
"nlink": 1,
"path": "/usr/bin/getsubids",
"pw_name": "root",
"readable": true,
"rgrp": true,
"roth": true,
"rusr": true,
"size": 16032,
"uid": 0,
"version": "2092964490",
"wgrp": false,
"woth": false,
"writeable": true,
"wusr": true,
"xgrp": true,
"xoth": true,
"xusr": true
}
}
TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:50
Wednesday 10 July 2024 05:14:56 +0000 (0:00:00.240) 0:03:01.364 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_user not in [\"root\", \"0\"]",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Check group with getsubids] ***********
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:55
Wednesday 10 July 2024 05:14:56 +0000 (0:00:00.018) 0:03:01.383 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_user not in [\"root\", \"0\"]",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ******
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:60
Wednesday 10 July 2024 05:14:56 +0000 (0:00:00.019) 0:03:01.402 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_user not in [\"root\", \"0\"]",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Get subuid file] **********************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:74
Wednesday 10 July 2024 05:14:56 +0000 (0:00:00.017) 0:03:01.420 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_stat_getsubids.stat.exists",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Get subgid file] **********************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:79
Wednesday 10 July 2024 05:14:56 +0000 (0:00:00.018) 0:03:01.439 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_stat_getsubids.stat.exists",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ******
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:84
Wednesday 10 July 2024 05:14:56 +0000 (0:00:00.017) 0:03:01.457 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_stat_getsubids.stat.exists",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ******
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:94
Wednesday 10 July 2024 05:14:56 +0000 (0:00:00.018) 0:03:01.475 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_stat_getsubids.stat.exists",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] *****
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:101
Wednesday 10 July 2024 05:14:56 +0000 (0:00:00.018) 0:03:01.493 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_stat_getsubids.stat.exists",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62
Wednesday 10 July 2024 05:14:56 +0000 (0:00:00.017) 0:03:01.511 ********
ok: [sut] => {
"ansible_facts": {
"__podman_activate_systemd_unit": true,
"__podman_images_found": [],
"__podman_kube_yamls_raw": [
"quadlet-demo.yml"
],
"__podman_service_name": "quadlet-demo.service",
"__podman_systemd_scope": "system",
"__podman_user_home_dir": "/root",
"__podman_xdg_runtime_dir": "/run/user/0"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73
Wednesday 10 July 2024 05:14:56 +0000 (0:00:00.035) 0:03:01.546 ********
ok: [sut] => {
"ansible_facts": {
"__podman_quadlet_path": "/etc/containers/systemd"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Get kube yaml contents] ***************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77
Wednesday 10 July 2024 05:14:56 +0000 (0:00:00.020) 0:03:01.566 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_state != \"absent\"",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87
Wednesday 10 July 2024 05:14:56 +0000 (0:00:00.052) 0:03:01.619 ********
ok: [sut] => {
"ansible_facts": {
"__podman_images": [],
"__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo.kube",
"__podman_volumes": []
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 6] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:103
Wednesday 10 July 2024 05:14:57 +0000 (0:00:00.048) 0:03:01.667 ********
ok: [sut] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : Cleanup quadlets] *********************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:110
Wednesday 10 July 2024 05:14:57 +0000 (0:00:00.024) 0:03:01.692 ********
included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml for sut
TASK [fedora.linux_system_roles.podman : Stat XDG_RUNTIME_DIR] *****************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:4
Wednesday 10 July 2024 05:14:57 +0000 (0:00:00.045) 0:03:01.738 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_rootless | bool",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Stop and disable service] *************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:12
Wednesday 10 July 2024 05:14:57 +0000 (0:00:00.018) 0:03:01.756 ********
changed: [sut] => {
"changed": true,
"enabled": false,
"failed_when_result": false,
"name": "quadlet-demo.service",
"state": "stopped",
"status": {
"AccessSELinuxContext": "system_u:object_r:systemd_unit_file_t:s0",
"ActiveEnterTimestamp": "Wed 2024-07-10 05:14:43 UTC",
"ActiveEnterTimestampMonotonic": "1003882093",
"ActiveExitTimestampMonotonic": "0",
"ActiveState": "active",
"After": "-.mount quadlet-demo-mysql.service basic.target quadlet-demo-network.service sysinit.target systemd-journald.socket system.slice",
"AllowIsolate": "no",
"AssertResult": "yes",
"AssertTimestamp": "Wed 2024-07-10 05:14:42 UTC",
"AssertTimestampMonotonic": "1002988075",
"Before": "shutdown.target multi-user.target",
"BlockIOAccounting": "no",
"BlockIOWeight": "[not set]",
"CPUAccounting": "yes",
"CPUAffinityFromNUMA": "no",
"CPUQuotaPerSecUSec": "infinity",
"CPUQuotaPeriodUSec": "infinity",
"CPUSchedulingPolicy": "0",
"CPUSchedulingPriority": "0",
"CPUSchedulingResetOnFork": "no",
"CPUShares": "[not set]",
"CPUUsageNSec": "562856000",
"CPUWeight": "[not set]",
"CacheDirectoryMode": "0755",
"CanFreeze": "yes",
"CanIsolate": "no",
"CanReload": "no",
"CanStart": "yes",
"CanStop": "yes",
"CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore",
"CleanResult": "success",
"CollectMode": "inactive",
"ConditionResult": "yes",
"ConditionTimestamp": "Wed 2024-07-10 05:14:42 UTC",
"ConditionTimestampMonotonic": "1002988072",
"ConfigurationDirectoryMode": "0755",
"Conflicts": "shutdown.target",
"ControlGroup": "/system.slice/quadlet-demo.service",
"ControlGroupId": "8519",
"ControlPID": "0",
"CoredumpFilter": "0x33",
"DefaultDependencies": "yes",
"DefaultMemoryLow": "0",
"DefaultMemoryMin": "0",
"DefaultStartupMemoryLow": "0",
"Delegate": "no",
"Description": "quadlet-demo.service",
"DevicePolicy": "auto",
"DropInPaths": "/usr/lib/systemd/system/service.d/10-timeout-abort.conf",
"DynamicUser": "no",
"EffectiveCPUs": "0-1",
"EffectiveMemoryNodes": "0",
"Environment": "PODMAN_SYSTEMD_UNIT=quadlet-demo.service",
"ExecMainCode": "0",
"ExecMainExitTimestampMonotonic": "0",
"ExecMainPID": "9332",
"ExecMainStartTimestamp": "Wed 2024-07-10 05:14:43 UTC",
"ExecMainStartTimestampMonotonic": "1003881995",
"ExecMainStatus": "0",
"ExecStart": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman kube play --replace --service-container=true --network=systemd-quadlet-demo --configmap /etc/containers/systemd/envoy-proxy-configmap.yml --publish 8000:8080 --publish 9000:9901 /etc/containers/systemd/quadlet-demo.yml ; ignore_errors=no ; start_time=[Wed 2024-07-10 05:14:42 UTC] ; stop_time=[n/a] ; pid=9293 ; code=(null) ; status=0/0 }",
"ExecStartEx": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman kube play --replace --service-container=true --network=systemd-quadlet-demo --configmap /etc/containers/systemd/envoy-proxy-configmap.yml --publish 8000:8080 --publish 9000:9901 /etc/containers/systemd/quadlet-demo.yml ; flags= ; start_time=[Wed 2024-07-10 05:14:42 UTC] ; stop_time=[n/a] ; pid=9293 ; code=(null) ; status=0/0 }",
"ExecStopPost": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman kube down /etc/containers/systemd/quadlet-demo.yml ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"ExecStopPostEx": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman kube down /etc/containers/systemd/quadlet-demo.yml ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"ExitType": "main",
"ExtensionImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent",
"FailureAction": "none",
"FileDescriptorStoreMax": "0",
"FileDescriptorStorePreserve": "restart",
"FinalKillSignal": "9",
"FragmentPath": "/run/systemd/generator/quadlet-demo.service",
"FreezerState": "running",
"GID": "[not set]",
"GuessMainPID": "yes",
"IOAccounting": "no",
"IOReadBytes": "18446744073709551615",
"IOReadOperations": "18446744073709551615",
"IOSchedulingClass": "2",
"IOSchedulingPriority": "4",
"IOWeight": "[not set]",
"IOWriteBytes": "18446744073709551615",
"IOWriteOperations": "18446744073709551615",
"IPAccounting": "no",
"IPEgressBytes": "[no data]",
"IPEgressPackets": "[no data]",
"IPIngressBytes": "[no data]",
"IPIngressPackets": "[no data]",
"Id": "quadlet-demo.service",
"IgnoreOnIsolate": "no",
"IgnoreSIGPIPE": "yes",
"InactiveEnterTimestampMonotonic": "0",
"InactiveExitTimestamp": "Wed 2024-07-10 05:14:42 UTC",
"InactiveExitTimestampMonotonic": "1002999439",
"InvocationID": "d800043225e649d39861ccd65ac46ecd",
"JobRunningTimeoutUSec": "infinity",
"JobTimeoutAction": "none",
"JobTimeoutUSec": "infinity",
"KeyringMode": "private",
"KillMode": "mixed",
"KillSignal": "15",
"LimitAS": "infinity",
"LimitASSoft": "infinity",
"LimitCORE": "infinity",
"LimitCORESoft": "infinity",
"LimitCPU": "infinity",
"LimitCPUSoft": "infinity",
"LimitDATA": "infinity",
"LimitDATASoft": "infinity",
"LimitFSIZE": "infinity",
"LimitFSIZESoft": "infinity",
"LimitLOCKS": "infinity",
"LimitLOCKSSoft": "infinity",
"LimitMEMLOCK": "8388608",
"LimitMEMLOCKSoft": "8388608",
"LimitMSGQUEUE": "819200",
"LimitMSGQUEUESoft": "819200",
"LimitNICE": "0",
"LimitNICESoft": "0",
"LimitNOFILE": "524288",
"LimitNOFILESoft": "1024",
"LimitNPROC": "14756",
"LimitNPROCSoft": "14756",
"LimitRSS": "infinity",
"LimitRSSSoft": "infinity",
"LimitRTPRIO": "0",
"LimitRTPRIOSoft": "0",
"LimitRTTIME": "infinity",
"LimitRTTIMESoft": "infinity",
"LimitSIGPENDING": "14756",
"LimitSIGPENDINGSoft": "14756",
"LimitSTACK": "infinity",
"LimitSTACKSoft": "8388608",
"LoadState": "loaded",
"LockPersonality": "no",
"LogLevelMax": "-1",
"LogRateLimitBurst": "0",
"LogRateLimitIntervalUSec": "0",
"LogsDirectoryMode": "0755",
"MainPID": "9332",
"ManagedOOMMemoryPressure": "auto",
"ManagedOOMMemoryPressureLimit": "0",
"ManagedOOMPreference": "none",
"ManagedOOMSwap": "auto",
"MemoryAccounting": "yes",
"MemoryAvailable": "infinity",
"MemoryCurrent": "4710400",
"MemoryDenyWriteExecute": "no",
"MemoryHigh": "infinity",
"MemoryKSM": "no",
"MemoryLimit": "infinity",
"MemoryLow": "0",
"MemoryMax": "infinity",
"MemoryMin": "0",
"MemoryPressureThresholdUSec": "200ms",
"MemoryPressureWatch": "auto",
"MemorySwapMax": "infinity",
"MemoryZSwapMax": "infinity",
"MountAPIVFS": "no",
"MountImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent",
"NFileDescriptorStore": "0",
"NRestarts": "0",
"NUMAPolicy": "n/a",
"Names": "quadlet-demo.service",
"NeedDaemonReload": "no",
"Nice": "0",
"NoNewPrivileges": "no",
"NonBlocking": "no",
"NotifyAccess": "all",
"OOMPolicy": "stop",
"OOMScoreAdjust": "0",
"OnFailureJobMode": "replace",
"OnSuccessJobMode": "fail",
"Perpetual": "no",
"PrivateDevices": "no",
"PrivateIPC": "no",
"PrivateMounts": "no",
"PrivateNetwork": "no",
"PrivateTmp": "no",
"PrivateUsers": "no",
"ProcSubset": "all",
"ProtectClock": "no",
"ProtectControlGroups": "no",
"ProtectHome": "no",
"ProtectHostname": "no",
"ProtectKernelLogs": "no",
"ProtectKernelModules": "no",
"ProtectKernelTunables": "no",
"ProtectProc": "default",
"ProtectSystem": "no",
"RefuseManualStart": "no",
"RefuseManualStop": "no",
"ReloadResult": "success",
"ReloadSignal": "1",
"RemainAfterExit": "no",
"RemoveIPC": "no",
"Requires": "sysinit.target system.slice -.mount quadlet-demo-mysql.service quadlet-demo-network.service",
"RequiresMountsFor": "/run/containers",
"Restart": "no",
"RestartKillSignal": "15",
"RestartMaxDelayUSec": "infinity",
"RestartMode": "normal",
"RestartSteps": "0",
"RestartUSec": "100ms",
"RestartUSecNext": "100ms",
"RestrictNamespaces": "no",
"RestrictRealtime": "no",
"RestrictSUIDSGID": "no",
"Result": "success",
"RootDirectoryStartOnly": "no",
"RootEphemeral": "no",
"RootImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent",
"RuntimeDirectoryMode": "0755",
"RuntimeDirectoryPreserve": "no",
"RuntimeMaxUSec": "infinity",
"RuntimeRandomizedExtraUSec": "0",
"SameProcessGroup": "no",
"SecureBits": "0",
"SendSIGHUP": "no",
"SendSIGKILL": "yes",
"Slice": "system.slice",
"SourcePath": "/etc/containers/systemd/quadlet-demo.kube",
"StandardError": "inherit",
"StandardInput": "null",
"StandardOutput": "journal",
"StartLimitAction": "none",
"StartLimitBurst": "5",
"StartLimitIntervalUSec": "10s",
"StartupBlockIOWeight": "[not set]",
"StartupCPUShares": "[not set]",
"StartupCPUWeight": "[not set]",
"StartupIOWeight": "[not set]",
"StartupMemoryHigh": "18446744073709551615",
"StartupMemoryLow": "0",
"StartupMemoryMax": "18446744073709551615",
"StartupMemorySwapMax": "18446744073709551615",
"StartupMemoryZSwapMax": "18446744073709551615",
"StateChangeTimestamp": "Wed 2024-07-10 05:14:43 UTC",
"StateChangeTimestampMonotonic": "1003882093",
"StateDirectoryMode": "0755",
"StatusErrno": "0",
"StopWhenUnneeded": "no",
"SubState": "running",
"SuccessAction": "none",
"SyslogFacility": "3",
"SyslogIdentifier": "quadlet-demo",
"SyslogLevel": "6",
"SyslogLevelPrefix": "yes",
"SyslogPriority": "30",
"SystemCallErrorNumber": "2147483646",
"TTYReset": "no",
"TTYVHangup": "no",
"TTYVTDisallocate": "no",
"TasksAccounting": "yes",
"TasksCurrent": "4",
"TasksMax": "4426",
"TimeoutAbortUSec": "45s",
"TimeoutCleanUSec": "infinity",
"TimeoutStartFailureMode": "terminate",
"TimeoutStartUSec": "45s",
"TimeoutStopFailureMode": "abort",
"TimeoutStopUSec": "45s",
"TimerSlackNSec": "50000",
"Transient": "no",
"Type": "notify",
"UID": "[not set]",
"UMask": "0022",
"UnitFilePreset": "disabled",
"UnitFileState": "generated",
"UtmpMode": "init",
"WantedBy": "multi-user.target",
"WatchdogSignal": "6",
"WatchdogTimestampMonotonic": "0",
"WatchdogUSec": "0"
}
}
TASK [fedora.linux_system_roles.podman : Remove quadlet file] ******************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:33
Wednesday 10 July 2024 05:14:58 +0000 (0:00:01.416) 0:03:03.173 ********
changed: [sut] => {
"changed": true,
"path": "/etc/containers/systemd/quadlet-demo.kube",
"state": "absent"
}
TASK [fedora.linux_system_roles.podman : Manage linger] ************************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:39
Wednesday 10 July 2024 05:14:58 +0000 (0:00:00.243) 0:03:03.416 ********
included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for sut
TASK [fedora.linux_system_roles.podman : Enable linger if needed] **************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12
Wednesday 10 July 2024 05:14:58 +0000 (0:00:00.031) 0:03:03.448 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_rootless | bool",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18
Wednesday 10 July 2024 05:14:58 +0000 (0:00:00.018) 0:03:03.467 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_rootless | bool",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22
Wednesday 10 July 2024 05:14:58 +0000 (0:00:00.018) 0:03:03.486 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_rootless | bool",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Reload systemctl] *********************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:47
Wednesday 10 July 2024 05:14:58 +0000 (0:00:00.018) 0:03:03.504 ********
ok: [sut] => {
"changed": false,
"name": null,
"status": {}
}
TASK [fedora.linux_system_roles.podman : Prune images no longer in use] ********
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:56
Wednesday 10 July 2024 05:14:59 +0000 (0:00:00.833) 0:03:04.338 ********
changed: [sut] => {
"changed": true,
"cmd": [
"podman",
"image",
"prune",
"--all",
"-f"
],
"delta": "0:00:00.544309",
"end": "2024-07-10 05:15:00.419831",
"rc": 0,
"start": "2024-07-10 05:14:59.875522"
}
STDOUT:
fcf3e41b8864a14d75a6d0627d3d02154e28a153aa57e8baa392cd744ffa0d0b
5af2585e22ed1562885d9407efab74010090427be79048c2cd6a226517cc1e1d
6c68feb81ac7fb684f7b4d483de9bffaff110fa95445c248359b77bb5d3b2295
TASK [fedora.linux_system_roles.podman : Remove networks no longer in use] *****
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:64
Wednesday 10 July 2024 05:15:00 +0000 (0:00:00.781) 0:03:05.119 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_quadlet_type == \"network\"",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Remove volumes no longer in use] ******
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:73
Wednesday 10 July 2024 05:15:00 +0000 (0:00:00.020) 0:03:05.139 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_quadlet_type == \"volume\"",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : For testing and debugging - images] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:82
Wednesday 10 July 2024 05:15:00 +0000 (0:00:00.020) 0:03:05.160 ********
ok: [sut] => {
"changed": false,
"cmd": [
"podman",
"images",
"-n"
],
"delta": "0:00:00.035798",
"end": "2024-07-10 05:15:00.733039",
"rc": 0,
"start": "2024-07-10 05:15:00.697241"
}
STDOUT:
quay.io/linux-system-roles/mysql 5.6 dd3b2a5dcb48 2 years ago 308 MB
TASK [fedora.linux_system_roles.podman : For testing and debugging - volumes] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:92
Wednesday 10 July 2024 05:15:00 +0000 (0:00:00.271) 0:03:05.432 ********
ok: [sut] => {
"changed": false,
"cmd": [
"podman",
"volume",
"ls",
"-n"
],
"delta": "0:00:00.032788",
"end": "2024-07-10 05:15:01.002321",
"rc": 0,
"start": "2024-07-10 05:15:00.969533"
}
STDOUT:
local systemd-quadlet-demo-mysql
local wp-pv-claim
local envoy-proxy-config
local envoy-certificates
TASK [fedora.linux_system_roles.podman : For testing and debugging - containers] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:102
Wednesday 10 July 2024 05:15:01 +0000 (0:00:00.305) 0:03:05.738 ********
ok: [sut] => {
"changed": false,
"cmd": [
"podman",
"ps",
"--noheading"
],
"delta": "0:00:00.039909",
"end": "2024-07-10 05:15:01.317214",
"rc": 0,
"start": "2024-07-10 05:15:01.277305"
}
STDOUT:
21b1326f5785 quay.io/linux-system-roles/mysql:5.6 mysqld 48 seconds ago Up 48 seconds (starting) quadlet-demo-mysql
TASK [fedora.linux_system_roles.podman : For testing and debugging - networks] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:112
Wednesday 10 July 2024 05:15:01 +0000 (0:00:00.278) 0:03:06.016 ********
ok: [sut] => {
"changed": false,
"cmd": [
"podman",
"network",
"ls",
"-n",
"-q"
],
"delta": "0:00:00.031987",
"end": "2024-07-10 05:15:01.587427",
"rc": 0,
"start": "2024-07-10 05:15:01.555440"
}
STDOUT:
podman
systemd-quadlet-demo
TASK [fedora.linux_system_roles.podman : For testing and debugging - secrets] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:122
Wednesday 10 July 2024 05:15:01 +0000 (0:00:00.270) 0:03:06.287 ********
ok: [sut] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : For testing and debugging - services] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:133
Wednesday 10 July 2024 05:15:01 +0000 (0:00:00.273) 0:03:06.561 ********
ok: [sut] => {
"ansible_facts": {
"services": {
"21b1326f57855b147efcb07cbd10c5f7b7edbabb6693b7752105182196a3f742.service": {
"name": "21b1326f57855b147efcb07cbd10c5f7b7edbabb6693b7752105182196a3f742.service",
"source": "systemd",
"state": "stopped",
"status": "failed"
},
"NetworkManager-dispatcher.service": {
"name": "NetworkManager-dispatcher.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"NetworkManager-wait-online.service": {
"name": "NetworkManager-wait-online.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"NetworkManager.service": {
"name": "NetworkManager.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"arp-ethers.service": {
"name": "arp-ethers.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"auditd.service": {
"name": "auditd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"auth-rpcgss-module.service": {
"name": "auth-rpcgss-module.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"autovt@.service": {
"name": "autovt@.service",
"source": "systemd",
"state": "unknown",
"status": "alias"
},
"blk-availability.service": {
"name": "blk-availability.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"bluetooth.service": {
"name": "bluetooth.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"certmonger.service": {
"name": "certmonger.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"chrony-wait.service": {
"name": "chrony-wait.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"chronyd-restricted.service": {
"name": "chronyd-restricted.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"chronyd.service": {
"name": "chronyd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"cloud-config.service": {
"name": "cloud-config.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"cloud-final.service": {
"name": "cloud-final.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"cloud-init-hotplugd.service": {
"name": "cloud-init-hotplugd.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"cloud-init-local.service": {
"name": "cloud-init-local.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"cloud-init.service": {
"name": "cloud-init.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"console-getty.service": {
"name": "console-getty.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"container-getty@.service": {
"name": "container-getty@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"dbus-broker.service": {
"name": "dbus-broker.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"dbus-org.bluez.service": {
"name": "dbus-org.bluez.service",
"source": "systemd",
"state": "inactive",
"status": "alias"
},
"dbus-org.fedoraproject.FirewallD1.service": {
"name": "dbus-org.fedoraproject.FirewallD1.service",
"source": "systemd",
"state": "active",
"status": "alias"
},
"dbus-org.freedesktop.home1.service": {
"name": "dbus-org.freedesktop.home1.service",
"source": "systemd",
"state": "active",
"status": "alias"
},
"dbus-org.freedesktop.hostname1.service": {
"name": "dbus-org.freedesktop.hostname1.service",
"source": "systemd",
"state": "inactive",
"status": "alias"
},
"dbus-org.freedesktop.locale1.service": {
"name": "dbus-org.freedesktop.locale1.service",
"source": "systemd",
"state": "inactive",
"status": "alias"
},
"dbus-org.freedesktop.login1.service": {
"name": "dbus-org.freedesktop.login1.service",
"source": "systemd",
"state": "active",
"status": "alias"
},
"dbus-org.freedesktop.nm-dispatcher.service": {
"name": "dbus-org.freedesktop.nm-dispatcher.service",
"source": "systemd",
"state": "inactive",
"status": "alias"
},
"dbus-org.freedesktop.oom1.service": {
"name": "dbus-org.freedesktop.oom1.service",
"source": "systemd",
"state": "active",
"status": "alias"
},
"dbus-org.freedesktop.portable1.service": {
"name": "dbus-org.freedesktop.portable1.service",
"source": "systemd",
"state": "inactive",
"status": "alias"
},
"dbus-org.freedesktop.resolve1.service": {
"name": "dbus-org.freedesktop.resolve1.service",
"source": "systemd",
"state": "active",
"status": "alias"
},
"dbus-org.freedesktop.timedate1.service": {
"name": "dbus-org.freedesktop.timedate1.service",
"source": "systemd",
"state": "inactive",
"status": "alias"
},
"dbus.service": {
"name": "dbus.service",
"source": "systemd",
"state": "active",
"status": "alias"
},
"debug-shell.service": {
"name": "debug-shell.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"display-manager.service": {
"name": "display-manager.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"dm-event.service": {
"name": "dm-event.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dnf-makecache.service": {
"name": "dnf-makecache.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dnf-system-upgrade-cleanup.service": {
"name": "dnf-system-upgrade-cleanup.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dnf-system-upgrade.service": {
"name": "dnf-system-upgrade.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"dracut-cmdline.service": {
"name": "dracut-cmdline.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-initqueue.service": {
"name": "dracut-initqueue.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-mount.service": {
"name": "dracut-mount.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-mount.service": {
"name": "dracut-pre-mount.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-pivot.service": {
"name": "dracut-pre-pivot.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-trigger.service": {
"name": "dracut-pre-trigger.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-udev.service": {
"name": "dracut-pre-udev.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-shutdown-onfailure.service": {
"name": "dracut-shutdown-onfailure.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-shutdown.service": {
"name": "dracut-shutdown.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"ebtables.service": {
"name": "ebtables.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"emergency.service": {
"name": "emergency.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"firewalld.service": {
"name": "firewalld.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"fsidd.service": {
"name": "fsidd.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"fstrim.service": {
"name": "fstrim.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"fwupd-offline-update.service": {
"name": "fwupd-offline-update.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"fwupd-refresh.service": {
"name": "fwupd-refresh.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"fwupd.service": {
"name": "fwupd.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"getty@.service": {
"name": "getty@.service",
"source": "systemd",
"state": "unknown",
"status": "enabled"
},
"getty@tty1.service": {
"name": "getty@tty1.service",
"source": "systemd",
"state": "running",
"status": "active"
},
"grub-boot-indeterminate.service": {
"name": "grub-boot-indeterminate.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"grub2-systemd-integration.service": {
"name": "grub2-systemd-integration.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"gssproxy.service": {
"name": "gssproxy.service",
"source": "systemd",
"state": "running",
"status": "disabled"
},
"hv_kvp_daemon.service": {
"name": "hv_kvp_daemon.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"import-state.service": {
"name": "import-state.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"initrd-cleanup.service": {
"name": "initrd-cleanup.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"initrd-parse-etc.service": {
"name": "initrd-parse-etc.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"initrd-switch-root.service": {
"name": "initrd-switch-root.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"initrd-udevadm-cleanup-db.service": {
"name": "initrd-udevadm-cleanup-db.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"ip6tables.service": {
"name": "ip6tables.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"ipset.service": {
"name": "ipset.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"iptables.service": {
"name": "iptables.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"kmod-static-nodes.service": {
"name": "kmod-static-nodes.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"ldconfig.service": {
"name": "ldconfig.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"loadmodules.service": {
"name": "loadmodules.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"lvm2-lvmpolld.service": {
"name": "lvm2-lvmpolld.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"lvm2-monitor.service": {
"name": "lvm2-monitor.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"man-db-cache-update.service": {
"name": "man-db-cache-update.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"man-db-restart-cache-update.service": {
"name": "man-db-restart-cache-update.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"mdadm-grow-continue@.service": {
"name": "mdadm-grow-continue@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"mdadm-last-resort@.service": {
"name": "mdadm-last-resort@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"mdcheck_continue.service": {
"name": "mdcheck_continue.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"mdcheck_start.service": {
"name": "mdcheck_start.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"mdmon@.service": {
"name": "mdmon@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"mdmonitor-oneshot.service": {
"name": "mdmonitor-oneshot.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"mdmonitor.service": {
"name": "mdmonitor.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"modprobe@.service": {
"name": "modprobe@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"modprobe@configfs.service": {
"name": "modprobe@configfs.service",
"source": "systemd",
"state": "stopped",
"status": "inactive"
},
"modprobe@dm_mod.service": {
"name": "modprobe@dm_mod.service",
"source": "systemd",
"state": "stopped",
"status": "inactive"
},
"modprobe@drm.service": {
"name": "modprobe@drm.service",
"source": "systemd",
"state": "stopped",
"status": "inactive"
},
"modprobe@efi_pstore.service": {
"name": "modprobe@efi_pstore.service",
"source": "systemd",
"state": "stopped",
"status": "inactive"
},
"modprobe@fuse.service": {
"name": "modprobe@fuse.service",
"source": "systemd",
"state": "stopped",
"status": "inactive"
},
"modprobe@loop.service": {
"name": "modprobe@loop.service",
"source": "systemd",
"state": "stopped",
"status": "inactive"
},
"netavark-dhcp-proxy.service": {
"name": "netavark-dhcp-proxy.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"netavark-firewalld-reload.service": {
"name": "netavark-firewalld-reload.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"network.service": {
"name": "network.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"nfs-blkmap.service": {
"name": "nfs-blkmap.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"nfs-idmapd.service": {
"name": "nfs-idmapd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nfs-mountd.service": {
"name": "nfs-mountd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nfs-server.service": {
"name": "nfs-server.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"nfs-utils.service": {
"name": "nfs-utils.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nfsdcld.service": {
"name": "nfsdcld.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nftables.service": {
"name": "nftables.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"nis-domainname.service": {
"name": "nis-domainname.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"nm-priv-helper.service": {
"name": "nm-priv-helper.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"ntpd.service": {
"name": "ntpd.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"ntpdate.service": {
"name": "ntpdate.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"pam_namespace.service": {
"name": "pam_namespace.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"pcscd.service": {
"name": "pcscd.service",
"source": "systemd",
"state": "stopped",
"status": "indirect"
},
"plymouth-halt.service": {
"name": "plymouth-halt.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-kexec.service": {
"name": "plymouth-kexec.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-poweroff.service": {
"name": "plymouth-poweroff.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-quit-wait.service": {
"name": "plymouth-quit-wait.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-quit.service": {
"name": "plymouth-quit.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-read-write.service": {
"name": "plymouth-read-write.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-reboot.service": {
"name": "plymouth-reboot.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-start.service": {
"name": "plymouth-start.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-switch-root-initramfs.service": {
"name": "plymouth-switch-root-initramfs.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-switch-root.service": {
"name": "plymouth-switch-root.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"podman-auto-update.service": {
"name": "podman-auto-update.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"podman-clean-transient.service": {
"name": "podman-clean-transient.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"podman-kube@.service": {
"name": "podman-kube@.service",
"source": "systemd",
"state": "unknown",
"status": "disabled"
},
"podman-restart.service": {
"name": "podman-restart.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"podman.service": {
"name": "podman.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"polkit.service": {
"name": "polkit.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"quadlet-demo-mysql-volume.service": {
"name": "quadlet-demo-mysql-volume.service",
"source": "systemd",
"state": "stopped",
"status": "generated"
},
"quadlet-demo-mysql.service": {
"name": "quadlet-demo-mysql.service",
"source": "systemd",
"state": "running",
"status": "generated"
},
"quadlet-demo-network.service": {
"name": "quadlet-demo-network.service",
"source": "systemd",
"state": "stopped",
"status": "generated"
},
"quotaon.service": {
"name": "quotaon.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"raid-check.service": {
"name": "raid-check.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rc-local.service": {
"name": "rc-local.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rescue.service": {
"name": "rescue.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"restraintd.service": {
"name": "restraintd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"rngd.service": {
"name": "rngd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"rpc-gssd.service": {
"name": "rpc-gssd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rpc-statd-notify.service": {
"name": "rpc-statd-notify.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rpc-statd.service": {
"name": "rpc-statd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rpc-svcgssd.service": {
"name": "rpc-svcgssd.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"rpcbind.service": {
"name": "rpcbind.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"rpmdb-migrate.service": {
"name": "rpmdb-migrate.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"rpmdb-rebuild.service": {
"name": "rpmdb-rebuild.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"selinux-autorelabel-mark.service": {
"name": "selinux-autorelabel-mark.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"selinux-autorelabel.service": {
"name": "selinux-autorelabel.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"selinux-check-proper-disable.service": {
"name": "selinux-check-proper-disable.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"serial-getty@.service": {
"name": "serial-getty@.service",
"source": "systemd",
"state": "unknown",
"status": "indirect"
},
"serial-getty@ttyS0.service": {
"name": "serial-getty@ttyS0.service",
"source": "systemd",
"state": "running",
"status": "active"
},
"snapd.seeded.service": {
"name": "snapd.seeded.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"sntp.service": {
"name": "sntp.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"ssh-host-keys-migration.service": {
"name": "ssh-host-keys-migration.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"sshd-keygen.service": {
"name": "sshd-keygen.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"sshd-keygen@.service": {
"name": "sshd-keygen@.service",
"source": "systemd",
"state": "unknown",
"status": "disabled"
},
"sshd-keygen@ecdsa.service": {
"name": "sshd-keygen@ecdsa.service",
"source": "systemd",
"state": "stopped",
"status": "inactive"
},
"sshd-keygen@ed25519.service": {
"name": "sshd-keygen@ed25519.service",
"source": "systemd",
"state": "stopped",
"status": "inactive"
},
"sshd-keygen@rsa.service": {
"name": "sshd-keygen@rsa.service",
"source": "systemd",
"state": "stopped",
"status": "inactive"
},
"sshd.service": {
"name": "sshd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"sshd@.service": {
"name": "sshd@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"sssd-autofs.service": {
"name": "sssd-autofs.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-kcm.service": {
"name": "sssd-kcm.service",
"source": "systemd",
"state": "stopped",
"status": "indirect"
},
"sssd-nss.service": {
"name": "sssd-nss.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-pac.service": {
"name": "sssd-pac.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-pam.service": {
"name": "sssd-pam.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-ssh.service": {
"name": "sssd-ssh.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-sudo.service": {
"name": "sssd-sudo.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd.service": {
"name": "sssd.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"syslog.service": {
"name": "syslog.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"system-update-cleanup.service": {
"name": "system-update-cleanup.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-ask-password-console.service": {
"name": "systemd-ask-password-console.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-ask-password-plymouth.service": {
"name": "systemd-ask-password-plymouth.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-ask-password-wall.service": {
"name": "systemd-ask-password-wall.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-backlight@.service": {
"name": "systemd-backlight@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-battery-check.service": {
"name": "systemd-battery-check.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-binfmt.service": {
"name": "systemd-binfmt.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-bless-boot.service": {
"name": "systemd-bless-boot.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-boot-check-no-failures.service": {
"name": "systemd-boot-check-no-failures.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"systemd-boot-random-seed.service": {
"name": "systemd-boot-random-seed.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-boot-update.service": {
"name": "systemd-boot-update.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"systemd-confext.service": {
"name": "systemd-confext.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"systemd-coredump@.service": {
"name": "systemd-coredump@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-exit.service": {
"name": "systemd-exit.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-firstboot.service": {
"name": "systemd-firstboot.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-fsck-root.service": {
"name": "systemd-fsck-root.service",
"source": "systemd",
"state": "stopped",
"status": "enabled-runtime"
},
"systemd-fsck@.service": {
"name": "systemd-fsck@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-growfs-root.service": {
"name": "systemd-growfs-root.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-growfs@.service": {
"name": "systemd-growfs@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-halt.service": {
"name": "systemd-halt.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-hibernate.service": {
"name": "systemd-hibernate.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-homed-activate.service": {
"name": "systemd-homed-activate.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"systemd-homed.service": {
"name": "systemd-homed.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"systemd-hostnamed.service": {
"name": "systemd-hostnamed.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-hwdb-update.service": {
"name": "systemd-hwdb-update.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-hybrid-sleep.service": {
"name": "systemd-hybrid-sleep.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-initctl.service": {
"name": "systemd-initctl.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-journal-catalog-update.service": {
"name": "systemd-journal-catalog-update.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-journal-flush.service": {
"name": "systemd-journal-flush.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-journald.service": {
"name": "systemd-journald.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"systemd-journald@.service": {
"name": "systemd-journald@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-kexec.service": {
"name": "systemd-kexec.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-localed.service": {
"name": "systemd-localed.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-logind.service": {
"name": "systemd-logind.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"systemd-machine-id-commit.service": {
"name": "systemd-machine-id-commit.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-modules-load.service": {
"name": "systemd-modules-load.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-network-generator.service": {
"name": "systemd-network-generator.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"systemd-networkd-wait-online.service": {
"name": "systemd-networkd-wait-online.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"systemd-networkd-wait-online@.service": {
"name": "systemd-networkd-wait-online@.service",
"source": "systemd",
"state": "unknown",
"status": "disabled"
},
"systemd-networkd.service": {
"name": "systemd-networkd.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"systemd-oomd.service": {
"name": "systemd-oomd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"systemd-pcrfs-root.service": {
"name": "systemd-pcrfs-root.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-pcrfs@.service": {
"name": "systemd-pcrfs@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-pcrmachine.service": {
"name": "systemd-pcrmachine.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-pcrphase-initrd.service": {
"name": "systemd-pcrphase-initrd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-pcrphase-sysinit.service": {
"name": "systemd-pcrphase-sysinit.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-pcrphase.service": {
"name": "systemd-pcrphase.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-portabled.service": {
"name": "systemd-portabled.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-poweroff.service": {
"name": "systemd-poweroff.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-pstore.service": {
"name": "systemd-pstore.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"systemd-quotacheck.service": {
"name": "systemd-quotacheck.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-random-seed.service": {
"name": "systemd-random-seed.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-reboot.service": {
"name": "systemd-reboot.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-remount-fs.service": {
"name": "systemd-remount-fs.service",
"source": "systemd",
"state": "stopped",
"status": "enabled-runtime"
},
"systemd-repart.service": {
"name": "systemd-repart.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-resolved.service": {
"name": "systemd-resolved.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"systemd-rfkill.service": {
"name": "systemd-rfkill.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-soft-reboot.service": {
"name": "systemd-soft-reboot.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-suspend-then-hibernate.service": {
"name": "systemd-suspend-then-hibernate.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-suspend.service": {
"name": "systemd-suspend.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-sysctl.service": {
"name": "systemd-sysctl.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-sysext.service": {
"name": "systemd-sysext.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"systemd-sysupdate-reboot.service": {
"name": "systemd-sysupdate-reboot.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"systemd-sysupdate.service": {
"name": "systemd-sysupdate.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"systemd-sysusers.service": {
"name": "systemd-sysusers.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-time-wait-sync.service": {
"name": "systemd-time-wait-sync.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"systemd-timedated.service": {
"name": "systemd-timedated.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-timesyncd.service": {
"name": "systemd-timesyncd.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"systemd-tmpfiles-clean.service": {
"name": "systemd-tmpfiles-clean.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-tmpfiles-setup-dev-early.service": {
"name": "systemd-tmpfiles-setup-dev-early.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-tmpfiles-setup-dev.service": {
"name": "systemd-tmpfiles-setup-dev.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-tmpfiles-setup.service": {
"name": "systemd-tmpfiles-setup.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-udev-settle.service": {
"name": "systemd-udev-settle.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-udev-trigger.service": {
"name": "systemd-udev-trigger.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-udevd.service": {
"name": "systemd-udevd.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"systemd-update-done.service": {
"name": "systemd-update-done.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-update-utmp-runlevel.service": {
"name": "systemd-update-utmp-runlevel.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-update-utmp.service": {
"name": "systemd-update-utmp.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-user-sessions.service": {
"name": "systemd-user-sessions.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-userdbd.service": {
"name": "systemd-userdbd.service",
"source": "systemd",
"state": "running",
"status": "indirect"
},
"systemd-vconsole-setup.service": {
"name": "systemd-vconsole-setup.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-volatile-root.service": {
"name": "systemd-volatile-root.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-zram-setup@.service": {
"name": "systemd-zram-setup@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-zram-setup@zram0.service": {
"name": "systemd-zram-setup@zram0.service",
"source": "systemd",
"state": "stopped",
"status": "active"
},
"udisks2.service": {
"name": "udisks2.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"unbound-anchor.service": {
"name": "unbound-anchor.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"user-runtime-dir@.service": {
"name": "user-runtime-dir@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"user-runtime-dir@0.service": {
"name": "user-runtime-dir@0.service",
"source": "systemd",
"state": "stopped",
"status": "active"
},
"user@.service": {
"name": "user@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"user@0.service": {
"name": "user@0.service",
"source": "systemd",
"state": "running",
"status": "active"
}
}
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Create and update quadlets] ***********
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:114
Wednesday 10 July 2024 05:15:04 +0000 (0:00:02.304) 0:03:08.866 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_state != \"absent\"",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14
Wednesday 10 July 2024 05:15:04 +0000 (0:00:00.021) 0:03:08.887 ********
ok: [sut] => {
"ansible_facts": {
"__podman_quadlet_file_src": "",
"__podman_quadlet_spec": {},
"__podman_quadlet_str": "---\napiVersion: v1\nkind: PersistentVolumeClaim\nmetadata:\n name: wp-pv-claim\n labels:\n app: wordpress\nspec:\n accessModes:\n - ReadWriteOnce\n resources:\n requests:\n storage: 20Gi\n---\napiVersion: v1\nkind: Pod\nmetadata:\n name: quadlet-demo\nspec:\n containers:\n - name: wordpress\n image: quay.io/linux-system-roles/wordpress:4.8-apache\n env:\n - name: WORDPRESS_DB_HOST\n value: quadlet-demo-mysql\n - name: WORDPRESS_DB_PASSWORD\n valueFrom:\n secretKeyRef:\n name: mysql-root-password-kube\n key: password\n volumeMounts:\n - name: wordpress-persistent-storage\n mountPath: /var/www/html\n resources:\n requests:\n memory: \"64Mi\"\n cpu: \"250m\"\n limits:\n memory: \"128Mi\"\n cpu: \"500m\"\n - name: envoy\n image: quay.io/linux-system-roles/envoyproxy:v1.25.0\n volumeMounts:\n - name: config-volume\n mountPath: /etc/envoy\n - name: certificates\n mountPath: /etc/envoy-certificates\n env:\n - name: ENVOY_UID\n value: \"0\"\n resources:\n requests:\n memory: \"64Mi\"\n cpu: \"250m\"\n limits:\n memory: \"128Mi\"\n cpu: \"500m\"\n volumes:\n - name: config-volume\n configMap:\n name: envoy-proxy-config\n - name: certificates\n secret:\n secretName: envoy-certificates\n - name: wordpress-persistent-storage\n persistentVolumeClaim:\n claimName: wp-pv-claim\n - name: www # not used - for testing hostpath\n hostPath:\n path: /tmp/httpd3\n - name: create # not used - for testing hostpath\n hostPath:\n path: /tmp/httpd3-create\n",
"__podman_quadlet_template_src": "quadlet-demo.yml.j2"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25
Wednesday 10 July 2024 05:15:04 +0000 (0:00:00.073) 0:03:08.961 ********
ok: [sut] => {
"ansible_facts": {
"__podman_continue_if_pull_fails": false,
"__podman_pull_image": true,
"__podman_state": "absent",
"__podman_systemd_unit_scope": "",
"__podman_user": "root"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] *****
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35
Wednesday 10 July 2024 05:15:04 +0000 (0:00:00.028) 0:03:08.989 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_quadlet_str",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48
Wednesday 10 July 2024 05:15:04 +0000 (0:00:00.023) 0:03:09.012 ********
ok: [sut] => {
"ansible_facts": {
"__podman_quadlet_name": "quadlet-demo",
"__podman_quadlet_type": "yml",
"__podman_rootless": false
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Check user and group information] *****
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57
Wednesday 10 July 2024 05:15:04 +0000 (0:00:00.032) 0:03:09.045 ********
included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for sut
TASK [fedora.linux_system_roles.podman : Get user information] *****************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2
Wednesday 10 July 2024 05:15:04 +0000 (0:00:00.040) 0:03:09.085 ********
skipping: [sut] => {
"changed": false,
"false_condition": "'getent_passwd' not in ansible_facts or __podman_user not in ansible_facts['getent_passwd']",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Fail if user does not exist] **********
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9
Wednesday 10 July 2024 05:15:04 +0000 (0:00:00.025) 0:03:09.111 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not ansible_facts[\"getent_passwd\"][__podman_user]",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set group for podman user] ************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16
Wednesday 10 July 2024 05:15:04 +0000 (0:00:00.025) 0:03:09.136 ********
ok: [sut] => {
"ansible_facts": {
"__podman_group": "0"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Get group information] ****************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:28
Wednesday 10 July 2024 05:15:04 +0000 (0:00:00.030) 0:03:09.166 ********
ok: [sut] => {
"ansible_facts": {
"getent_group": {
"root": [
"x",
"0",
""
]
}
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set group name] ***********************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:35
Wednesday 10 July 2024 05:15:04 +0000 (0:00:00.323) 0:03:09.490 ********
ok: [sut] => {
"ansible_facts": {
"__podman_group_name": "root"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : See if getsubids exists] **************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:39
Wednesday 10 July 2024 05:15:04 +0000 (0:00:00.031) 0:03:09.522 ********
ok: [sut] => {
"changed": false,
"stat": {
"atime": 1720588427.5164855,
"attr_flags": "e",
"attributes": [
"extents"
],
"block_size": 4096,
"blocks": 32,
"charset": "binary",
"checksum": "72ba28a3be19c467e89ab90f5d21f99b605159a7",
"ctime": 1720588377.3007717,
"dev": 51714,
"device_type": 0,
"executable": true,
"exists": true,
"gid": 0,
"gr_name": "root",
"inode": 203603,
"isblk": false,
"ischr": false,
"isdir": false,
"isfifo": false,
"isgid": false,
"islnk": false,
"isreg": true,
"issock": false,
"isuid": false,
"mimetype": "application/x-pie-executable",
"mode": "0755",
"mtime": 1696291200.0,
"nlink": 1,
"path": "/usr/bin/getsubids",
"pw_name": "root",
"readable": true,
"rgrp": true,
"roth": true,
"rusr": true,
"size": 16032,
"uid": 0,
"version": "2092964490",
"wgrp": false,
"woth": false,
"writeable": true,
"wusr": true,
"xgrp": true,
"xoth": true,
"xusr": true
}
}
TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:50
Wednesday 10 July 2024 05:15:05 +0000 (0:00:00.240) 0:03:09.763 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_user not in [\"root\", \"0\"]",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Check group with getsubids] ***********
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:55
Wednesday 10 July 2024 05:15:05 +0000 (0:00:00.020) 0:03:09.783 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_user not in [\"root\", \"0\"]",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ******
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:60
Wednesday 10 July 2024 05:15:05 +0000 (0:00:00.020) 0:03:09.804 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_user not in [\"root\", \"0\"]",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Get subuid file] **********************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:74
Wednesday 10 July 2024 05:15:05 +0000 (0:00:00.020) 0:03:09.825 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_stat_getsubids.stat.exists",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Get subgid file] **********************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:79
Wednesday 10 July 2024 05:15:05 +0000 (0:00:00.020) 0:03:09.846 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_stat_getsubids.stat.exists",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ******
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:84
Wednesday 10 July 2024 05:15:05 +0000 (0:00:00.019) 0:03:09.865 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_stat_getsubids.stat.exists",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ******
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:94
Wednesday 10 July 2024 05:15:05 +0000 (0:00:00.020) 0:03:09.886 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_stat_getsubids.stat.exists",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] *****
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:101
Wednesday 10 July 2024 05:15:05 +0000 (0:00:00.020) 0:03:09.906 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_stat_getsubids.stat.exists",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62
Wednesday 10 July 2024 05:15:05 +0000 (0:00:00.020) 0:03:09.927 ********
ok: [sut] => {
"ansible_facts": {
"__podman_activate_systemd_unit": true,
"__podman_images_found": [],
"__podman_kube_yamls_raw": "",
"__podman_service_name": "",
"__podman_systemd_scope": "system",
"__podman_user_home_dir": "/root",
"__podman_xdg_runtime_dir": "/run/user/0"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73
Wednesday 10 July 2024 05:15:05 +0000 (0:00:00.037) 0:03:09.964 ********
ok: [sut] => {
"ansible_facts": {
"__podman_quadlet_path": "/etc/containers/systemd"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Get kube yaml contents] ***************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77
Wednesday 10 July 2024 05:15:05 +0000 (0:00:00.022) 0:03:09.987 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_state != \"absent\"",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87
Wednesday 10 July 2024 05:15:05 +0000 (0:00:00.020) 0:03:10.007 ********
ok: [sut] => {
"ansible_facts": {
"__podman_images": [],
"__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo.yml",
"__podman_volumes": []
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 6] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:103
Wednesday 10 July 2024 05:15:05 +0000 (0:00:00.087) 0:03:10.094 ********
ok: [sut] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : Cleanup quadlets] *********************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:110
Wednesday 10 July 2024 05:15:05 +0000 (0:00:00.028) 0:03:10.123 ********
included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml for sut
TASK [fedora.linux_system_roles.podman : Stat XDG_RUNTIME_DIR] *****************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:4
Wednesday 10 July 2024 05:15:05 +0000 (0:00:00.048) 0:03:10.171 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_rootless | bool",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Stop and disable service] *************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:12
Wednesday 10 July 2024 05:15:05 +0000 (0:00:00.021) 0:03:10.193 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_service_name | length > 0",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Remove quadlet file] ******************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:33
Wednesday 10 July 2024 05:15:05 +0000 (0:00:00.023) 0:03:10.216 ********
changed: [sut] => {
"changed": true,
"path": "/etc/containers/systemd/quadlet-demo.yml",
"state": "absent"
}
TASK [fedora.linux_system_roles.podman : Manage linger] ************************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:39
Wednesday 10 July 2024 05:15:05 +0000 (0:00:00.240) 0:03:10.456 ********
included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for sut
TASK [fedora.linux_system_roles.podman : Enable linger if needed] **************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12
Wednesday 10 July 2024 05:15:05 +0000 (0:00:00.035) 0:03:10.492 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_rootless | bool",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18
Wednesday 10 July 2024 05:15:05 +0000 (0:00:00.021) 0:03:10.513 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_rootless | bool",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22
Wednesday 10 July 2024 05:15:05 +0000 (0:00:00.020) 0:03:10.533 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_rootless | bool",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Reload systemctl] *********************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:47
Wednesday 10 July 2024 05:15:05 +0000 (0:00:00.020) 0:03:10.553 ********
ok: [sut] => {
"changed": false,
"name": null,
"status": {}
}
TASK [fedora.linux_system_roles.podman : Prune images no longer in use] ********
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:56
Wednesday 10 July 2024 05:15:06 +0000 (0:00:00.825) 0:03:11.379 ********
changed: [sut] => {
"changed": true,
"cmd": [
"podman",
"image",
"prune",
"--all",
"-f"
],
"delta": "0:00:00.032062",
"end": "2024-07-10 05:15:06.947090",
"rc": 0,
"start": "2024-07-10 05:15:06.915028"
}
TASK [fedora.linux_system_roles.podman : Remove networks no longer in use] *****
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:64
Wednesday 10 July 2024 05:15:06 +0000 (0:00:00.270) 0:03:11.649 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_quadlet_type == \"network\"",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Remove volumes no longer in use] ******
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:73
Wednesday 10 July 2024 05:15:07 +0000 (0:00:00.022) 0:03:11.671 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_quadlet_type == \"volume\"",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : For testing and debugging - images] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:82
Wednesday 10 July 2024 05:15:07 +0000 (0:00:00.060) 0:03:11.731 ********
ok: [sut] => {
"changed": false,
"cmd": [
"podman",
"images",
"-n"
],
"delta": "0:00:00.034605",
"end": "2024-07-10 05:15:07.306336",
"rc": 0,
"start": "2024-07-10 05:15:07.271731"
}
STDOUT:
quay.io/linux-system-roles/mysql 5.6 dd3b2a5dcb48 2 years ago 308 MB
TASK [fedora.linux_system_roles.podman : For testing and debugging - volumes] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:92
Wednesday 10 July 2024 05:15:07 +0000 (0:00:00.278) 0:03:12.010 ********
ok: [sut] => {
"changed": false,
"cmd": [
"podman",
"volume",
"ls",
"-n"
],
"delta": "0:00:00.035355",
"end": "2024-07-10 05:15:07.582134",
"rc": 0,
"start": "2024-07-10 05:15:07.546779"
}
STDOUT:
local systemd-quadlet-demo-mysql
local wp-pv-claim
local envoy-proxy-config
local envoy-certificates
TASK [fedora.linux_system_roles.podman : For testing and debugging - containers] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:102
Wednesday 10 July 2024 05:15:07 +0000 (0:00:00.273) 0:03:12.284 ********
ok: [sut] => {
"changed": false,
"cmd": [
"podman",
"ps",
"--noheading"
],
"delta": "0:00:00.039611",
"end": "2024-07-10 05:15:07.861356",
"rc": 0,
"start": "2024-07-10 05:15:07.821745"
}
STDOUT:
21b1326f5785 quay.io/linux-system-roles/mysql:5.6 mysqld 55 seconds ago Up 54 seconds (starting) quadlet-demo-mysql
TASK [fedora.linux_system_roles.podman : For testing and debugging - networks] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:112
Wednesday 10 July 2024 05:15:07 +0000 (0:00:00.279) 0:03:12.563 ********
ok: [sut] => {
"changed": false,
"cmd": [
"podman",
"network",
"ls",
"-n",
"-q"
],
"delta": "0:00:00.032467",
"end": "2024-07-10 05:15:08.134329",
"rc": 0,
"start": "2024-07-10 05:15:08.101862"
}
STDOUT:
podman
systemd-quadlet-demo
TASK [fedora.linux_system_roles.podman : For testing and debugging - secrets] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:122
Wednesday 10 July 2024 05:15:08 +0000 (0:00:00.274) 0:03:12.838 ********
ok: [sut] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : For testing and debugging - services] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:133
Wednesday 10 July 2024 05:15:08 +0000 (0:00:00.271) 0:03:13.110 ********
ok: [sut] => {
"ansible_facts": {
"services": {
"21b1326f57855b147efcb07cbd10c5f7b7edbabb6693b7752105182196a3f742.service": {
"name": "21b1326f57855b147efcb07cbd10c5f7b7edbabb6693b7752105182196a3f742.service",
"source": "systemd",
"state": "stopped",
"status": "failed"
},
"NetworkManager-dispatcher.service": {
"name": "NetworkManager-dispatcher.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"NetworkManager-wait-online.service": {
"name": "NetworkManager-wait-online.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"NetworkManager.service": {
"name": "NetworkManager.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"arp-ethers.service": {
"name": "arp-ethers.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"auditd.service": {
"name": "auditd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"auth-rpcgss-module.service": {
"name": "auth-rpcgss-module.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"autovt@.service": {
"name": "autovt@.service",
"source": "systemd",
"state": "unknown",
"status": "alias"
},
"blk-availability.service": {
"name": "blk-availability.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"bluetooth.service": {
"name": "bluetooth.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"certmonger.service": {
"name": "certmonger.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"chrony-wait.service": {
"name": "chrony-wait.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"chronyd-restricted.service": {
"name": "chronyd-restricted.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"chronyd.service": {
"name": "chronyd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"cloud-config.service": {
"name": "cloud-config.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"cloud-final.service": {
"name": "cloud-final.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"cloud-init-hotplugd.service": {
"name": "cloud-init-hotplugd.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"cloud-init-local.service": {
"name": "cloud-init-local.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"cloud-init.service": {
"name": "cloud-init.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"console-getty.service": {
"name": "console-getty.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"container-getty@.service": {
"name": "container-getty@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"dbus-broker.service": {
"name": "dbus-broker.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"dbus-org.bluez.service": {
"name": "dbus-org.bluez.service",
"source": "systemd",
"state": "inactive",
"status": "alias"
},
"dbus-org.fedoraproject.FirewallD1.service": {
"name": "dbus-org.fedoraproject.FirewallD1.service",
"source": "systemd",
"state": "active",
"status": "alias"
},
"dbus-org.freedesktop.home1.service": {
"name": "dbus-org.freedesktop.home1.service",
"source": "systemd",
"state": "active",
"status": "alias"
},
"dbus-org.freedesktop.hostname1.service": {
"name": "dbus-org.freedesktop.hostname1.service",
"source": "systemd",
"state": "inactive",
"status": "alias"
},
"dbus-org.freedesktop.locale1.service": {
"name": "dbus-org.freedesktop.locale1.service",
"source": "systemd",
"state": "inactive",
"status": "alias"
},
"dbus-org.freedesktop.login1.service": {
"name": "dbus-org.freedesktop.login1.service",
"source": "systemd",
"state": "active",
"status": "alias"
},
"dbus-org.freedesktop.nm-dispatcher.service": {
"name": "dbus-org.freedesktop.nm-dispatcher.service",
"source": "systemd",
"state": "inactive",
"status": "alias"
},
"dbus-org.freedesktop.oom1.service": {
"name": "dbus-org.freedesktop.oom1.service",
"source": "systemd",
"state": "active",
"status": "alias"
},
"dbus-org.freedesktop.portable1.service": {
"name": "dbus-org.freedesktop.portable1.service",
"source": "systemd",
"state": "inactive",
"status": "alias"
},
"dbus-org.freedesktop.resolve1.service": {
"name": "dbus-org.freedesktop.resolve1.service",
"source": "systemd",
"state": "active",
"status": "alias"
},
"dbus-org.freedesktop.timedate1.service": {
"name": "dbus-org.freedesktop.timedate1.service",
"source": "systemd",
"state": "inactive",
"status": "alias"
},
"dbus.service": {
"name": "dbus.service",
"source": "systemd",
"state": "active",
"status": "alias"
},
"debug-shell.service": {
"name": "debug-shell.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"display-manager.service": {
"name": "display-manager.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"dm-event.service": {
"name": "dm-event.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dnf-makecache.service": {
"name": "dnf-makecache.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dnf-system-upgrade-cleanup.service": {
"name": "dnf-system-upgrade-cleanup.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dnf-system-upgrade.service": {
"name": "dnf-system-upgrade.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"dracut-cmdline.service": {
"name": "dracut-cmdline.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-initqueue.service": {
"name": "dracut-initqueue.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-mount.service": {
"name": "dracut-mount.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-mount.service": {
"name": "dracut-pre-mount.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-pivot.service": {
"name": "dracut-pre-pivot.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-trigger.service": {
"name": "dracut-pre-trigger.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-udev.service": {
"name": "dracut-pre-udev.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-shutdown-onfailure.service": {
"name": "dracut-shutdown-onfailure.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-shutdown.service": {
"name": "dracut-shutdown.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"ebtables.service": {
"name": "ebtables.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"emergency.service": {
"name": "emergency.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"firewalld.service": {
"name": "firewalld.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"fsidd.service": {
"name": "fsidd.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"fstrim.service": {
"name": "fstrim.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"fwupd-offline-update.service": {
"name": "fwupd-offline-update.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"fwupd-refresh.service": {
"name": "fwupd-refresh.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"fwupd.service": {
"name": "fwupd.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"getty@.service": {
"name": "getty@.service",
"source": "systemd",
"state": "unknown",
"status": "enabled"
},
"getty@tty1.service": {
"name": "getty@tty1.service",
"source": "systemd",
"state": "running",
"status": "active"
},
"grub-boot-indeterminate.service": {
"name": "grub-boot-indeterminate.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"grub2-systemd-integration.service": {
"name": "grub2-systemd-integration.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"gssproxy.service": {
"name": "gssproxy.service",
"source": "systemd",
"state": "running",
"status": "disabled"
},
"hv_kvp_daemon.service": {
"name": "hv_kvp_daemon.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"import-state.service": {
"name": "import-state.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"initrd-cleanup.service": {
"name": "initrd-cleanup.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"initrd-parse-etc.service": {
"name": "initrd-parse-etc.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"initrd-switch-root.service": {
"name": "initrd-switch-root.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"initrd-udevadm-cleanup-db.service": {
"name": "initrd-udevadm-cleanup-db.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"ip6tables.service": {
"name": "ip6tables.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"ipset.service": {
"name": "ipset.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"iptables.service": {
"name": "iptables.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"kmod-static-nodes.service": {
"name": "kmod-static-nodes.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"ldconfig.service": {
"name": "ldconfig.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"loadmodules.service": {
"name": "loadmodules.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"lvm2-lvmpolld.service": {
"name": "lvm2-lvmpolld.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"lvm2-monitor.service": {
"name": "lvm2-monitor.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"man-db-cache-update.service": {
"name": "man-db-cache-update.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"man-db-restart-cache-update.service": {
"name": "man-db-restart-cache-update.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"mdadm-grow-continue@.service": {
"name": "mdadm-grow-continue@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"mdadm-last-resort@.service": {
"name": "mdadm-last-resort@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"mdcheck_continue.service": {
"name": "mdcheck_continue.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"mdcheck_start.service": {
"name": "mdcheck_start.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"mdmon@.service": {
"name": "mdmon@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"mdmonitor-oneshot.service": {
"name": "mdmonitor-oneshot.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"mdmonitor.service": {
"name": "mdmonitor.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"modprobe@.service": {
"name": "modprobe@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"modprobe@configfs.service": {
"name": "modprobe@configfs.service",
"source": "systemd",
"state": "stopped",
"status": "inactive"
},
"modprobe@dm_mod.service": {
"name": "modprobe@dm_mod.service",
"source": "systemd",
"state": "stopped",
"status": "inactive"
},
"modprobe@drm.service": {
"name": "modprobe@drm.service",
"source": "systemd",
"state": "stopped",
"status": "inactive"
},
"modprobe@efi_pstore.service": {
"name": "modprobe@efi_pstore.service",
"source": "systemd",
"state": "stopped",
"status": "inactive"
},
"modprobe@fuse.service": {
"name": "modprobe@fuse.service",
"source": "systemd",
"state": "stopped",
"status": "inactive"
},
"modprobe@loop.service": {
"name": "modprobe@loop.service",
"source": "systemd",
"state": "stopped",
"status": "inactive"
},
"netavark-dhcp-proxy.service": {
"name": "netavark-dhcp-proxy.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"netavark-firewalld-reload.service": {
"name": "netavark-firewalld-reload.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"network.service": {
"name": "network.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"nfs-blkmap.service": {
"name": "nfs-blkmap.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"nfs-idmapd.service": {
"name": "nfs-idmapd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nfs-mountd.service": {
"name": "nfs-mountd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nfs-server.service": {
"name": "nfs-server.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"nfs-utils.service": {
"name": "nfs-utils.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nfsdcld.service": {
"name": "nfsdcld.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nftables.service": {
"name": "nftables.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"nis-domainname.service": {
"name": "nis-domainname.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"nm-priv-helper.service": {
"name": "nm-priv-helper.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"ntpd.service": {
"name": "ntpd.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"ntpdate.service": {
"name": "ntpdate.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"pam_namespace.service": {
"name": "pam_namespace.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"pcscd.service": {
"name": "pcscd.service",
"source": "systemd",
"state": "stopped",
"status": "indirect"
},
"plymouth-halt.service": {
"name": "plymouth-halt.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-kexec.service": {
"name": "plymouth-kexec.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-poweroff.service": {
"name": "plymouth-poweroff.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-quit-wait.service": {
"name": "plymouth-quit-wait.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-quit.service": {
"name": "plymouth-quit.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-read-write.service": {
"name": "plymouth-read-write.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-reboot.service": {
"name": "plymouth-reboot.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-start.service": {
"name": "plymouth-start.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-switch-root-initramfs.service": {
"name": "plymouth-switch-root-initramfs.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-switch-root.service": {
"name": "plymouth-switch-root.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"podman-auto-update.service": {
"name": "podman-auto-update.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"podman-clean-transient.service": {
"name": "podman-clean-transient.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"podman-kube@.service": {
"name": "podman-kube@.service",
"source": "systemd",
"state": "unknown",
"status": "disabled"
},
"podman-restart.service": {
"name": "podman-restart.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"podman.service": {
"name": "podman.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"polkit.service": {
"name": "polkit.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"quadlet-demo-mysql-volume.service": {
"name": "quadlet-demo-mysql-volume.service",
"source": "systemd",
"state": "stopped",
"status": "generated"
},
"quadlet-demo-mysql.service": {
"name": "quadlet-demo-mysql.service",
"source": "systemd",
"state": "running",
"status": "generated"
},
"quadlet-demo-network.service": {
"name": "quadlet-demo-network.service",
"source": "systemd",
"state": "stopped",
"status": "generated"
},
"quotaon.service": {
"name": "quotaon.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"raid-check.service": {
"name": "raid-check.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rc-local.service": {
"name": "rc-local.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rescue.service": {
"name": "rescue.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"restraintd.service": {
"name": "restraintd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"rngd.service": {
"name": "rngd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"rpc-gssd.service": {
"name": "rpc-gssd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rpc-statd-notify.service": {
"name": "rpc-statd-notify.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rpc-statd.service": {
"name": "rpc-statd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rpc-svcgssd.service": {
"name": "rpc-svcgssd.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"rpcbind.service": {
"name": "rpcbind.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"rpmdb-migrate.service": {
"name": "rpmdb-migrate.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"rpmdb-rebuild.service": {
"name": "rpmdb-rebuild.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"selinux-autorelabel-mark.service": {
"name": "selinux-autorelabel-mark.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"selinux-autorelabel.service": {
"name": "selinux-autorelabel.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"selinux-check-proper-disable.service": {
"name": "selinux-check-proper-disable.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"serial-getty@.service": {
"name": "serial-getty@.service",
"source": "systemd",
"state": "unknown",
"status": "indirect"
},
"serial-getty@ttyS0.service": {
"name": "serial-getty@ttyS0.service",
"source": "systemd",
"state": "running",
"status": "active"
},
"snapd.seeded.service": {
"name": "snapd.seeded.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"sntp.service": {
"name": "sntp.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"ssh-host-keys-migration.service": {
"name": "ssh-host-keys-migration.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"sshd-keygen.service": {
"name": "sshd-keygen.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"sshd-keygen@.service": {
"name": "sshd-keygen@.service",
"source": "systemd",
"state": "unknown",
"status": "disabled"
},
"sshd-keygen@ecdsa.service": {
"name": "sshd-keygen@ecdsa.service",
"source": "systemd",
"state": "stopped",
"status": "inactive"
},
"sshd-keygen@ed25519.service": {
"name": "sshd-keygen@ed25519.service",
"source": "systemd",
"state": "stopped",
"status": "inactive"
},
"sshd-keygen@rsa.service": {
"name": "sshd-keygen@rsa.service",
"source": "systemd",
"state": "stopped",
"status": "inactive"
},
"sshd.service": {
"name": "sshd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"sshd@.service": {
"name": "sshd@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"sssd-autofs.service": {
"name": "sssd-autofs.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-kcm.service": {
"name": "sssd-kcm.service",
"source": "systemd",
"state": "stopped",
"status": "indirect"
},
"sssd-nss.service": {
"name": "sssd-nss.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-pac.service": {
"name": "sssd-pac.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-pam.service": {
"name": "sssd-pam.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-ssh.service": {
"name": "sssd-ssh.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-sudo.service": {
"name": "sssd-sudo.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd.service": {
"name": "sssd.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"syslog.service": {
"name": "syslog.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"system-update-cleanup.service": {
"name": "system-update-cleanup.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-ask-password-console.service": {
"name": "systemd-ask-password-console.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-ask-password-plymouth.service": {
"name": "systemd-ask-password-plymouth.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-ask-password-wall.service": {
"name": "systemd-ask-password-wall.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-backlight@.service": {
"name": "systemd-backlight@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-battery-check.service": {
"name": "systemd-battery-check.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-binfmt.service": {
"name": "systemd-binfmt.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-bless-boot.service": {
"name": "systemd-bless-boot.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-boot-check-no-failures.service": {
"name": "systemd-boot-check-no-failures.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"systemd-boot-random-seed.service": {
"name": "systemd-boot-random-seed.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-boot-update.service": {
"name": "systemd-boot-update.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"systemd-confext.service": {
"name": "systemd-confext.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"systemd-coredump@.service": {
"name": "systemd-coredump@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-exit.service": {
"name": "systemd-exit.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-firstboot.service": {
"name": "systemd-firstboot.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-fsck-root.service": {
"name": "systemd-fsck-root.service",
"source": "systemd",
"state": "stopped",
"status": "enabled-runtime"
},
"systemd-fsck@.service": {
"name": "systemd-fsck@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-growfs-root.service": {
"name": "systemd-growfs-root.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-growfs@.service": {
"name": "systemd-growfs@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-halt.service": {
"name": "systemd-halt.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-hibernate.service": {
"name": "systemd-hibernate.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-homed-activate.service": {
"name": "systemd-homed-activate.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"systemd-homed.service": {
"name": "systemd-homed.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"systemd-hostnamed.service": {
"name": "systemd-hostnamed.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-hwdb-update.service": {
"name": "systemd-hwdb-update.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-hybrid-sleep.service": {
"name": "systemd-hybrid-sleep.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-initctl.service": {
"name": "systemd-initctl.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-journal-catalog-update.service": {
"name": "systemd-journal-catalog-update.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-journal-flush.service": {
"name": "systemd-journal-flush.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-journald.service": {
"name": "systemd-journald.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"systemd-journald@.service": {
"name": "systemd-journald@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-kexec.service": {
"name": "systemd-kexec.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-localed.service": {
"name": "systemd-localed.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-logind.service": {
"name": "systemd-logind.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"systemd-machine-id-commit.service": {
"name": "systemd-machine-id-commit.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-modules-load.service": {
"name": "systemd-modules-load.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-network-generator.service": {
"name": "systemd-network-generator.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"systemd-networkd-wait-online.service": {
"name": "systemd-networkd-wait-online.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"systemd-networkd-wait-online@.service": {
"name": "systemd-networkd-wait-online@.service",
"source": "systemd",
"state": "unknown",
"status": "disabled"
},
"systemd-networkd.service": {
"name": "systemd-networkd.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"systemd-oomd.service": {
"name": "systemd-oomd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"systemd-pcrfs-root.service": {
"name": "systemd-pcrfs-root.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-pcrfs@.service": {
"name": "systemd-pcrfs@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-pcrmachine.service": {
"name": "systemd-pcrmachine.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-pcrphase-initrd.service": {
"name": "systemd-pcrphase-initrd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-pcrphase-sysinit.service": {
"name": "systemd-pcrphase-sysinit.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-pcrphase.service": {
"name": "systemd-pcrphase.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-portabled.service": {
"name": "systemd-portabled.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-poweroff.service": {
"name": "systemd-poweroff.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-pstore.service": {
"name": "systemd-pstore.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"systemd-quotacheck.service": {
"name": "systemd-quotacheck.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-random-seed.service": {
"name": "systemd-random-seed.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-reboot.service": {
"name": "systemd-reboot.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-remount-fs.service": {
"name": "systemd-remount-fs.service",
"source": "systemd",
"state": "stopped",
"status": "enabled-runtime"
},
"systemd-repart.service": {
"name": "systemd-repart.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-resolved.service": {
"name": "systemd-resolved.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"systemd-rfkill.service": {
"name": "systemd-rfkill.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-soft-reboot.service": {
"name": "systemd-soft-reboot.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-suspend-then-hibernate.service": {
"name": "systemd-suspend-then-hibernate.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-suspend.service": {
"name": "systemd-suspend.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-sysctl.service": {
"name": "systemd-sysctl.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-sysext.service": {
"name": "systemd-sysext.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"systemd-sysupdate-reboot.service": {
"name": "systemd-sysupdate-reboot.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"systemd-sysupdate.service": {
"name": "systemd-sysupdate.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"systemd-sysusers.service": {
"name": "systemd-sysusers.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-time-wait-sync.service": {
"name": "systemd-time-wait-sync.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"systemd-timedated.service": {
"name": "systemd-timedated.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-timesyncd.service": {
"name": "systemd-timesyncd.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"systemd-tmpfiles-clean.service": {
"name": "systemd-tmpfiles-clean.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-tmpfiles-setup-dev-early.service": {
"name": "systemd-tmpfiles-setup-dev-early.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-tmpfiles-setup-dev.service": {
"name": "systemd-tmpfiles-setup-dev.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-tmpfiles-setup.service": {
"name": "systemd-tmpfiles-setup.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-udev-settle.service": {
"name": "systemd-udev-settle.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-udev-trigger.service": {
"name": "systemd-udev-trigger.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-udevd.service": {
"name": "systemd-udevd.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"systemd-update-done.service": {
"name": "systemd-update-done.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-update-utmp-runlevel.service": {
"name": "systemd-update-utmp-runlevel.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-update-utmp.service": {
"name": "systemd-update-utmp.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-user-sessions.service": {
"name": "systemd-user-sessions.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-userdbd.service": {
"name": "systemd-userdbd.service",
"source": "systemd",
"state": "running",
"status": "indirect"
},
"systemd-vconsole-setup.service": {
"name": "systemd-vconsole-setup.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-volatile-root.service": {
"name": "systemd-volatile-root.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-zram-setup@.service": {
"name": "systemd-zram-setup@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-zram-setup@zram0.service": {
"name": "systemd-zram-setup@zram0.service",
"source": "systemd",
"state": "stopped",
"status": "active"
},
"udisks2.service": {
"name": "udisks2.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"unbound-anchor.service": {
"name": "unbound-anchor.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"user-runtime-dir@.service": {
"name": "user-runtime-dir@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"user-runtime-dir@0.service": {
"name": "user-runtime-dir@0.service",
"source": "systemd",
"state": "stopped",
"status": "active"
},
"user@.service": {
"name": "user@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"user@0.service": {
"name": "user@0.service",
"source": "systemd",
"state": "running",
"status": "active"
}
}
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Create and update quadlets] ***********
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:114
Wednesday 10 July 2024 05:15:10 +0000 (0:00:02.210) 0:03:15.320 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_state != \"absent\"",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14
Wednesday 10 July 2024 05:15:10 +0000 (0:00:00.021) 0:03:15.342 ********
ok: [sut] => {
"ansible_facts": {
"__podman_quadlet_file_src": "envoy-proxy-configmap.yml",
"__podman_quadlet_spec": {},
"__podman_quadlet_str": "---\napiVersion: v1\nkind: ConfigMap\nmetadata:\n name: envoy-proxy-config\ndata:\n envoy.yaml: |\n admin:\n address:\n socket_address:\n address: 0.0.0.0\n port_value: 9901\n\n static_resources:\n listeners:\n - name: listener_0\n address:\n socket_address:\n address: 0.0.0.0\n port_value: 8080\n filter_chains:\n - filters:\n - name: envoy.filters.network.http_connection_manager\n typed_config:\n \"@type\": type.googleapis.com/envoy.extensions.filters.network.http_connection_manager.v3.HttpConnectionManager\n stat_prefix: ingress_http\n codec_type: AUTO\n route_config:\n name: local_route\n virtual_hosts:\n - name: local_service\n domains: [\"*\"]\n routes:\n - match:\n prefix: \"/\"\n route:\n cluster: backend\n http_filters:\n - name: envoy.filters.http.router\n typed_config:\n \"@type\": type.googleapis.com/envoy.extensions.filters.http.router.v3.Router\n transport_socket:\n name: envoy.transport_sockets.tls\n typed_config:\n \"@type\": type.googleapis.com/envoy.extensions.transport_sockets.tls.v3.DownstreamTlsContext\n common_tls_context:\n tls_certificates:\n - certificate_chain:\n filename: /etc/envoy-certificates/certificate.pem\n private_key:\n filename: /etc/envoy-certificates/certificate.key\n clusters:\n - name: backend\n connect_timeout: 5s\n type: STATIC\n dns_refresh_rate: 1800s\n lb_policy: ROUND_ROBIN\n load_assignment:\n cluster_name: backend\n endpoints:\n - lb_endpoints:\n - endpoint:\n address:\n socket_address:\n address: 127.0.0.1\n port_value: 80",
"__podman_quadlet_template_src": ""
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25
Wednesday 10 July 2024 05:15:10 +0000 (0:00:00.030) 0:03:15.373 ********
ok: [sut] => {
"ansible_facts": {
"__podman_continue_if_pull_fails": false,
"__podman_pull_image": true,
"__podman_state": "absent",
"__podman_systemd_unit_scope": "",
"__podman_user": "root"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] *****
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35
Wednesday 10 July 2024 05:15:10 +0000 (0:00:00.026) 0:03:15.400 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_quadlet_file_src",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48
Wednesday 10 July 2024 05:15:10 +0000 (0:00:00.022) 0:03:15.422 ********
ok: [sut] => {
"ansible_facts": {
"__podman_quadlet_name": "envoy-proxy-configmap",
"__podman_quadlet_type": "yml",
"__podman_rootless": false
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Check user and group information] *****
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57
Wednesday 10 July 2024 05:15:10 +0000 (0:00:00.032) 0:03:15.454 ********
included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for sut
TASK [fedora.linux_system_roles.podman : Get user information] *****************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2
Wednesday 10 July 2024 05:15:10 +0000 (0:00:00.040) 0:03:15.495 ********
skipping: [sut] => {
"changed": false,
"false_condition": "'getent_passwd' not in ansible_facts or __podman_user not in ansible_facts['getent_passwd']",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Fail if user does not exist] **********
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9
Wednesday 10 July 2024 05:15:10 +0000 (0:00:00.063) 0:03:15.558 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not ansible_facts[\"getent_passwd\"][__podman_user]",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set group for podman user] ************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16
Wednesday 10 July 2024 05:15:10 +0000 (0:00:00.026) 0:03:15.585 ********
ok: [sut] => {
"ansible_facts": {
"__podman_group": "0"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Get group information] ****************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:28
Wednesday 10 July 2024 05:15:10 +0000 (0:00:00.032) 0:03:15.617 ********
ok: [sut] => {
"ansible_facts": {
"getent_group": {
"root": [
"x",
"0",
""
]
}
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set group name] ***********************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:35
Wednesday 10 July 2024 05:15:11 +0000 (0:00:00.242) 0:03:15.859 ********
ok: [sut] => {
"ansible_facts": {
"__podman_group_name": "root"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : See if getsubids exists] **************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:39
Wednesday 10 July 2024 05:15:11 +0000 (0:00:00.030) 0:03:15.889 ********
ok: [sut] => {
"changed": false,
"stat": {
"atime": 1720588427.5164855,
"attr_flags": "e",
"attributes": [
"extents"
],
"block_size": 4096,
"blocks": 32,
"charset": "binary",
"checksum": "72ba28a3be19c467e89ab90f5d21f99b605159a7",
"ctime": 1720588377.3007717,
"dev": 51714,
"device_type": 0,
"executable": true,
"exists": true,
"gid": 0,
"gr_name": "root",
"inode": 203603,
"isblk": false,
"ischr": false,
"isdir": false,
"isfifo": false,
"isgid": false,
"islnk": false,
"isreg": true,
"issock": false,
"isuid": false,
"mimetype": "application/x-pie-executable",
"mode": "0755",
"mtime": 1696291200.0,
"nlink": 1,
"path": "/usr/bin/getsubids",
"pw_name": "root",
"readable": true,
"rgrp": true,
"roth": true,
"rusr": true,
"size": 16032,
"uid": 0,
"version": "2092964490",
"wgrp": false,
"woth": false,
"writeable": true,
"wusr": true,
"xgrp": true,
"xoth": true,
"xusr": true
}
}
TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:50
Wednesday 10 July 2024 05:15:11 +0000 (0:00:00.241) 0:03:16.131 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_user not in [\"root\", \"0\"]",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Check group with getsubids] ***********
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:55
Wednesday 10 July 2024 05:15:11 +0000 (0:00:00.021) 0:03:16.152 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_user not in [\"root\", \"0\"]",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ******
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:60
Wednesday 10 July 2024 05:15:11 +0000 (0:00:00.020) 0:03:16.173 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_user not in [\"root\", \"0\"]",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Get subuid file] **********************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:74
Wednesday 10 July 2024 05:15:11 +0000 (0:00:00.020) 0:03:16.193 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_stat_getsubids.stat.exists",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Get subgid file] **********************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:79
Wednesday 10 July 2024 05:15:11 +0000 (0:00:00.020) 0:03:16.214 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_stat_getsubids.stat.exists",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ******
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:84
Wednesday 10 July 2024 05:15:11 +0000 (0:00:00.020) 0:03:16.234 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_stat_getsubids.stat.exists",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ******
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:94
Wednesday 10 July 2024 05:15:11 +0000 (0:00:00.019) 0:03:16.253 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_stat_getsubids.stat.exists",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] *****
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:101
Wednesday 10 July 2024 05:15:11 +0000 (0:00:00.020) 0:03:16.274 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_stat_getsubids.stat.exists",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62
Wednesday 10 July 2024 05:15:11 +0000 (0:00:00.020) 0:03:16.294 ********
ok: [sut] => {
"ansible_facts": {
"__podman_activate_systemd_unit": true,
"__podman_images_found": [],
"__podman_kube_yamls_raw": "",
"__podman_service_name": "",
"__podman_systemd_scope": "system",
"__podman_user_home_dir": "/root",
"__podman_xdg_runtime_dir": "/run/user/0"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73
Wednesday 10 July 2024 05:15:11 +0000 (0:00:00.037) 0:03:16.331 ********
ok: [sut] => {
"ansible_facts": {
"__podman_quadlet_path": "/etc/containers/systemd"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Get kube yaml contents] ***************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77
Wednesday 10 July 2024 05:15:11 +0000 (0:00:00.061) 0:03:16.393 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_state != \"absent\"",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87
Wednesday 10 July 2024 05:15:11 +0000 (0:00:00.022) 0:03:16.415 ********
ok: [sut] => {
"ansible_facts": {
"__podman_images": [],
"__podman_quadlet_file": "/etc/containers/systemd/envoy-proxy-configmap.yml",
"__podman_volumes": []
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 6] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:103
Wednesday 10 July 2024 05:15:11 +0000 (0:00:00.049) 0:03:16.464 ********
ok: [sut] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : Cleanup quadlets] *********************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:110
Wednesday 10 July 2024 05:15:11 +0000 (0:00:00.025) 0:03:16.490 ********
included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml for sut
TASK [fedora.linux_system_roles.podman : Stat XDG_RUNTIME_DIR] *****************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:4
Wednesday 10 July 2024 05:15:11 +0000 (0:00:00.047) 0:03:16.538 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_rootless | bool",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Stop and disable service] *************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:12
Wednesday 10 July 2024 05:15:11 +0000 (0:00:00.020) 0:03:16.559 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_service_name | length > 0",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Remove quadlet file] ******************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:33
Wednesday 10 July 2024 05:15:11 +0000 (0:00:00.022) 0:03:16.581 ********
changed: [sut] => {
"changed": true,
"path": "/etc/containers/systemd/envoy-proxy-configmap.yml",
"state": "absent"
}
TASK [fedora.linux_system_roles.podman : Manage linger] ************************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:39
Wednesday 10 July 2024 05:15:12 +0000 (0:00:00.240) 0:03:16.821 ********
included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for sut
TASK [fedora.linux_system_roles.podman : Enable linger if needed] **************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12
Wednesday 10 July 2024 05:15:12 +0000 (0:00:00.035) 0:03:16.856 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_rootless | bool",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18
Wednesday 10 July 2024 05:15:12 +0000 (0:00:00.020) 0:03:16.877 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_rootless | bool",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22
Wednesday 10 July 2024 05:15:12 +0000 (0:00:00.019) 0:03:16.897 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_rootless | bool",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Reload systemctl] *********************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:47
Wednesday 10 July 2024 05:15:12 +0000 (0:00:00.020) 0:03:16.918 ********
ok: [sut] => {
"changed": false,
"name": null,
"status": {}
}
TASK [fedora.linux_system_roles.podman : Prune images no longer in use] ********
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:56
Wednesday 10 July 2024 05:15:13 +0000 (0:00:00.843) 0:03:17.761 ********
changed: [sut] => {
"changed": true,
"cmd": [
"podman",
"image",
"prune",
"--all",
"-f"
],
"delta": "0:00:00.031594",
"end": "2024-07-10 05:15:13.330891",
"rc": 0,
"start": "2024-07-10 05:15:13.299297"
}
TASK [fedora.linux_system_roles.podman : Remove networks no longer in use] *****
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:64
Wednesday 10 July 2024 05:15:13 +0000 (0:00:00.272) 0:03:18.034 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_quadlet_type == \"network\"",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Remove volumes no longer in use] ******
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:73
Wednesday 10 July 2024 05:15:13 +0000 (0:00:00.022) 0:03:18.057 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_quadlet_type == \"volume\"",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : For testing and debugging - images] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:82
Wednesday 10 July 2024 05:15:13 +0000 (0:00:00.022) 0:03:18.079 ********
ok: [sut] => {
"changed": false,
"cmd": [
"podman",
"images",
"-n"
],
"delta": "0:00:00.037608",
"end": "2024-07-10 05:15:13.652381",
"rc": 0,
"start": "2024-07-10 05:15:13.614773"
}
STDOUT:
quay.io/linux-system-roles/mysql 5.6 dd3b2a5dcb48 2 years ago 308 MB
TASK [fedora.linux_system_roles.podman : For testing and debugging - volumes] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:92
Wednesday 10 July 2024 05:15:13 +0000 (0:00:00.273) 0:03:18.353 ********
ok: [sut] => {
"changed": false,
"cmd": [
"podman",
"volume",
"ls",
"-n"
],
"delta": "0:00:00.061854",
"end": "2024-07-10 05:15:13.953104",
"rc": 0,
"start": "2024-07-10 05:15:13.891250"
}
STDOUT:
local systemd-quadlet-demo-mysql
local wp-pv-claim
local envoy-proxy-config
local envoy-certificates
TASK [fedora.linux_system_roles.podman : For testing and debugging - containers] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:102
Wednesday 10 July 2024 05:15:13 +0000 (0:00:00.307) 0:03:18.660 ********
ok: [sut] => {
"changed": false,
"cmd": [
"podman",
"ps",
"--noheading"
],
"delta": "0:00:00.104094",
"end": "2024-07-10 05:15:14.399793",
"rc": 0,
"start": "2024-07-10 05:15:14.295699"
}
TASK [fedora.linux_system_roles.podman : For testing and debugging - networks] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:112
Wednesday 10 July 2024 05:15:14 +0000 (0:00:00.443) 0:03:19.103 ********
ok: [sut] => {
"changed": false,
"cmd": [
"podman",
"network",
"ls",
"-n",
"-q"
],
"delta": "0:00:00.032065",
"end": "2024-07-10 05:15:14.674350",
"rc": 0,
"start": "2024-07-10 05:15:14.642285"
}
STDOUT:
podman
systemd-quadlet-demo
TASK [fedora.linux_system_roles.podman : For testing and debugging - secrets] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:122
Wednesday 10 July 2024 05:15:14 +0000 (0:00:00.273) 0:03:19.377 ********
ok: [sut] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : For testing and debugging - services] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:133
Wednesday 10 July 2024 05:15:14 +0000 (0:00:00.264) 0:03:19.641 ********
ok: [sut] => {
"ansible_facts": {
"services": {
"NetworkManager-dispatcher.service": {
"name": "NetworkManager-dispatcher.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"NetworkManager-wait-online.service": {
"name": "NetworkManager-wait-online.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"NetworkManager.service": {
"name": "NetworkManager.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"arp-ethers.service": {
"name": "arp-ethers.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"auditd.service": {
"name": "auditd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"auth-rpcgss-module.service": {
"name": "auth-rpcgss-module.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"autovt@.service": {
"name": "autovt@.service",
"source": "systemd",
"state": "unknown",
"status": "alias"
},
"blk-availability.service": {
"name": "blk-availability.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"bluetooth.service": {
"name": "bluetooth.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"certmonger.service": {
"name": "certmonger.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"chrony-wait.service": {
"name": "chrony-wait.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"chronyd-restricted.service": {
"name": "chronyd-restricted.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"chronyd.service": {
"name": "chronyd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"cloud-config.service": {
"name": "cloud-config.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"cloud-final.service": {
"name": "cloud-final.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"cloud-init-hotplugd.service": {
"name": "cloud-init-hotplugd.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"cloud-init-local.service": {
"name": "cloud-init-local.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"cloud-init.service": {
"name": "cloud-init.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"console-getty.service": {
"name": "console-getty.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"container-getty@.service": {
"name": "container-getty@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"dbus-broker.service": {
"name": "dbus-broker.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"dbus-org.bluez.service": {
"name": "dbus-org.bluez.service",
"source": "systemd",
"state": "inactive",
"status": "alias"
},
"dbus-org.fedoraproject.FirewallD1.service": {
"name": "dbus-org.fedoraproject.FirewallD1.service",
"source": "systemd",
"state": "active",
"status": "alias"
},
"dbus-org.freedesktop.home1.service": {
"name": "dbus-org.freedesktop.home1.service",
"source": "systemd",
"state": "active",
"status": "alias"
},
"dbus-org.freedesktop.hostname1.service": {
"name": "dbus-org.freedesktop.hostname1.service",
"source": "systemd",
"state": "inactive",
"status": "alias"
},
"dbus-org.freedesktop.locale1.service": {
"name": "dbus-org.freedesktop.locale1.service",
"source": "systemd",
"state": "inactive",
"status": "alias"
},
"dbus-org.freedesktop.login1.service": {
"name": "dbus-org.freedesktop.login1.service",
"source": "systemd",
"state": "active",
"status": "alias"
},
"dbus-org.freedesktop.nm-dispatcher.service": {
"name": "dbus-org.freedesktop.nm-dispatcher.service",
"source": "systemd",
"state": "active",
"status": "alias"
},
"dbus-org.freedesktop.oom1.service": {
"name": "dbus-org.freedesktop.oom1.service",
"source": "systemd",
"state": "active",
"status": "alias"
},
"dbus-org.freedesktop.portable1.service": {
"name": "dbus-org.freedesktop.portable1.service",
"source": "systemd",
"state": "inactive",
"status": "alias"
},
"dbus-org.freedesktop.resolve1.service": {
"name": "dbus-org.freedesktop.resolve1.service",
"source": "systemd",
"state": "active",
"status": "alias"
},
"dbus-org.freedesktop.timedate1.service": {
"name": "dbus-org.freedesktop.timedate1.service",
"source": "systemd",
"state": "inactive",
"status": "alias"
},
"dbus.service": {
"name": "dbus.service",
"source": "systemd",
"state": "active",
"status": "alias"
},
"debug-shell.service": {
"name": "debug-shell.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"display-manager.service": {
"name": "display-manager.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"dm-event.service": {
"name": "dm-event.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dnf-makecache.service": {
"name": "dnf-makecache.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dnf-system-upgrade-cleanup.service": {
"name": "dnf-system-upgrade-cleanup.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dnf-system-upgrade.service": {
"name": "dnf-system-upgrade.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"dracut-cmdline.service": {
"name": "dracut-cmdline.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-initqueue.service": {
"name": "dracut-initqueue.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-mount.service": {
"name": "dracut-mount.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-mount.service": {
"name": "dracut-pre-mount.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-pivot.service": {
"name": "dracut-pre-pivot.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-trigger.service": {
"name": "dracut-pre-trigger.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-udev.service": {
"name": "dracut-pre-udev.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-shutdown-onfailure.service": {
"name": "dracut-shutdown-onfailure.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-shutdown.service": {
"name": "dracut-shutdown.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"ebtables.service": {
"name": "ebtables.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"emergency.service": {
"name": "emergency.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"firewalld.service": {
"name": "firewalld.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"fsidd.service": {
"name": "fsidd.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"fstrim.service": {
"name": "fstrim.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"fwupd-offline-update.service": {
"name": "fwupd-offline-update.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"fwupd-refresh.service": {
"name": "fwupd-refresh.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"fwupd.service": {
"name": "fwupd.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"getty@.service": {
"name": "getty@.service",
"source": "systemd",
"state": "unknown",
"status": "enabled"
},
"getty@tty1.service": {
"name": "getty@tty1.service",
"source": "systemd",
"state": "running",
"status": "active"
},
"grub-boot-indeterminate.service": {
"name": "grub-boot-indeterminate.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"grub2-systemd-integration.service": {
"name": "grub2-systemd-integration.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"gssproxy.service": {
"name": "gssproxy.service",
"source": "systemd",
"state": "running",
"status": "disabled"
},
"hv_kvp_daemon.service": {
"name": "hv_kvp_daemon.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"import-state.service": {
"name": "import-state.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"initrd-cleanup.service": {
"name": "initrd-cleanup.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"initrd-parse-etc.service": {
"name": "initrd-parse-etc.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"initrd-switch-root.service": {
"name": "initrd-switch-root.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"initrd-udevadm-cleanup-db.service": {
"name": "initrd-udevadm-cleanup-db.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"ip6tables.service": {
"name": "ip6tables.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"ipset.service": {
"name": "ipset.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"iptables.service": {
"name": "iptables.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"kmod-static-nodes.service": {
"name": "kmod-static-nodes.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"ldconfig.service": {
"name": "ldconfig.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"loadmodules.service": {
"name": "loadmodules.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"lvm2-lvmpolld.service": {
"name": "lvm2-lvmpolld.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"lvm2-monitor.service": {
"name": "lvm2-monitor.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"man-db-cache-update.service": {
"name": "man-db-cache-update.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"man-db-restart-cache-update.service": {
"name": "man-db-restart-cache-update.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"mdadm-grow-continue@.service": {
"name": "mdadm-grow-continue@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"mdadm-last-resort@.service": {
"name": "mdadm-last-resort@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"mdcheck_continue.service": {
"name": "mdcheck_continue.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"mdcheck_start.service": {
"name": "mdcheck_start.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"mdmon@.service": {
"name": "mdmon@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"mdmonitor-oneshot.service": {
"name": "mdmonitor-oneshot.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"mdmonitor.service": {
"name": "mdmonitor.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"modprobe@.service": {
"name": "modprobe@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"modprobe@configfs.service": {
"name": "modprobe@configfs.service",
"source": "systemd",
"state": "stopped",
"status": "inactive"
},
"modprobe@dm_mod.service": {
"name": "modprobe@dm_mod.service",
"source": "systemd",
"state": "stopped",
"status": "inactive"
},
"modprobe@drm.service": {
"name": "modprobe@drm.service",
"source": "systemd",
"state": "stopped",
"status": "inactive"
},
"modprobe@efi_pstore.service": {
"name": "modprobe@efi_pstore.service",
"source": "systemd",
"state": "stopped",
"status": "inactive"
},
"modprobe@fuse.service": {
"name": "modprobe@fuse.service",
"source": "systemd",
"state": "stopped",
"status": "inactive"
},
"modprobe@loop.service": {
"name": "modprobe@loop.service",
"source": "systemd",
"state": "stopped",
"status": "inactive"
},
"netavark-dhcp-proxy.service": {
"name": "netavark-dhcp-proxy.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"netavark-firewalld-reload.service": {
"name": "netavark-firewalld-reload.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"network.service": {
"name": "network.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"nfs-blkmap.service": {
"name": "nfs-blkmap.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"nfs-idmapd.service": {
"name": "nfs-idmapd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nfs-mountd.service": {
"name": "nfs-mountd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nfs-server.service": {
"name": "nfs-server.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"nfs-utils.service": {
"name": "nfs-utils.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nfsdcld.service": {
"name": "nfsdcld.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nftables.service": {
"name": "nftables.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"nis-domainname.service": {
"name": "nis-domainname.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"nm-priv-helper.service": {
"name": "nm-priv-helper.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"ntpd.service": {
"name": "ntpd.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"ntpdate.service": {
"name": "ntpdate.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"pam_namespace.service": {
"name": "pam_namespace.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"pcscd.service": {
"name": "pcscd.service",
"source": "systemd",
"state": "stopped",
"status": "indirect"
},
"plymouth-halt.service": {
"name": "plymouth-halt.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-kexec.service": {
"name": "plymouth-kexec.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-poweroff.service": {
"name": "plymouth-poweroff.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-quit-wait.service": {
"name": "plymouth-quit-wait.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-quit.service": {
"name": "plymouth-quit.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-read-write.service": {
"name": "plymouth-read-write.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-reboot.service": {
"name": "plymouth-reboot.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-start.service": {
"name": "plymouth-start.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-switch-root-initramfs.service": {
"name": "plymouth-switch-root-initramfs.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-switch-root.service": {
"name": "plymouth-switch-root.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"podman-auto-update.service": {
"name": "podman-auto-update.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"podman-clean-transient.service": {
"name": "podman-clean-transient.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"podman-kube@.service": {
"name": "podman-kube@.service",
"source": "systemd",
"state": "unknown",
"status": "disabled"
},
"podman-restart.service": {
"name": "podman-restart.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"podman.service": {
"name": "podman.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"polkit.service": {
"name": "polkit.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"quadlet-demo-mysql-volume.service": {
"name": "quadlet-demo-mysql-volume.service",
"source": "systemd",
"state": "stopped",
"status": "generated"
},
"quadlet-demo-mysql.service": {
"name": "quadlet-demo-mysql.service",
"source": "systemd",
"state": "stopped",
"status": "failed"
},
"quadlet-demo-network.service": {
"name": "quadlet-demo-network.service",
"source": "systemd",
"state": "stopped",
"status": "generated"
},
"quotaon.service": {
"name": "quotaon.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"raid-check.service": {
"name": "raid-check.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rc-local.service": {
"name": "rc-local.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rescue.service": {
"name": "rescue.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"restraintd.service": {
"name": "restraintd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"rngd.service": {
"name": "rngd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"rpc-gssd.service": {
"name": "rpc-gssd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rpc-statd-notify.service": {
"name": "rpc-statd-notify.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rpc-statd.service": {
"name": "rpc-statd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rpc-svcgssd.service": {
"name": "rpc-svcgssd.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"rpcbind.service": {
"name": "rpcbind.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"rpmdb-migrate.service": {
"name": "rpmdb-migrate.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"rpmdb-rebuild.service": {
"name": "rpmdb-rebuild.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"selinux-autorelabel-mark.service": {
"name": "selinux-autorelabel-mark.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"selinux-autorelabel.service": {
"name": "selinux-autorelabel.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"selinux-check-proper-disable.service": {
"name": "selinux-check-proper-disable.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"serial-getty@.service": {
"name": "serial-getty@.service",
"source": "systemd",
"state": "unknown",
"status": "indirect"
},
"serial-getty@ttyS0.service": {
"name": "serial-getty@ttyS0.service",
"source": "systemd",
"state": "running",
"status": "active"
},
"snapd.seeded.service": {
"name": "snapd.seeded.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"sntp.service": {
"name": "sntp.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"ssh-host-keys-migration.service": {
"name": "ssh-host-keys-migration.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"sshd-keygen.service": {
"name": "sshd-keygen.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"sshd-keygen@.service": {
"name": "sshd-keygen@.service",
"source": "systemd",
"state": "unknown",
"status": "disabled"
},
"sshd-keygen@ecdsa.service": {
"name": "sshd-keygen@ecdsa.service",
"source": "systemd",
"state": "stopped",
"status": "inactive"
},
"sshd-keygen@ed25519.service": {
"name": "sshd-keygen@ed25519.service",
"source": "systemd",
"state": "stopped",
"status": "inactive"
},
"sshd-keygen@rsa.service": {
"name": "sshd-keygen@rsa.service",
"source": "systemd",
"state": "stopped",
"status": "inactive"
},
"sshd.service": {
"name": "sshd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"sshd@.service": {
"name": "sshd@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"sssd-autofs.service": {
"name": "sssd-autofs.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-kcm.service": {
"name": "sssd-kcm.service",
"source": "systemd",
"state": "stopped",
"status": "indirect"
},
"sssd-nss.service": {
"name": "sssd-nss.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-pac.service": {
"name": "sssd-pac.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-pam.service": {
"name": "sssd-pam.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-ssh.service": {
"name": "sssd-ssh.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-sudo.service": {
"name": "sssd-sudo.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd.service": {
"name": "sssd.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"syslog.service": {
"name": "syslog.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"system-update-cleanup.service": {
"name": "system-update-cleanup.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-ask-password-console.service": {
"name": "systemd-ask-password-console.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-ask-password-plymouth.service": {
"name": "systemd-ask-password-plymouth.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-ask-password-wall.service": {
"name": "systemd-ask-password-wall.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-backlight@.service": {
"name": "systemd-backlight@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-battery-check.service": {
"name": "systemd-battery-check.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-binfmt.service": {
"name": "systemd-binfmt.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-bless-boot.service": {
"name": "systemd-bless-boot.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-boot-check-no-failures.service": {
"name": "systemd-boot-check-no-failures.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"systemd-boot-random-seed.service": {
"name": "systemd-boot-random-seed.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-boot-update.service": {
"name": "systemd-boot-update.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"systemd-confext.service": {
"name": "systemd-confext.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"systemd-coredump@.service": {
"name": "systemd-coredump@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-exit.service": {
"name": "systemd-exit.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-firstboot.service": {
"name": "systemd-firstboot.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-fsck-root.service": {
"name": "systemd-fsck-root.service",
"source": "systemd",
"state": "stopped",
"status": "enabled-runtime"
},
"systemd-fsck@.service": {
"name": "systemd-fsck@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-growfs-root.service": {
"name": "systemd-growfs-root.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-growfs@.service": {
"name": "systemd-growfs@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-halt.service": {
"name": "systemd-halt.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-hibernate.service": {
"name": "systemd-hibernate.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-homed-activate.service": {
"name": "systemd-homed-activate.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"systemd-homed.service": {
"name": "systemd-homed.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"systemd-hostnamed.service": {
"name": "systemd-hostnamed.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-hwdb-update.service": {
"name": "systemd-hwdb-update.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-hybrid-sleep.service": {
"name": "systemd-hybrid-sleep.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-initctl.service": {
"name": "systemd-initctl.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-journal-catalog-update.service": {
"name": "systemd-journal-catalog-update.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-journal-flush.service": {
"name": "systemd-journal-flush.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-journald.service": {
"name": "systemd-journald.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"systemd-journald@.service": {
"name": "systemd-journald@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-kexec.service": {
"name": "systemd-kexec.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-localed.service": {
"name": "systemd-localed.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-logind.service": {
"name": "systemd-logind.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"systemd-machine-id-commit.service": {
"name": "systemd-machine-id-commit.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-modules-load.service": {
"name": "systemd-modules-load.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-network-generator.service": {
"name": "systemd-network-generator.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"systemd-networkd-wait-online.service": {
"name": "systemd-networkd-wait-online.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"systemd-networkd-wait-online@.service": {
"name": "systemd-networkd-wait-online@.service",
"source": "systemd",
"state": "unknown",
"status": "disabled"
},
"systemd-networkd.service": {
"name": "systemd-networkd.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"systemd-oomd.service": {
"name": "systemd-oomd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"systemd-pcrfs-root.service": {
"name": "systemd-pcrfs-root.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-pcrfs@.service": {
"name": "systemd-pcrfs@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-pcrmachine.service": {
"name": "systemd-pcrmachine.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-pcrphase-initrd.service": {
"name": "systemd-pcrphase-initrd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-pcrphase-sysinit.service": {
"name": "systemd-pcrphase-sysinit.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-pcrphase.service": {
"name": "systemd-pcrphase.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-portabled.service": {
"name": "systemd-portabled.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-poweroff.service": {
"name": "systemd-poweroff.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-pstore.service": {
"name": "systemd-pstore.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"systemd-quotacheck.service": {
"name": "systemd-quotacheck.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-random-seed.service": {
"name": "systemd-random-seed.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-reboot.service": {
"name": "systemd-reboot.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-remount-fs.service": {
"name": "systemd-remount-fs.service",
"source": "systemd",
"state": "stopped",
"status": "enabled-runtime"
},
"systemd-repart.service": {
"name": "systemd-repart.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-resolved.service": {
"name": "systemd-resolved.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"systemd-rfkill.service": {
"name": "systemd-rfkill.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-soft-reboot.service": {
"name": "systemd-soft-reboot.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-suspend-then-hibernate.service": {
"name": "systemd-suspend-then-hibernate.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-suspend.service": {
"name": "systemd-suspend.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-sysctl.service": {
"name": "systemd-sysctl.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-sysext.service": {
"name": "systemd-sysext.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"systemd-sysupdate-reboot.service": {
"name": "systemd-sysupdate-reboot.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"systemd-sysupdate.service": {
"name": "systemd-sysupdate.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"systemd-sysusers.service": {
"name": "systemd-sysusers.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-time-wait-sync.service": {
"name": "systemd-time-wait-sync.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"systemd-timedated.service": {
"name": "systemd-timedated.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-timesyncd.service": {
"name": "systemd-timesyncd.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"systemd-tmpfiles-clean.service": {
"name": "systemd-tmpfiles-clean.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-tmpfiles-setup-dev-early.service": {
"name": "systemd-tmpfiles-setup-dev-early.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-tmpfiles-setup-dev.service": {
"name": "systemd-tmpfiles-setup-dev.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-tmpfiles-setup.service": {
"name": "systemd-tmpfiles-setup.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-udev-settle.service": {
"name": "systemd-udev-settle.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-udev-trigger.service": {
"name": "systemd-udev-trigger.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-udevd.service": {
"name": "systemd-udevd.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"systemd-update-done.service": {
"name": "systemd-update-done.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-update-utmp-runlevel.service": {
"name": "systemd-update-utmp-runlevel.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-update-utmp.service": {
"name": "systemd-update-utmp.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-user-sessions.service": {
"name": "systemd-user-sessions.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-userdbd.service": {
"name": "systemd-userdbd.service",
"source": "systemd",
"state": "running",
"status": "indirect"
},
"systemd-vconsole-setup.service": {
"name": "systemd-vconsole-setup.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-volatile-root.service": {
"name": "systemd-volatile-root.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-zram-setup@.service": {
"name": "systemd-zram-setup@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-zram-setup@zram0.service": {
"name": "systemd-zram-setup@zram0.service",
"source": "systemd",
"state": "stopped",
"status": "active"
},
"udisks2.service": {
"name": "udisks2.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"unbound-anchor.service": {
"name": "unbound-anchor.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"user-runtime-dir@.service": {
"name": "user-runtime-dir@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"user-runtime-dir@0.service": {
"name": "user-runtime-dir@0.service",
"source": "systemd",
"state": "stopped",
"status": "active"
},
"user@.service": {
"name": "user@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"user@0.service": {
"name": "user@0.service",
"source": "systemd",
"state": "running",
"status": "active"
}
}
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Create and update quadlets] ***********
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:114
Wednesday 10 July 2024 05:15:17 +0000 (0:00:02.182) 0:03:21.823 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_state != \"absent\"",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14
Wednesday 10 July 2024 05:15:17 +0000 (0:00:00.020) 0:03:21.844 ********
ok: [sut] => {
"ansible_facts": {
"__podman_quadlet_file_src": "",
"__podman_quadlet_spec": {},
"__podman_quadlet_str": "[Install]\nWantedBy=default.target\n\n[Container]\nImage=quay.io/linux-system-roles/mysql:5.6\nContainerName=quadlet-demo-mysql\nVolume=quadlet-demo-mysql.volume:/var/lib/mysql\nVolume=/tmp/quadlet_demo:/var/lib/quadlet_demo:Z\nNetwork=quadlet-demo.network\nSecret=mysql-root-password-container,type=env,target=MYSQL_ROOT_PASSWORD\nHealthCmd=/usr/bin/true\nHealthOnFailure=kill\n",
"__podman_quadlet_template_src": "quadlet-demo-mysql.container.j2"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25
Wednesday 10 July 2024 05:15:17 +0000 (0:00:00.070) 0:03:21.915 ********
ok: [sut] => {
"ansible_facts": {
"__podman_continue_if_pull_fails": false,
"__podman_pull_image": true,
"__podman_state": "absent",
"__podman_systemd_unit_scope": "",
"__podman_user": "root"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] *****
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35
Wednesday 10 July 2024 05:15:17 +0000 (0:00:00.026) 0:03:21.942 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_quadlet_str",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48
Wednesday 10 July 2024 05:15:17 +0000 (0:00:00.022) 0:03:21.964 ********
ok: [sut] => {
"ansible_facts": {
"__podman_quadlet_name": "quadlet-demo-mysql",
"__podman_quadlet_type": "container",
"__podman_rootless": false
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Check user and group information] *****
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57
Wednesday 10 July 2024 05:15:17 +0000 (0:00:00.073) 0:03:22.037 ********
included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for sut
TASK [fedora.linux_system_roles.podman : Get user information] *****************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2
Wednesday 10 July 2024 05:15:17 +0000 (0:00:00.042) 0:03:22.080 ********
skipping: [sut] => {
"changed": false,
"false_condition": "'getent_passwd' not in ansible_facts or __podman_user not in ansible_facts['getent_passwd']",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Fail if user does not exist] **********
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9
Wednesday 10 July 2024 05:15:17 +0000 (0:00:00.026) 0:03:22.106 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not ansible_facts[\"getent_passwd\"][__podman_user]",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set group for podman user] ************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16
Wednesday 10 July 2024 05:15:17 +0000 (0:00:00.025) 0:03:22.132 ********
ok: [sut] => {
"ansible_facts": {
"__podman_group": "0"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Get group information] ****************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:28
Wednesday 10 July 2024 05:15:17 +0000 (0:00:00.032) 0:03:22.165 ********
ok: [sut] => {
"ansible_facts": {
"getent_group": {
"root": [
"x",
"0",
""
]
}
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set group name] ***********************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:35
Wednesday 10 July 2024 05:15:17 +0000 (0:00:00.243) 0:03:22.408 ********
ok: [sut] => {
"ansible_facts": {
"__podman_group_name": "root"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : See if getsubids exists] **************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:39
Wednesday 10 July 2024 05:15:17 +0000 (0:00:00.032) 0:03:22.441 ********
ok: [sut] => {
"changed": false,
"stat": {
"atime": 1720588427.5164855,
"attr_flags": "e",
"attributes": [
"extents"
],
"block_size": 4096,
"blocks": 32,
"charset": "binary",
"checksum": "72ba28a3be19c467e89ab90f5d21f99b605159a7",
"ctime": 1720588377.3007717,
"dev": 51714,
"device_type": 0,
"executable": true,
"exists": true,
"gid": 0,
"gr_name": "root",
"inode": 203603,
"isblk": false,
"ischr": false,
"isdir": false,
"isfifo": false,
"isgid": false,
"islnk": false,
"isreg": true,
"issock": false,
"isuid": false,
"mimetype": "application/x-pie-executable",
"mode": "0755",
"mtime": 1696291200.0,
"nlink": 1,
"path": "/usr/bin/getsubids",
"pw_name": "root",
"readable": true,
"rgrp": true,
"roth": true,
"rusr": true,
"size": 16032,
"uid": 0,
"version": "2092964490",
"wgrp": false,
"woth": false,
"writeable": true,
"wusr": true,
"xgrp": true,
"xoth": true,
"xusr": true
}
}
TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:50
Wednesday 10 July 2024 05:15:18 +0000 (0:00:00.241) 0:03:22.682 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_user not in [\"root\", \"0\"]",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Check group with getsubids] ***********
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:55
Wednesday 10 July 2024 05:15:18 +0000 (0:00:00.021) 0:03:22.703 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_user not in [\"root\", \"0\"]",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ******
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:60
Wednesday 10 July 2024 05:15:18 +0000 (0:00:00.020) 0:03:22.724 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_user not in [\"root\", \"0\"]",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Get subuid file] **********************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:74
Wednesday 10 July 2024 05:15:18 +0000 (0:00:00.020) 0:03:22.744 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_stat_getsubids.stat.exists",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Get subgid file] **********************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:79
Wednesday 10 July 2024 05:15:18 +0000 (0:00:00.020) 0:03:22.764 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_stat_getsubids.stat.exists",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ******
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:84
Wednesday 10 July 2024 05:15:18 +0000 (0:00:00.019) 0:03:22.784 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_stat_getsubids.stat.exists",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ******
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:94
Wednesday 10 July 2024 05:15:18 +0000 (0:00:00.019) 0:03:22.804 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_stat_getsubids.stat.exists",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] *****
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:101
Wednesday 10 July 2024 05:15:18 +0000 (0:00:00.060) 0:03:22.865 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_stat_getsubids.stat.exists",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62
Wednesday 10 July 2024 05:15:18 +0000 (0:00:00.022) 0:03:22.888 ********
ok: [sut] => {
"ansible_facts": {
"__podman_activate_systemd_unit": true,
"__podman_images_found": [
"quay.io/linux-system-roles/mysql:5.6"
],
"__podman_kube_yamls_raw": "",
"__podman_service_name": "quadlet-demo-mysql.service",
"__podman_systemd_scope": "system",
"__podman_user_home_dir": "/root",
"__podman_xdg_runtime_dir": "/run/user/0"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73
Wednesday 10 July 2024 05:15:18 +0000 (0:00:00.039) 0:03:22.927 ********
ok: [sut] => {
"ansible_facts": {
"__podman_quadlet_path": "/etc/containers/systemd"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Get kube yaml contents] ***************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77
Wednesday 10 July 2024 05:15:18 +0000 (0:00:00.023) 0:03:22.950 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_state != \"absent\"",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87
Wednesday 10 July 2024 05:15:18 +0000 (0:00:00.020) 0:03:22.971 ********
ok: [sut] => {
"ansible_facts": {
"__podman_images": [
"quay.io/linux-system-roles/mysql:5.6"
],
"__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo-mysql.container",
"__podman_volumes": [
"/tmp/quadlet_demo"
]
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 6] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:103
Wednesday 10 July 2024 05:15:18 +0000 (0:00:00.049) 0:03:23.020 ********
ok: [sut] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : Cleanup quadlets] *********************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:110
Wednesday 10 July 2024 05:15:18 +0000 (0:00:00.026) 0:03:23.047 ********
included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml for sut
TASK [fedora.linux_system_roles.podman : Stat XDG_RUNTIME_DIR] *****************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:4
Wednesday 10 July 2024 05:15:18 +0000 (0:00:00.046) 0:03:23.093 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_rootless | bool",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Stop and disable service] *************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:12
Wednesday 10 July 2024 05:15:18 +0000 (0:00:00.020) 0:03:23.114 ********
changed: [sut] => {
"changed": true,
"enabled": false,
"failed_when_result": false,
"name": "quadlet-demo-mysql.service",
"state": "stopped",
"status": {
"AccessSELinuxContext": "system_u:object_r:systemd_unit_file_t:s0",
"ActiveEnterTimestamp": "Wed 2024-07-10 05:14:13 UTC",
"ActiveEnterTimestampMonotonic": "973531941",
"ActiveExitTimestamp": "Wed 2024-07-10 05:15:14 UTC",
"ActiveExitTimestampMonotonic": "1034753169",
"ActiveState": "failed",
"After": "basic.target system.slice systemd-journald.socket quadlet-demo-mysql-volume.service quadlet-demo-network.service sysinit.target -.mount tmp.mount",
"AllowIsolate": "no",
"AssertResult": "yes",
"AssertTimestamp": "Wed 2024-07-10 05:14:12 UTC",
"AssertTimestampMonotonic": "973072307",
"Before": "shutdown.target multi-user.target",
"BlockIOAccounting": "no",
"BlockIOWeight": "[not set]",
"CPUAccounting": "yes",
"CPUAffinityFromNUMA": "no",
"CPUQuotaPerSecUSec": "infinity",
"CPUQuotaPeriodUSec": "infinity",
"CPUSchedulingPolicy": "0",
"CPUSchedulingPriority": "0",
"CPUSchedulingResetOnFork": "no",
"CPUShares": "[not set]",
"CPUUsageNSec": "3750354000",
"CPUWeight": "[not set]",
"CacheDirectoryMode": "0755",
"CanFreeze": "yes",
"CanIsolate": "no",
"CanReload": "no",
"CanStart": "yes",
"CanStop": "yes",
"CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore",
"CleanResult": "success",
"CollectMode": "inactive",
"ConditionResult": "yes",
"ConditionTimestamp": "Wed 2024-07-10 05:14:12 UTC",
"ConditionTimestampMonotonic": "973072303",
"ConfigurationDirectoryMode": "0755",
"Conflicts": "shutdown.target",
"ControlGroupId": "7462",
"ControlPID": "0",
"CoredumpFilter": "0x33",
"DefaultDependencies": "yes",
"DefaultMemoryLow": "0",
"DefaultMemoryMin": "0",
"DefaultStartupMemoryLow": "0",
"Delegate": "yes",
"DelegateControllers": "cpu cpuset io memory pids",
"Description": "quadlet-demo-mysql.service",
"DevicePolicy": "auto",
"DropInPaths": "/usr/lib/systemd/system/service.d/10-timeout-abort.conf",
"DynamicUser": "no",
"Environment": "PODMAN_SYSTEMD_UNIT=quadlet-demo-mysql.service",
"ExecMainCode": "1",
"ExecMainExitTimestamp": "Wed 2024-07-10 05:15:14 UTC",
"ExecMainExitTimestampMonotonic": "1034735500",
"ExecMainPID": "8040",
"ExecMainStartTimestamp": "Wed 2024-07-10 05:14:13 UTC",
"ExecMainStartTimestampMonotonic": "973531872",
"ExecMainStatus": "137",
"ExecStart": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman run --name=quadlet-demo-mysql --cidfile=/run/quadlet-demo-mysql.cid --replace --rm --cgroups=split --network=systemd-quadlet-demo --sdnotify=conmon -d -v systemd-quadlet-demo-mysql:/var/lib/mysql -v /tmp/quadlet_demo:/var/lib/quadlet_demo:Z --secret mysql-root-password-container,type=env,target=MYSQL_ROOT_PASSWORD --health-cmd /usr/bin/true --health-on-failure kill quay.io/linux-system-roles/mysql:5.6 ; ignore_errors=no ; start_time=[Wed 2024-07-10 05:14:13 UTC] ; stop_time=[Wed 2024-07-10 05:15:14 UTC] ; pid=8040 ; code=exited ; status=137 }",
"ExecStartEx": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman run --name=quadlet-demo-mysql --cidfile=/run/quadlet-demo-mysql.cid --replace --rm --cgroups=split --network=systemd-quadlet-demo --sdnotify=conmon -d -v systemd-quadlet-demo-mysql:/var/lib/mysql -v /tmp/quadlet_demo:/var/lib/quadlet_demo:Z --secret mysql-root-password-container,type=env,target=MYSQL_ROOT_PASSWORD --health-cmd /usr/bin/true --health-on-failure kill quay.io/linux-system-roles/mysql:5.6 ; flags= ; start_time=[Wed 2024-07-10 05:14:13 UTC] ; stop_time=[Wed 2024-07-10 05:15:14 UTC] ; pid=8040 ; code=exited ; status=137 }",
"ExecStop": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman rm -v -f -i --cidfile=/run/quadlet-demo-mysql.cid ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"ExecStopEx": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman rm -v -f -i --cidfile=/run/quadlet-demo-mysql.cid ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"ExecStopPost": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman rm -v -f -i --cidfile=/run/quadlet-demo-mysql.cid ; ignore_errors=yes ; start_time=[Wed 2024-07-10 05:15:14 UTC] ; stop_time=[Wed 2024-07-10 05:15:14 UTC] ; pid=11298 ; code=exited ; status=0 }",
"ExecStopPostEx": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman rm -v -f -i --cidfile=/run/quadlet-demo-mysql.cid ; flags=ignore-failure ; start_time=[Wed 2024-07-10 05:15:14 UTC] ; stop_time=[Wed 2024-07-10 05:15:14 UTC] ; pid=11298 ; code=exited ; status=0 }",
"ExitType": "main",
"ExtensionImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent",
"FailureAction": "none",
"FileDescriptorStoreMax": "0",
"FileDescriptorStorePreserve": "restart",
"FinalKillSignal": "9",
"FragmentPath": "/run/systemd/generator/quadlet-demo-mysql.service",
"FreezerState": "running",
"GID": "[not set]",
"GuessMainPID": "yes",
"IOAccounting": "no",
"IOReadBytes": "18446744073709551615",
"IOReadOperations": "18446744073709551615",
"IOSchedulingClass": "2",
"IOSchedulingPriority": "4",
"IOWeight": "[not set]",
"IOWriteBytes": "18446744073709551615",
"IOWriteOperations": "18446744073709551615",
"IPAccounting": "no",
"IPEgressBytes": "[no data]",
"IPEgressPackets": "[no data]",
"IPIngressBytes": "[no data]",
"IPIngressPackets": "[no data]",
"Id": "quadlet-demo-mysql.service",
"IgnoreOnIsolate": "no",
"IgnoreSIGPIPE": "yes",
"InactiveEnterTimestamp": "Wed 2024-07-10 05:15:14 UTC",
"InactiveEnterTimestampMonotonic": "1034788061",
"InactiveExitTimestamp": "Wed 2024-07-10 05:14:12 UTC",
"InactiveExitTimestampMonotonic": "973080495",
"InvocationID": "16ff4258ca4845f99e8a96e4512592e3",
"JobRunningTimeoutUSec": "infinity",
"JobTimeoutAction": "none",
"JobTimeoutUSec": "infinity",
"KeyringMode": "private",
"KillMode": "mixed",
"KillSignal": "15",
"LimitAS": "infinity",
"LimitASSoft": "infinity",
"LimitCORE": "infinity",
"LimitCORESoft": "infinity",
"LimitCPU": "infinity",
"LimitCPUSoft": "infinity",
"LimitDATA": "infinity",
"LimitDATASoft": "infinity",
"LimitFSIZE": "infinity",
"LimitFSIZESoft": "infinity",
"LimitLOCKS": "infinity",
"LimitLOCKSSoft": "infinity",
"LimitMEMLOCK": "8388608",
"LimitMEMLOCKSoft": "8388608",
"LimitMSGQUEUE": "819200",
"LimitMSGQUEUESoft": "819200",
"LimitNICE": "0",
"LimitNICESoft": "0",
"LimitNOFILE": "524288",
"LimitNOFILESoft": "1024",
"LimitNPROC": "14756",
"LimitNPROCSoft": "14756",
"LimitRSS": "infinity",
"LimitRSSSoft": "infinity",
"LimitRTPRIO": "0",
"LimitRTPRIOSoft": "0",
"LimitRTTIME": "infinity",
"LimitRTTIMESoft": "infinity",
"LimitSIGPENDING": "14756",
"LimitSIGPENDINGSoft": "14756",
"LimitSTACK": "infinity",
"LimitSTACKSoft": "8388608",
"LoadState": "loaded",
"LockPersonality": "no",
"LogLevelMax": "-1",
"LogRateLimitBurst": "0",
"LogRateLimitIntervalUSec": "0",
"LogsDirectoryMode": "0755",
"MainPID": "0",
"ManagedOOMMemoryPressure": "auto",
"ManagedOOMMemoryPressureLimit": "0",
"ManagedOOMPreference": "none",
"ManagedOOMSwap": "auto",
"MemoryAccounting": "yes",
"MemoryAvailable": "infinity",
"MemoryCurrent": "[not set]",
"MemoryDenyWriteExecute": "no",
"MemoryHigh": "infinity",
"MemoryKSM": "no",
"MemoryLimit": "infinity",
"MemoryLow": "0",
"MemoryMax": "infinity",
"MemoryMin": "0",
"MemoryPressureThresholdUSec": "200ms",
"MemoryPressureWatch": "auto",
"MemorySwapMax": "infinity",
"MemoryZSwapMax": "infinity",
"MountAPIVFS": "no",
"MountImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent",
"NFileDescriptorStore": "0",
"NRestarts": "0",
"NUMAPolicy": "n/a",
"Names": "quadlet-demo-mysql.service",
"NeedDaemonReload": "no",
"Nice": "0",
"NoNewPrivileges": "no",
"NonBlocking": "no",
"NotifyAccess": "all",
"OOMPolicy": "continue",
"OOMScoreAdjust": "0",
"OnFailureJobMode": "replace",
"OnSuccessJobMode": "fail",
"Perpetual": "no",
"PrivateDevices": "no",
"PrivateIPC": "no",
"PrivateMounts": "no",
"PrivateNetwork": "no",
"PrivateTmp": "no",
"PrivateUsers": "no",
"ProcSubset": "all",
"ProtectClock": "no",
"ProtectControlGroups": "no",
"ProtectHome": "no",
"ProtectHostname": "no",
"ProtectKernelLogs": "no",
"ProtectKernelModules": "no",
"ProtectKernelTunables": "no",
"ProtectProc": "default",
"ProtectSystem": "no",
"RefuseManualStart": "no",
"RefuseManualStop": "no",
"ReloadResult": "success",
"ReloadSignal": "1",
"RemainAfterExit": "no",
"RemoveIPC": "no",
"Requires": "quadlet-demo-network.service -.mount sysinit.target system.slice quadlet-demo-mysql-volume.service tmp.mount",
"RequiresMountsFor": "/tmp/quadlet_demo /run/containers",
"Restart": "no",
"RestartKillSignal": "15",
"RestartMaxDelayUSec": "infinity",
"RestartMode": "normal",
"RestartSteps": "0",
"RestartUSec": "100ms",
"RestartUSecNext": "100ms",
"RestrictNamespaces": "no",
"RestrictRealtime": "no",
"RestrictSUIDSGID": "no",
"Result": "exit-code",
"RootDirectoryStartOnly": "no",
"RootEphemeral": "no",
"RootImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent",
"RuntimeDirectoryMode": "0755",
"RuntimeDirectoryPreserve": "no",
"RuntimeMaxUSec": "infinity",
"RuntimeRandomizedExtraUSec": "0",
"SameProcessGroup": "no",
"SecureBits": "0",
"SendSIGHUP": "no",
"SendSIGKILL": "yes",
"Slice": "system.slice",
"SourcePath": "/etc/containers/systemd/quadlet-demo-mysql.container",
"StandardError": "inherit",
"StandardInput": "null",
"StandardOutput": "journal",
"StartLimitAction": "none",
"StartLimitBurst": "5",
"StartLimitIntervalUSec": "10s",
"StartupBlockIOWeight": "[not set]",
"StartupCPUShares": "[not set]",
"StartupCPUWeight": "[not set]",
"StartupIOWeight": "[not set]",
"StartupMemoryHigh": "18446744073709551615",
"StartupMemoryLow": "0",
"StartupMemoryMax": "18446744073709551615",
"StartupMemorySwapMax": "18446744073709551615",
"StartupMemoryZSwapMax": "18446744073709551615",
"StateChangeTimestamp": "Wed 2024-07-10 05:15:14 UTC",
"StateChangeTimestampMonotonic": "1034788061",
"StateDirectoryMode": "0755",
"StatusErrno": "0",
"StopWhenUnneeded": "no",
"SubState": "failed",
"SuccessAction": "none",
"SyslogFacility": "3",
"SyslogIdentifier": "quadlet-demo-mysql",
"SyslogLevel": "6",
"SyslogLevelPrefix": "yes",
"SyslogPriority": "30",
"SystemCallErrorNumber": "2147483646",
"TTYReset": "no",
"TTYVHangup": "no",
"TTYVTDisallocate": "no",
"TasksAccounting": "yes",
"TasksCurrent": "[not set]",
"TasksMax": "4426",
"TimeoutAbortUSec": "45s",
"TimeoutCleanUSec": "infinity",
"TimeoutStartFailureMode": "terminate",
"TimeoutStartUSec": "45s",
"TimeoutStopFailureMode": "abort",
"TimeoutStopUSec": "45s",
"TimerSlackNSec": "50000",
"Transient": "no",
"Type": "notify",
"UID": "[not set]",
"UMask": "0022",
"UnitFilePreset": "disabled",
"UnitFileState": "generated",
"UtmpMode": "init",
"WantedBy": "multi-user.target",
"WatchdogSignal": "6",
"WatchdogTimestampMonotonic": "0",
"WatchdogUSec": "0"
}
}
TASK [fedora.linux_system_roles.podman : Remove quadlet file] ******************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:33
Wednesday 10 July 2024 05:15:19 +0000 (0:00:00.837) 0:03:23.952 ********
changed: [sut] => {
"changed": true,
"path": "/etc/containers/systemd/quadlet-demo-mysql.container",
"state": "absent"
}
TASK [fedora.linux_system_roles.podman : Manage linger] ************************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:39
Wednesday 10 July 2024 05:15:19 +0000 (0:00:00.240) 0:03:24.192 ********
included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for sut
TASK [fedora.linux_system_roles.podman : Enable linger if needed] **************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12
Wednesday 10 July 2024 05:15:19 +0000 (0:00:00.034) 0:03:24.226 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_rootless | bool",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18
Wednesday 10 July 2024 05:15:19 +0000 (0:00:00.062) 0:03:24.289 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_rootless | bool",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22
Wednesday 10 July 2024 05:15:19 +0000 (0:00:00.022) 0:03:24.311 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_rootless | bool",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Reload systemctl] *********************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:47
Wednesday 10 July 2024 05:15:19 +0000 (0:00:00.020) 0:03:24.332 ********
ok: [sut] => {
"changed": false,
"name": null,
"status": {}
}
TASK [fedora.linux_system_roles.podman : Prune images no longer in use] ********
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:56
Wednesday 10 July 2024 05:15:20 +0000 (0:00:00.878) 0:03:25.210 ********
changed: [sut] => {
"changed": true,
"cmd": [
"podman",
"image",
"prune",
"--all",
"-f"
],
"delta": "0:00:00.229999",
"end": "2024-07-10 05:15:20.974256",
"rc": 0,
"start": "2024-07-10 05:15:20.744257"
}
STDOUT:
dd3b2a5dcb48ff61113592ed5ddd762581be4387c7bc552375a2159422aa6bf5
TASK [fedora.linux_system_roles.podman : Remove networks no longer in use] *****
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:64
Wednesday 10 July 2024 05:15:21 +0000 (0:00:00.466) 0:03:25.677 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_quadlet_type == \"network\"",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Remove volumes no longer in use] ******
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:73
Wednesday 10 July 2024 05:15:21 +0000 (0:00:00.022) 0:03:25.699 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_quadlet_type == \"volume\"",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : For testing and debugging - images] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:82
Wednesday 10 July 2024 05:15:21 +0000 (0:00:00.022) 0:03:25.722 ********
ok: [sut] => {
"changed": false,
"cmd": [
"podman",
"images",
"-n"
],
"delta": "0:00:00.036569",
"end": "2024-07-10 05:15:21.294266",
"rc": 0,
"start": "2024-07-10 05:15:21.257697"
}
TASK [fedora.linux_system_roles.podman : For testing and debugging - volumes] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:92
Wednesday 10 July 2024 05:15:21 +0000 (0:00:00.275) 0:03:25.998 ********
ok: [sut] => {
"changed": false,
"cmd": [
"podman",
"volume",
"ls",
"-n"
],
"delta": "0:00:00.036158",
"end": "2024-07-10 05:15:21.569057",
"rc": 0,
"start": "2024-07-10 05:15:21.532899"
}
STDOUT:
local systemd-quadlet-demo-mysql
local wp-pv-claim
local envoy-proxy-config
local envoy-certificates
TASK [fedora.linux_system_roles.podman : For testing and debugging - containers] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:102
Wednesday 10 July 2024 05:15:21 +0000 (0:00:00.274) 0:03:26.272 ********
ok: [sut] => {
"changed": false,
"cmd": [
"podman",
"ps",
"--noheading"
],
"delta": "0:00:00.036024",
"end": "2024-07-10 05:15:21.845993",
"rc": 0,
"start": "2024-07-10 05:15:21.809969"
}
TASK [fedora.linux_system_roles.podman : For testing and debugging - networks] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:112
Wednesday 10 July 2024 05:15:21 +0000 (0:00:00.274) 0:03:26.547 ********
ok: [sut] => {
"changed": false,
"cmd": [
"podman",
"network",
"ls",
"-n",
"-q"
],
"delta": "0:00:00.035355",
"end": "2024-07-10 05:15:22.120706",
"rc": 0,
"start": "2024-07-10 05:15:22.085351"
}
STDOUT:
podman
systemd-quadlet-demo
TASK [fedora.linux_system_roles.podman : For testing and debugging - secrets] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:122
Wednesday 10 July 2024 05:15:22 +0000 (0:00:00.276) 0:03:26.823 ********
ok: [sut] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : For testing and debugging - services] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:133
Wednesday 10 July 2024 05:15:22 +0000 (0:00:00.267) 0:03:27.090 ********
ok: [sut] => {
"ansible_facts": {
"services": {
"NetworkManager-dispatcher.service": {
"name": "NetworkManager-dispatcher.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"NetworkManager-wait-online.service": {
"name": "NetworkManager-wait-online.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"NetworkManager.service": {
"name": "NetworkManager.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"arp-ethers.service": {
"name": "arp-ethers.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"auditd.service": {
"name": "auditd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"auth-rpcgss-module.service": {
"name": "auth-rpcgss-module.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"autovt@.service": {
"name": "autovt@.service",
"source": "systemd",
"state": "unknown",
"status": "alias"
},
"blk-availability.service": {
"name": "blk-availability.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"bluetooth.service": {
"name": "bluetooth.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"certmonger.service": {
"name": "certmonger.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"chrony-wait.service": {
"name": "chrony-wait.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"chronyd-restricted.service": {
"name": "chronyd-restricted.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"chronyd.service": {
"name": "chronyd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"cloud-config.service": {
"name": "cloud-config.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"cloud-final.service": {
"name": "cloud-final.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"cloud-init-hotplugd.service": {
"name": "cloud-init-hotplugd.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"cloud-init-local.service": {
"name": "cloud-init-local.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"cloud-init.service": {
"name": "cloud-init.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"console-getty.service": {
"name": "console-getty.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"container-getty@.service": {
"name": "container-getty@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"dbus-broker.service": {
"name": "dbus-broker.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"dbus-org.bluez.service": {
"name": "dbus-org.bluez.service",
"source": "systemd",
"state": "inactive",
"status": "alias"
},
"dbus-org.fedoraproject.FirewallD1.service": {
"name": "dbus-org.fedoraproject.FirewallD1.service",
"source": "systemd",
"state": "active",
"status": "alias"
},
"dbus-org.freedesktop.home1.service": {
"name": "dbus-org.freedesktop.home1.service",
"source": "systemd",
"state": "active",
"status": "alias"
},
"dbus-org.freedesktop.hostname1.service": {
"name": "dbus-org.freedesktop.hostname1.service",
"source": "systemd",
"state": "inactive",
"status": "alias"
},
"dbus-org.freedesktop.locale1.service": {
"name": "dbus-org.freedesktop.locale1.service",
"source": "systemd",
"state": "inactive",
"status": "alias"
},
"dbus-org.freedesktop.login1.service": {
"name": "dbus-org.freedesktop.login1.service",
"source": "systemd",
"state": "active",
"status": "alias"
},
"dbus-org.freedesktop.nm-dispatcher.service": {
"name": "dbus-org.freedesktop.nm-dispatcher.service",
"source": "systemd",
"state": "active",
"status": "alias"
},
"dbus-org.freedesktop.oom1.service": {
"name": "dbus-org.freedesktop.oom1.service",
"source": "systemd",
"state": "active",
"status": "alias"
},
"dbus-org.freedesktop.portable1.service": {
"name": "dbus-org.freedesktop.portable1.service",
"source": "systemd",
"state": "inactive",
"status": "alias"
},
"dbus-org.freedesktop.resolve1.service": {
"name": "dbus-org.freedesktop.resolve1.service",
"source": "systemd",
"state": "active",
"status": "alias"
},
"dbus-org.freedesktop.timedate1.service": {
"name": "dbus-org.freedesktop.timedate1.service",
"source": "systemd",
"state": "inactive",
"status": "alias"
},
"dbus.service": {
"name": "dbus.service",
"source": "systemd",
"state": "active",
"status": "alias"
},
"debug-shell.service": {
"name": "debug-shell.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"display-manager.service": {
"name": "display-manager.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"dm-event.service": {
"name": "dm-event.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dnf-makecache.service": {
"name": "dnf-makecache.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dnf-system-upgrade-cleanup.service": {
"name": "dnf-system-upgrade-cleanup.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dnf-system-upgrade.service": {
"name": "dnf-system-upgrade.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"dracut-cmdline.service": {
"name": "dracut-cmdline.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-initqueue.service": {
"name": "dracut-initqueue.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-mount.service": {
"name": "dracut-mount.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-mount.service": {
"name": "dracut-pre-mount.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-pivot.service": {
"name": "dracut-pre-pivot.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-trigger.service": {
"name": "dracut-pre-trigger.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-udev.service": {
"name": "dracut-pre-udev.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-shutdown-onfailure.service": {
"name": "dracut-shutdown-onfailure.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-shutdown.service": {
"name": "dracut-shutdown.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"ebtables.service": {
"name": "ebtables.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"emergency.service": {
"name": "emergency.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"firewalld.service": {
"name": "firewalld.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"fsidd.service": {
"name": "fsidd.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"fstrim.service": {
"name": "fstrim.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"fwupd-offline-update.service": {
"name": "fwupd-offline-update.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"fwupd-refresh.service": {
"name": "fwupd-refresh.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"fwupd.service": {
"name": "fwupd.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"getty@.service": {
"name": "getty@.service",
"source": "systemd",
"state": "unknown",
"status": "enabled"
},
"getty@tty1.service": {
"name": "getty@tty1.service",
"source": "systemd",
"state": "running",
"status": "active"
},
"grub-boot-indeterminate.service": {
"name": "grub-boot-indeterminate.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"grub2-systemd-integration.service": {
"name": "grub2-systemd-integration.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"gssproxy.service": {
"name": "gssproxy.service",
"source": "systemd",
"state": "running",
"status": "disabled"
},
"hv_kvp_daemon.service": {
"name": "hv_kvp_daemon.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"import-state.service": {
"name": "import-state.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"initrd-cleanup.service": {
"name": "initrd-cleanup.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"initrd-parse-etc.service": {
"name": "initrd-parse-etc.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"initrd-switch-root.service": {
"name": "initrd-switch-root.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"initrd-udevadm-cleanup-db.service": {
"name": "initrd-udevadm-cleanup-db.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"ip6tables.service": {
"name": "ip6tables.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"ipset.service": {
"name": "ipset.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"iptables.service": {
"name": "iptables.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"kmod-static-nodes.service": {
"name": "kmod-static-nodes.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"ldconfig.service": {
"name": "ldconfig.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"loadmodules.service": {
"name": "loadmodules.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"lvm2-lvmpolld.service": {
"name": "lvm2-lvmpolld.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"lvm2-monitor.service": {
"name": "lvm2-monitor.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"man-db-cache-update.service": {
"name": "man-db-cache-update.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"man-db-restart-cache-update.service": {
"name": "man-db-restart-cache-update.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"mdadm-grow-continue@.service": {
"name": "mdadm-grow-continue@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"mdadm-last-resort@.service": {
"name": "mdadm-last-resort@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"mdcheck_continue.service": {
"name": "mdcheck_continue.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"mdcheck_start.service": {
"name": "mdcheck_start.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"mdmon@.service": {
"name": "mdmon@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"mdmonitor-oneshot.service": {
"name": "mdmonitor-oneshot.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"mdmonitor.service": {
"name": "mdmonitor.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"modprobe@.service": {
"name": "modprobe@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"modprobe@configfs.service": {
"name": "modprobe@configfs.service",
"source": "systemd",
"state": "stopped",
"status": "inactive"
},
"modprobe@dm_mod.service": {
"name": "modprobe@dm_mod.service",
"source": "systemd",
"state": "stopped",
"status": "inactive"
},
"modprobe@drm.service": {
"name": "modprobe@drm.service",
"source": "systemd",
"state": "stopped",
"status": "inactive"
},
"modprobe@efi_pstore.service": {
"name": "modprobe@efi_pstore.service",
"source": "systemd",
"state": "stopped",
"status": "inactive"
},
"modprobe@fuse.service": {
"name": "modprobe@fuse.service",
"source": "systemd",
"state": "stopped",
"status": "inactive"
},
"modprobe@loop.service": {
"name": "modprobe@loop.service",
"source": "systemd",
"state": "stopped",
"status": "inactive"
},
"netavark-dhcp-proxy.service": {
"name": "netavark-dhcp-proxy.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"netavark-firewalld-reload.service": {
"name": "netavark-firewalld-reload.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"network.service": {
"name": "network.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"nfs-blkmap.service": {
"name": "nfs-blkmap.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"nfs-idmapd.service": {
"name": "nfs-idmapd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nfs-mountd.service": {
"name": "nfs-mountd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nfs-server.service": {
"name": "nfs-server.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"nfs-utils.service": {
"name": "nfs-utils.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nfsdcld.service": {
"name": "nfsdcld.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nftables.service": {
"name": "nftables.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"nis-domainname.service": {
"name": "nis-domainname.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"nm-priv-helper.service": {
"name": "nm-priv-helper.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"ntpd.service": {
"name": "ntpd.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"ntpdate.service": {
"name": "ntpdate.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"pam_namespace.service": {
"name": "pam_namespace.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"pcscd.service": {
"name": "pcscd.service",
"source": "systemd",
"state": "stopped",
"status": "indirect"
},
"plymouth-halt.service": {
"name": "plymouth-halt.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-kexec.service": {
"name": "plymouth-kexec.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-poweroff.service": {
"name": "plymouth-poweroff.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-quit-wait.service": {
"name": "plymouth-quit-wait.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-quit.service": {
"name": "plymouth-quit.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-read-write.service": {
"name": "plymouth-read-write.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-reboot.service": {
"name": "plymouth-reboot.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-start.service": {
"name": "plymouth-start.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-switch-root-initramfs.service": {
"name": "plymouth-switch-root-initramfs.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-switch-root.service": {
"name": "plymouth-switch-root.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"podman-auto-update.service": {
"name": "podman-auto-update.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"podman-clean-transient.service": {
"name": "podman-clean-transient.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"podman-kube@.service": {
"name": "podman-kube@.service",
"source": "systemd",
"state": "unknown",
"status": "disabled"
},
"podman-restart.service": {
"name": "podman-restart.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"podman.service": {
"name": "podman.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"polkit.service": {
"name": "polkit.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"quadlet-demo-mysql-volume.service": {
"name": "quadlet-demo-mysql-volume.service",
"source": "systemd",
"state": "stopped",
"status": "generated"
},
"quadlet-demo-mysql.service": {
"name": "quadlet-demo-mysql.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"quadlet-demo-network.service": {
"name": "quadlet-demo-network.service",
"source": "systemd",
"state": "stopped",
"status": "generated"
},
"quotaon.service": {
"name": "quotaon.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"raid-check.service": {
"name": "raid-check.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rc-local.service": {
"name": "rc-local.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rescue.service": {
"name": "rescue.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"restraintd.service": {
"name": "restraintd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"rngd.service": {
"name": "rngd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"rpc-gssd.service": {
"name": "rpc-gssd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rpc-statd-notify.service": {
"name": "rpc-statd-notify.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rpc-statd.service": {
"name": "rpc-statd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rpc-svcgssd.service": {
"name": "rpc-svcgssd.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"rpcbind.service": {
"name": "rpcbind.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"rpmdb-migrate.service": {
"name": "rpmdb-migrate.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"rpmdb-rebuild.service": {
"name": "rpmdb-rebuild.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"selinux-autorelabel-mark.service": {
"name": "selinux-autorelabel-mark.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"selinux-autorelabel.service": {
"name": "selinux-autorelabel.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"selinux-check-proper-disable.service": {
"name": "selinux-check-proper-disable.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"serial-getty@.service": {
"name": "serial-getty@.service",
"source": "systemd",
"state": "unknown",
"status": "indirect"
},
"serial-getty@ttyS0.service": {
"name": "serial-getty@ttyS0.service",
"source": "systemd",
"state": "running",
"status": "active"
},
"snapd.seeded.service": {
"name": "snapd.seeded.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"sntp.service": {
"name": "sntp.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"ssh-host-keys-migration.service": {
"name": "ssh-host-keys-migration.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"sshd-keygen.service": {
"name": "sshd-keygen.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"sshd-keygen@.service": {
"name": "sshd-keygen@.service",
"source": "systemd",
"state": "unknown",
"status": "disabled"
},
"sshd-keygen@ecdsa.service": {
"name": "sshd-keygen@ecdsa.service",
"source": "systemd",
"state": "stopped",
"status": "inactive"
},
"sshd-keygen@ed25519.service": {
"name": "sshd-keygen@ed25519.service",
"source": "systemd",
"state": "stopped",
"status": "inactive"
},
"sshd-keygen@rsa.service": {
"name": "sshd-keygen@rsa.service",
"source": "systemd",
"state": "stopped",
"status": "inactive"
},
"sshd.service": {
"name": "sshd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"sshd@.service": {
"name": "sshd@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"sssd-autofs.service": {
"name": "sssd-autofs.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-kcm.service": {
"name": "sssd-kcm.service",
"source": "systemd",
"state": "stopped",
"status": "indirect"
},
"sssd-nss.service": {
"name": "sssd-nss.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-pac.service": {
"name": "sssd-pac.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-pam.service": {
"name": "sssd-pam.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-ssh.service": {
"name": "sssd-ssh.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-sudo.service": {
"name": "sssd-sudo.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd.service": {
"name": "sssd.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"syslog.service": {
"name": "syslog.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"system-update-cleanup.service": {
"name": "system-update-cleanup.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-ask-password-console.service": {
"name": "systemd-ask-password-console.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-ask-password-plymouth.service": {
"name": "systemd-ask-password-plymouth.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-ask-password-wall.service": {
"name": "systemd-ask-password-wall.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-backlight@.service": {
"name": "systemd-backlight@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-battery-check.service": {
"name": "systemd-battery-check.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-binfmt.service": {
"name": "systemd-binfmt.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-bless-boot.service": {
"name": "systemd-bless-boot.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-boot-check-no-failures.service": {
"name": "systemd-boot-check-no-failures.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"systemd-boot-random-seed.service": {
"name": "systemd-boot-random-seed.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-boot-update.service": {
"name": "systemd-boot-update.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"systemd-confext.service": {
"name": "systemd-confext.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"systemd-coredump@.service": {
"name": "systemd-coredump@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-exit.service": {
"name": "systemd-exit.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-firstboot.service": {
"name": "systemd-firstboot.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-fsck-root.service": {
"name": "systemd-fsck-root.service",
"source": "systemd",
"state": "stopped",
"status": "enabled-runtime"
},
"systemd-fsck@.service": {
"name": "systemd-fsck@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-growfs-root.service": {
"name": "systemd-growfs-root.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-growfs@.service": {
"name": "systemd-growfs@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-halt.service": {
"name": "systemd-halt.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-hibernate.service": {
"name": "systemd-hibernate.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-homed-activate.service": {
"name": "systemd-homed-activate.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"systemd-homed.service": {
"name": "systemd-homed.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"systemd-hostnamed.service": {
"name": "systemd-hostnamed.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-hwdb-update.service": {
"name": "systemd-hwdb-update.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-hybrid-sleep.service": {
"name": "systemd-hybrid-sleep.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-initctl.service": {
"name": "systemd-initctl.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-journal-catalog-update.service": {
"name": "systemd-journal-catalog-update.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-journal-flush.service": {
"name": "systemd-journal-flush.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-journald.service": {
"name": "systemd-journald.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"systemd-journald@.service": {
"name": "systemd-journald@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-kexec.service": {
"name": "systemd-kexec.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-localed.service": {
"name": "systemd-localed.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-logind.service": {
"name": "systemd-logind.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"systemd-machine-id-commit.service": {
"name": "systemd-machine-id-commit.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-modules-load.service": {
"name": "systemd-modules-load.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-network-generator.service": {
"name": "systemd-network-generator.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"systemd-networkd-wait-online.service": {
"name": "systemd-networkd-wait-online.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"systemd-networkd-wait-online@.service": {
"name": "systemd-networkd-wait-online@.service",
"source": "systemd",
"state": "unknown",
"status": "disabled"
},
"systemd-networkd.service": {
"name": "systemd-networkd.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"systemd-oomd.service": {
"name": "systemd-oomd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"systemd-pcrfs-root.service": {
"name": "systemd-pcrfs-root.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-pcrfs@.service": {
"name": "systemd-pcrfs@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-pcrmachine.service": {
"name": "systemd-pcrmachine.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-pcrphase-initrd.service": {
"name": "systemd-pcrphase-initrd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-pcrphase-sysinit.service": {
"name": "systemd-pcrphase-sysinit.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-pcrphase.service": {
"name": "systemd-pcrphase.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-portabled.service": {
"name": "systemd-portabled.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-poweroff.service": {
"name": "systemd-poweroff.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-pstore.service": {
"name": "systemd-pstore.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"systemd-quotacheck.service": {
"name": "systemd-quotacheck.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-random-seed.service": {
"name": "systemd-random-seed.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-reboot.service": {
"name": "systemd-reboot.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-remount-fs.service": {
"name": "systemd-remount-fs.service",
"source": "systemd",
"state": "stopped",
"status": "enabled-runtime"
},
"systemd-repart.service": {
"name": "systemd-repart.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-resolved.service": {
"name": "systemd-resolved.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"systemd-rfkill.service": {
"name": "systemd-rfkill.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-soft-reboot.service": {
"name": "systemd-soft-reboot.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-suspend-then-hibernate.service": {
"name": "systemd-suspend-then-hibernate.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-suspend.service": {
"name": "systemd-suspend.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-sysctl.service": {
"name": "systemd-sysctl.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-sysext.service": {
"name": "systemd-sysext.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"systemd-sysupdate-reboot.service": {
"name": "systemd-sysupdate-reboot.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"systemd-sysupdate.service": {
"name": "systemd-sysupdate.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"systemd-sysusers.service": {
"name": "systemd-sysusers.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-time-wait-sync.service": {
"name": "systemd-time-wait-sync.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"systemd-timedated.service": {
"name": "systemd-timedated.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-timesyncd.service": {
"name": "systemd-timesyncd.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"systemd-tmpfiles-clean.service": {
"name": "systemd-tmpfiles-clean.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-tmpfiles-setup-dev-early.service": {
"name": "systemd-tmpfiles-setup-dev-early.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-tmpfiles-setup-dev.service": {
"name": "systemd-tmpfiles-setup-dev.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-tmpfiles-setup.service": {
"name": "systemd-tmpfiles-setup.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-udev-settle.service": {
"name": "systemd-udev-settle.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-udev-trigger.service": {
"name": "systemd-udev-trigger.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-udevd.service": {
"name": "systemd-udevd.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"systemd-update-done.service": {
"name": "systemd-update-done.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-update-utmp-runlevel.service": {
"name": "systemd-update-utmp-runlevel.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-update-utmp.service": {
"name": "systemd-update-utmp.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-user-sessions.service": {
"name": "systemd-user-sessions.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-userdbd.service": {
"name": "systemd-userdbd.service",
"source": "systemd",
"state": "running",
"status": "indirect"
},
"systemd-vconsole-setup.service": {
"name": "systemd-vconsole-setup.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-volatile-root.service": {
"name": "systemd-volatile-root.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-zram-setup@.service": {
"name": "systemd-zram-setup@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-zram-setup@zram0.service": {
"name": "systemd-zram-setup@zram0.service",
"source": "systemd",
"state": "stopped",
"status": "active"
},
"udisks2.service": {
"name": "udisks2.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"unbound-anchor.service": {
"name": "unbound-anchor.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"user-runtime-dir@.service": {
"name": "user-runtime-dir@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"user-runtime-dir@0.service": {
"name": "user-runtime-dir@0.service",
"source": "systemd",
"state": "stopped",
"status": "active"
},
"user@.service": {
"name": "user@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"user@0.service": {
"name": "user@0.service",
"source": "systemd",
"state": "running",
"status": "active"
}
}
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Create and update quadlets] ***********
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:114
Wednesday 10 July 2024 05:15:24 +0000 (0:00:02.186) 0:03:29.277 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_state != \"absent\"",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14
Wednesday 10 July 2024 05:15:24 +0000 (0:00:00.020) 0:03:29.297 ********
ok: [sut] => {
"ansible_facts": {
"__podman_quadlet_file_src": "quadlet-demo-mysql.volume",
"__podman_quadlet_spec": {},
"__podman_quadlet_str": "[Volume]",
"__podman_quadlet_template_src": ""
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25
Wednesday 10 July 2024 05:15:24 +0000 (0:00:00.073) 0:03:29.371 ********
ok: [sut] => {
"ansible_facts": {
"__podman_continue_if_pull_fails": false,
"__podman_pull_image": true,
"__podman_state": "absent",
"__podman_systemd_unit_scope": "",
"__podman_user": "root"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] *****
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35
Wednesday 10 July 2024 05:15:24 +0000 (0:00:00.028) 0:03:29.399 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_quadlet_file_src",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48
Wednesday 10 July 2024 05:15:24 +0000 (0:00:00.023) 0:03:29.422 ********
ok: [sut] => {
"ansible_facts": {
"__podman_quadlet_name": "quadlet-demo-mysql",
"__podman_quadlet_type": "volume",
"__podman_rootless": false
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Check user and group information] *****
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57
Wednesday 10 July 2024 05:15:24 +0000 (0:00:00.032) 0:03:29.455 ********
included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for sut
TASK [fedora.linux_system_roles.podman : Get user information] *****************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2
Wednesday 10 July 2024 05:15:24 +0000 (0:00:00.041) 0:03:29.497 ********
skipping: [sut] => {
"changed": false,
"false_condition": "'getent_passwd' not in ansible_facts or __podman_user not in ansible_facts['getent_passwd']",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Fail if user does not exist] **********
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9
Wednesday 10 July 2024 05:15:24 +0000 (0:00:00.024) 0:03:29.521 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not ansible_facts[\"getent_passwd\"][__podman_user]",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set group for podman user] ************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16
Wednesday 10 July 2024 05:15:24 +0000 (0:00:00.024) 0:03:29.546 ********
ok: [sut] => {
"ansible_facts": {
"__podman_group": "0"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Get group information] ****************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:28
Wednesday 10 July 2024 05:15:24 +0000 (0:00:00.031) 0:03:29.578 ********
ok: [sut] => {
"ansible_facts": {
"getent_group": {
"root": [
"x",
"0",
""
]
}
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set group name] ***********************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:35
Wednesday 10 July 2024 05:15:25 +0000 (0:00:00.239) 0:03:29.818 ********
ok: [sut] => {
"ansible_facts": {
"__podman_group_name": "root"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : See if getsubids exists] **************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:39
Wednesday 10 July 2024 05:15:25 +0000 (0:00:00.029) 0:03:29.847 ********
ok: [sut] => {
"changed": false,
"stat": {
"atime": 1720588427.5164855,
"attr_flags": "e",
"attributes": [
"extents"
],
"block_size": 4096,
"blocks": 32,
"charset": "binary",
"checksum": "72ba28a3be19c467e89ab90f5d21f99b605159a7",
"ctime": 1720588377.3007717,
"dev": 51714,
"device_type": 0,
"executable": true,
"exists": true,
"gid": 0,
"gr_name": "root",
"inode": 203603,
"isblk": false,
"ischr": false,
"isdir": false,
"isfifo": false,
"isgid": false,
"islnk": false,
"isreg": true,
"issock": false,
"isuid": false,
"mimetype": "application/x-pie-executable",
"mode": "0755",
"mtime": 1696291200.0,
"nlink": 1,
"path": "/usr/bin/getsubids",
"pw_name": "root",
"readable": true,
"rgrp": true,
"roth": true,
"rusr": true,
"size": 16032,
"uid": 0,
"version": "2092964490",
"wgrp": false,
"woth": false,
"writeable": true,
"wusr": true,
"xgrp": true,
"xoth": true,
"xusr": true
}
}
TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:50
Wednesday 10 July 2024 05:15:25 +0000 (0:00:00.235) 0:03:30.082 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_user not in [\"root\", \"0\"]",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Check group with getsubids] ***********
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:55
Wednesday 10 July 2024 05:15:25 +0000 (0:00:00.020) 0:03:30.103 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_user not in [\"root\", \"0\"]",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ******
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:60
Wednesday 10 July 2024 05:15:25 +0000 (0:00:00.020) 0:03:30.123 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_user not in [\"root\", \"0\"]",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Get subuid file] **********************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:74
Wednesday 10 July 2024 05:15:25 +0000 (0:00:00.019) 0:03:30.143 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_stat_getsubids.stat.exists",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Get subgid file] **********************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:79
Wednesday 10 July 2024 05:15:25 +0000 (0:00:00.058) 0:03:30.201 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_stat_getsubids.stat.exists",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ******
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:84
Wednesday 10 July 2024 05:15:25 +0000 (0:00:00.022) 0:03:30.224 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_stat_getsubids.stat.exists",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ******
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:94
Wednesday 10 July 2024 05:15:25 +0000 (0:00:00.020) 0:03:30.244 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_stat_getsubids.stat.exists",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] *****
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:101
Wednesday 10 July 2024 05:15:25 +0000 (0:00:00.020) 0:03:30.265 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_stat_getsubids.stat.exists",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62
Wednesday 10 July 2024 05:15:25 +0000 (0:00:00.020) 0:03:30.286 ********
ok: [sut] => {
"ansible_facts": {
"__podman_activate_systemd_unit": true,
"__podman_images_found": [],
"__podman_kube_yamls_raw": "",
"__podman_service_name": "quadlet-demo-mysql-volume.service",
"__podman_systemd_scope": "system",
"__podman_user_home_dir": "/root",
"__podman_xdg_runtime_dir": "/run/user/0"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73
Wednesday 10 July 2024 05:15:25 +0000 (0:00:00.038) 0:03:30.325 ********
ok: [sut] => {
"ansible_facts": {
"__podman_quadlet_path": "/etc/containers/systemd"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Get kube yaml contents] ***************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77
Wednesday 10 July 2024 05:15:25 +0000 (0:00:00.022) 0:03:30.348 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_state != \"absent\"",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87
Wednesday 10 July 2024 05:15:25 +0000 (0:00:00.020) 0:03:30.369 ********
ok: [sut] => {
"ansible_facts": {
"__podman_images": [],
"__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo-mysql.volume",
"__podman_volumes": []
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 6] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:103
Wednesday 10 July 2024 05:15:25 +0000 (0:00:00.049) 0:03:30.418 ********
ok: [sut] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : Cleanup quadlets] *********************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:110
Wednesday 10 July 2024 05:15:25 +0000 (0:00:00.026) 0:03:30.445 ********
included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml for sut
TASK [fedora.linux_system_roles.podman : Stat XDG_RUNTIME_DIR] *****************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:4
Wednesday 10 July 2024 05:15:25 +0000 (0:00:00.046) 0:03:30.492 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_rootless | bool",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Stop and disable service] *************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:12
Wednesday 10 July 2024 05:15:25 +0000 (0:00:00.020) 0:03:30.512 ********
changed: [sut] => {
"changed": true,
"enabled": false,
"failed_when_result": false,
"name": "quadlet-demo-mysql-volume.service",
"state": "stopped",
"status": {
"AccessSELinuxContext": "system_u:object_r:systemd_unit_file_t:s0",
"ActiveEnterTimestamp": "Wed 2024-07-10 05:14:02 UTC",
"ActiveEnterTimestampMonotonic": "962640855",
"ActiveExitTimestampMonotonic": "0",
"ActiveState": "active",
"After": "-.mount system.slice sysinit.target systemd-journald.socket basic.target",
"AllowIsolate": "no",
"AssertResult": "yes",
"AssertTimestamp": "Wed 2024-07-10 05:14:02 UTC",
"AssertTimestampMonotonic": "962587608",
"Before": "shutdown.target",
"BlockIOAccounting": "no",
"BlockIOWeight": "[not set]",
"CPUAccounting": "yes",
"CPUAffinityFromNUMA": "no",
"CPUQuotaPerSecUSec": "infinity",
"CPUQuotaPeriodUSec": "infinity",
"CPUSchedulingPolicy": "0",
"CPUSchedulingPriority": "0",
"CPUSchedulingResetOnFork": "no",
"CPUShares": "[not set]",
"CPUUsageNSec": "38624000",
"CPUWeight": "[not set]",
"CacheDirectoryMode": "0755",
"CanFreeze": "yes",
"CanIsolate": "no",
"CanReload": "no",
"CanStart": "yes",
"CanStop": "yes",
"CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore",
"CleanResult": "success",
"CollectMode": "inactive",
"ConditionResult": "yes",
"ConditionTimestamp": "Wed 2024-07-10 05:14:02 UTC",
"ConditionTimestampMonotonic": "962587604",
"ConfigurationDirectoryMode": "0755",
"Conflicts": "shutdown.target",
"ControlGroupId": "0",
"ControlPID": "0",
"CoredumpFilter": "0x33",
"DefaultDependencies": "yes",
"DefaultMemoryLow": "0",
"DefaultMemoryMin": "0",
"DefaultStartupMemoryLow": "0",
"Delegate": "no",
"Description": "quadlet-demo-mysql-volume.service",
"DevicePolicy": "auto",
"DropInPaths": "/usr/lib/systemd/system/service.d/10-timeout-abort.conf",
"DynamicUser": "no",
"ExecMainCode": "1",
"ExecMainExitTimestamp": "Wed 2024-07-10 05:14:02 UTC",
"ExecMainExitTimestampMonotonic": "962640687",
"ExecMainPID": "7517",
"ExecMainStartTimestamp": "Wed 2024-07-10 05:14:02 UTC",
"ExecMainStartTimestampMonotonic": "962595096",
"ExecMainStatus": "0",
"ExecStart": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman volume create --ignore systemd-quadlet-demo-mysql ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"ExecStartEx": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman volume create --ignore systemd-quadlet-demo-mysql ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"ExitType": "main",
"ExtensionImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent",
"FailureAction": "none",
"FileDescriptorStoreMax": "0",
"FileDescriptorStorePreserve": "restart",
"FinalKillSignal": "9",
"FragmentPath": "/run/systemd/generator/quadlet-demo-mysql-volume.service",
"FreezerState": "running",
"GID": "[not set]",
"GuessMainPID": "yes",
"IOAccounting": "no",
"IOReadBytes": "18446744073709551615",
"IOReadOperations": "18446744073709551615",
"IOSchedulingClass": "2",
"IOSchedulingPriority": "4",
"IOWeight": "[not set]",
"IOWriteBytes": "18446744073709551615",
"IOWriteOperations": "18446744073709551615",
"IPAccounting": "no",
"IPEgressBytes": "[no data]",
"IPEgressPackets": "[no data]",
"IPIngressBytes": "[no data]",
"IPIngressPackets": "[no data]",
"Id": "quadlet-demo-mysql-volume.service",
"IgnoreOnIsolate": "no",
"IgnoreSIGPIPE": "yes",
"InactiveEnterTimestampMonotonic": "0",
"InactiveExitTimestamp": "Wed 2024-07-10 05:14:02 UTC",
"InactiveExitTimestampMonotonic": "962595461",
"InvocationID": "816a65673bcb4a1b9babf8b8555173f9",
"JobRunningTimeoutUSec": "infinity",
"JobTimeoutAction": "none",
"JobTimeoutUSec": "infinity",
"KeyringMode": "private",
"KillMode": "control-group",
"KillSignal": "15",
"LimitAS": "infinity",
"LimitASSoft": "infinity",
"LimitCORE": "infinity",
"LimitCORESoft": "infinity",
"LimitCPU": "infinity",
"LimitCPUSoft": "infinity",
"LimitDATA": "infinity",
"LimitDATASoft": "infinity",
"LimitFSIZE": "infinity",
"LimitFSIZESoft": "infinity",
"LimitLOCKS": "infinity",
"LimitLOCKSSoft": "infinity",
"LimitMEMLOCK": "8388608",
"LimitMEMLOCKSoft": "8388608",
"LimitMSGQUEUE": "819200",
"LimitMSGQUEUESoft": "819200",
"LimitNICE": "0",
"LimitNICESoft": "0",
"LimitNOFILE": "524288",
"LimitNOFILESoft": "1024",
"LimitNPROC": "14756",
"LimitNPROCSoft": "14756",
"LimitRSS": "infinity",
"LimitRSSSoft": "infinity",
"LimitRTPRIO": "0",
"LimitRTPRIOSoft": "0",
"LimitRTTIME": "infinity",
"LimitRTTIMESoft": "infinity",
"LimitSIGPENDING": "14756",
"LimitSIGPENDINGSoft": "14756",
"LimitSTACK": "infinity",
"LimitSTACKSoft": "8388608",
"LoadState": "loaded",
"LockPersonality": "no",
"LogLevelMax": "-1",
"LogRateLimitBurst": "0",
"LogRateLimitIntervalUSec": "0",
"LogsDirectoryMode": "0755",
"MainPID": "0",
"ManagedOOMMemoryPressure": "auto",
"ManagedOOMMemoryPressureLimit": "0",
"ManagedOOMPreference": "none",
"ManagedOOMSwap": "auto",
"MemoryAccounting": "yes",
"MemoryAvailable": "infinity",
"MemoryCurrent": "[not set]",
"MemoryDenyWriteExecute": "no",
"MemoryHigh": "infinity",
"MemoryKSM": "no",
"MemoryLimit": "infinity",
"MemoryLow": "0",
"MemoryMax": "infinity",
"MemoryMin": "0",
"MemoryPressureThresholdUSec": "200ms",
"MemoryPressureWatch": "auto",
"MemorySwapMax": "infinity",
"MemoryZSwapMax": "infinity",
"MountAPIVFS": "no",
"MountImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent",
"NFileDescriptorStore": "0",
"NRestarts": "0",
"NUMAPolicy": "n/a",
"Names": "quadlet-demo-mysql-volume.service",
"NeedDaemonReload": "no",
"Nice": "0",
"NoNewPrivileges": "no",
"NonBlocking": "no",
"NotifyAccess": "none",
"OOMPolicy": "stop",
"OOMScoreAdjust": "0",
"OnFailureJobMode": "replace",
"OnSuccessJobMode": "fail",
"Perpetual": "no",
"PrivateDevices": "no",
"PrivateIPC": "no",
"PrivateMounts": "no",
"PrivateNetwork": "no",
"PrivateTmp": "no",
"PrivateUsers": "no",
"ProcSubset": "all",
"ProtectClock": "no",
"ProtectControlGroups": "no",
"ProtectHome": "no",
"ProtectHostname": "no",
"ProtectKernelLogs": "no",
"ProtectKernelModules": "no",
"ProtectKernelTunables": "no",
"ProtectProc": "default",
"ProtectSystem": "no",
"RefuseManualStart": "no",
"RefuseManualStop": "no",
"ReloadResult": "success",
"ReloadSignal": "1",
"RemainAfterExit": "yes",
"RemoveIPC": "no",
"Requires": "system.slice sysinit.target -.mount",
"RequiresMountsFor": "/run/containers",
"Restart": "no",
"RestartKillSignal": "15",
"RestartMaxDelayUSec": "infinity",
"RestartMode": "normal",
"RestartSteps": "0",
"RestartUSec": "100ms",
"RestartUSecNext": "100ms",
"RestrictNamespaces": "no",
"RestrictRealtime": "no",
"RestrictSUIDSGID": "no",
"Result": "success",
"RootDirectoryStartOnly": "no",
"RootEphemeral": "no",
"RootImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent",
"RuntimeDirectoryMode": "0755",
"RuntimeDirectoryPreserve": "no",
"RuntimeMaxUSec": "infinity",
"RuntimeRandomizedExtraUSec": "0",
"SameProcessGroup": "no",
"SecureBits": "0",
"SendSIGHUP": "no",
"SendSIGKILL": "yes",
"Slice": "system.slice",
"StandardError": "inherit",
"StandardInput": "null",
"StandardOutput": "journal",
"StartLimitAction": "none",
"StartLimitBurst": "5",
"StartLimitIntervalUSec": "10s",
"StartupBlockIOWeight": "[not set]",
"StartupCPUShares": "[not set]",
"StartupCPUWeight": "[not set]",
"StartupIOWeight": "[not set]",
"StartupMemoryHigh": "18446744073709551615",
"StartupMemoryLow": "0",
"StartupMemoryMax": "18446744073709551615",
"StartupMemorySwapMax": "18446744073709551615",
"StartupMemoryZSwapMax": "18446744073709551615",
"StateChangeTimestamp": "Wed 2024-07-10 05:14:02 UTC",
"StateChangeTimestampMonotonic": "962640855",
"StateDirectoryMode": "0755",
"StatusErrno": "0",
"StopWhenUnneeded": "no",
"SubState": "exited",
"SuccessAction": "none",
"SyslogFacility": "3",
"SyslogIdentifier": "quadlet-demo-mysql-volume",
"SyslogLevel": "6",
"SyslogLevelPrefix": "yes",
"SyslogPriority": "30",
"SystemCallErrorNumber": "2147483646",
"TTYReset": "no",
"TTYVHangup": "no",
"TTYVTDisallocate": "no",
"TasksAccounting": "yes",
"TasksCurrent": "[not set]",
"TasksMax": "4426",
"TimeoutAbortUSec": "45s",
"TimeoutCleanUSec": "infinity",
"TimeoutStartFailureMode": "terminate",
"TimeoutStartUSec": "infinity",
"TimeoutStopFailureMode": "abort",
"TimeoutStopUSec": "45s",
"TimerSlackNSec": "50000",
"Transient": "no",
"Type": "oneshot",
"UID": "[not set]",
"UMask": "0022",
"UnitFilePreset": "disabled",
"UnitFileState": "generated",
"UtmpMode": "init",
"WatchdogSignal": "6",
"WatchdogTimestampMonotonic": "0",
"WatchdogUSec": "0"
}
}
TASK [fedora.linux_system_roles.podman : Remove quadlet file] ******************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:33
Wednesday 10 July 2024 05:15:26 +0000 (0:00:00.843) 0:03:31.356 ********
changed: [sut] => {
"changed": true,
"path": "/etc/containers/systemd/quadlet-demo-mysql.volume",
"state": "absent"
}
TASK [fedora.linux_system_roles.podman : Manage linger] ************************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:39
Wednesday 10 July 2024 05:15:26 +0000 (0:00:00.279) 0:03:31.636 ********
included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for sut
TASK [fedora.linux_system_roles.podman : Enable linger if needed] **************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12
Wednesday 10 July 2024 05:15:27 +0000 (0:00:00.035) 0:03:31.671 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_rootless | bool",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18
Wednesday 10 July 2024 05:15:27 +0000 (0:00:00.021) 0:03:31.692 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_rootless | bool",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22
Wednesday 10 July 2024 05:15:27 +0000 (0:00:00.020) 0:03:31.713 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_rootless | bool",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Reload systemctl] *********************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:47
Wednesday 10 July 2024 05:15:27 +0000 (0:00:00.020) 0:03:31.734 ********
ok: [sut] => {
"changed": false,
"name": null,
"status": {}
}
TASK [fedora.linux_system_roles.podman : Prune images no longer in use] ********
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:56
Wednesday 10 July 2024 05:15:27 +0000 (0:00:00.786) 0:03:32.521 ********
changed: [sut] => {
"changed": true,
"cmd": [
"podman",
"image",
"prune",
"--all",
"-f"
],
"delta": "0:00:00.033872",
"end": "2024-07-10 05:15:28.088749",
"rc": 0,
"start": "2024-07-10 05:15:28.054877"
}
TASK [fedora.linux_system_roles.podman : Remove networks no longer in use] *****
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:64
Wednesday 10 July 2024 05:15:28 +0000 (0:00:00.272) 0:03:32.793 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_quadlet_type == \"network\"",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Remove volumes no longer in use] ******
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:73
Wednesday 10 July 2024 05:15:28 +0000 (0:00:00.023) 0:03:32.816 ********
changed: [sut] => {
"changed": true,
"cmd": [
"podman",
"volume",
"prune",
"-f"
],
"delta": "0:00:00.131415",
"end": "2024-07-10 05:15:28.481581",
"rc": 0,
"start": "2024-07-10 05:15:28.350166"
}
STDOUT:
systemd-quadlet-demo-mysql
wp-pv-claim
envoy-proxy-config
envoy-certificates
TASK [fedora.linux_system_roles.podman : For testing and debugging - images] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:82
Wednesday 10 July 2024 05:15:28 +0000 (0:00:00.365) 0:03:33.182 ********
ok: [sut] => {
"changed": false,
"cmd": [
"podman",
"images",
"-n"
],
"delta": "0:00:00.035502",
"end": "2024-07-10 05:15:28.754719",
"rc": 0,
"start": "2024-07-10 05:15:28.719217"
}
TASK [fedora.linux_system_roles.podman : For testing and debugging - volumes] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:92
Wednesday 10 July 2024 05:15:28 +0000 (0:00:00.274) 0:03:33.456 ********
ok: [sut] => {
"changed": false,
"cmd": [
"podman",
"volume",
"ls",
"-n"
],
"delta": "0:00:00.031639",
"end": "2024-07-10 05:15:29.022810",
"rc": 0,
"start": "2024-07-10 05:15:28.991171"
}
TASK [fedora.linux_system_roles.podman : For testing and debugging - containers] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:102
Wednesday 10 July 2024 05:15:29 +0000 (0:00:00.269) 0:03:33.726 ********
ok: [sut] => {
"changed": false,
"cmd": [
"podman",
"ps",
"--noheading"
],
"delta": "0:00:00.036405",
"end": "2024-07-10 05:15:29.297830",
"rc": 0,
"start": "2024-07-10 05:15:29.261425"
}
TASK [fedora.linux_system_roles.podman : For testing and debugging - networks] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:112
Wednesday 10 July 2024 05:15:29 +0000 (0:00:00.273) 0:03:33.999 ********
ok: [sut] => {
"changed": false,
"cmd": [
"podman",
"network",
"ls",
"-n",
"-q"
],
"delta": "0:00:00.037909",
"end": "2024-07-10 05:15:29.574206",
"rc": 0,
"start": "2024-07-10 05:15:29.536297"
}
STDOUT:
podman
systemd-quadlet-demo
TASK [fedora.linux_system_roles.podman : For testing and debugging - secrets] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:122
Wednesday 10 July 2024 05:15:29 +0000 (0:00:00.276) 0:03:34.276 ********
ok: [sut] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : For testing and debugging - services] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:133
Wednesday 10 July 2024 05:15:29 +0000 (0:00:00.271) 0:03:34.547 ********
ok: [sut] => {
"ansible_facts": {
"services": {
"NetworkManager-dispatcher.service": {
"name": "NetworkManager-dispatcher.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"NetworkManager-wait-online.service": {
"name": "NetworkManager-wait-online.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"NetworkManager.service": {
"name": "NetworkManager.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"arp-ethers.service": {
"name": "arp-ethers.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"auditd.service": {
"name": "auditd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"auth-rpcgss-module.service": {
"name": "auth-rpcgss-module.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"autovt@.service": {
"name": "autovt@.service",
"source": "systemd",
"state": "unknown",
"status": "alias"
},
"blk-availability.service": {
"name": "blk-availability.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"bluetooth.service": {
"name": "bluetooth.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"certmonger.service": {
"name": "certmonger.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"chrony-wait.service": {
"name": "chrony-wait.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"chronyd-restricted.service": {
"name": "chronyd-restricted.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"chronyd.service": {
"name": "chronyd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"cloud-config.service": {
"name": "cloud-config.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"cloud-final.service": {
"name": "cloud-final.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"cloud-init-hotplugd.service": {
"name": "cloud-init-hotplugd.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"cloud-init-local.service": {
"name": "cloud-init-local.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"cloud-init.service": {
"name": "cloud-init.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"console-getty.service": {
"name": "console-getty.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"container-getty@.service": {
"name": "container-getty@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"dbus-broker.service": {
"name": "dbus-broker.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"dbus-org.bluez.service": {
"name": "dbus-org.bluez.service",
"source": "systemd",
"state": "inactive",
"status": "alias"
},
"dbus-org.fedoraproject.FirewallD1.service": {
"name": "dbus-org.fedoraproject.FirewallD1.service",
"source": "systemd",
"state": "active",
"status": "alias"
},
"dbus-org.freedesktop.home1.service": {
"name": "dbus-org.freedesktop.home1.service",
"source": "systemd",
"state": "active",
"status": "alias"
},
"dbus-org.freedesktop.hostname1.service": {
"name": "dbus-org.freedesktop.hostname1.service",
"source": "systemd",
"state": "inactive",
"status": "alias"
},
"dbus-org.freedesktop.locale1.service": {
"name": "dbus-org.freedesktop.locale1.service",
"source": "systemd",
"state": "inactive",
"status": "alias"
},
"dbus-org.freedesktop.login1.service": {
"name": "dbus-org.freedesktop.login1.service",
"source": "systemd",
"state": "active",
"status": "alias"
},
"dbus-org.freedesktop.nm-dispatcher.service": {
"name": "dbus-org.freedesktop.nm-dispatcher.service",
"source": "systemd",
"state": "inactive",
"status": "alias"
},
"dbus-org.freedesktop.oom1.service": {
"name": "dbus-org.freedesktop.oom1.service",
"source": "systemd",
"state": "active",
"status": "alias"
},
"dbus-org.freedesktop.portable1.service": {
"name": "dbus-org.freedesktop.portable1.service",
"source": "systemd",
"state": "inactive",
"status": "alias"
},
"dbus-org.freedesktop.resolve1.service": {
"name": "dbus-org.freedesktop.resolve1.service",
"source": "systemd",
"state": "active",
"status": "alias"
},
"dbus-org.freedesktop.timedate1.service": {
"name": "dbus-org.freedesktop.timedate1.service",
"source": "systemd",
"state": "inactive",
"status": "alias"
},
"dbus.service": {
"name": "dbus.service",
"source": "systemd",
"state": "active",
"status": "alias"
},
"debug-shell.service": {
"name": "debug-shell.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"display-manager.service": {
"name": "display-manager.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"dm-event.service": {
"name": "dm-event.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dnf-makecache.service": {
"name": "dnf-makecache.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dnf-system-upgrade-cleanup.service": {
"name": "dnf-system-upgrade-cleanup.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dnf-system-upgrade.service": {
"name": "dnf-system-upgrade.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"dracut-cmdline.service": {
"name": "dracut-cmdline.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-initqueue.service": {
"name": "dracut-initqueue.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-mount.service": {
"name": "dracut-mount.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-mount.service": {
"name": "dracut-pre-mount.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-pivot.service": {
"name": "dracut-pre-pivot.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-trigger.service": {
"name": "dracut-pre-trigger.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-udev.service": {
"name": "dracut-pre-udev.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-shutdown-onfailure.service": {
"name": "dracut-shutdown-onfailure.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-shutdown.service": {
"name": "dracut-shutdown.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"ebtables.service": {
"name": "ebtables.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"emergency.service": {
"name": "emergency.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"firewalld.service": {
"name": "firewalld.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"fsidd.service": {
"name": "fsidd.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"fstrim.service": {
"name": "fstrim.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"fwupd-offline-update.service": {
"name": "fwupd-offline-update.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"fwupd-refresh.service": {
"name": "fwupd-refresh.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"fwupd.service": {
"name": "fwupd.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"getty@.service": {
"name": "getty@.service",
"source": "systemd",
"state": "unknown",
"status": "enabled"
},
"getty@tty1.service": {
"name": "getty@tty1.service",
"source": "systemd",
"state": "running",
"status": "active"
},
"grub-boot-indeterminate.service": {
"name": "grub-boot-indeterminate.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"grub2-systemd-integration.service": {
"name": "grub2-systemd-integration.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"gssproxy.service": {
"name": "gssproxy.service",
"source": "systemd",
"state": "running",
"status": "disabled"
},
"hv_kvp_daemon.service": {
"name": "hv_kvp_daemon.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"import-state.service": {
"name": "import-state.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"initrd-cleanup.service": {
"name": "initrd-cleanup.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"initrd-parse-etc.service": {
"name": "initrd-parse-etc.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"initrd-switch-root.service": {
"name": "initrd-switch-root.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"initrd-udevadm-cleanup-db.service": {
"name": "initrd-udevadm-cleanup-db.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"ip6tables.service": {
"name": "ip6tables.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"ipset.service": {
"name": "ipset.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"iptables.service": {
"name": "iptables.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"kmod-static-nodes.service": {
"name": "kmod-static-nodes.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"ldconfig.service": {
"name": "ldconfig.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"loadmodules.service": {
"name": "loadmodules.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"lvm2-lvmpolld.service": {
"name": "lvm2-lvmpolld.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"lvm2-monitor.service": {
"name": "lvm2-monitor.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"man-db-cache-update.service": {
"name": "man-db-cache-update.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"man-db-restart-cache-update.service": {
"name": "man-db-restart-cache-update.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"mdadm-grow-continue@.service": {
"name": "mdadm-grow-continue@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"mdadm-last-resort@.service": {
"name": "mdadm-last-resort@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"mdcheck_continue.service": {
"name": "mdcheck_continue.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"mdcheck_start.service": {
"name": "mdcheck_start.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"mdmon@.service": {
"name": "mdmon@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"mdmonitor-oneshot.service": {
"name": "mdmonitor-oneshot.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"mdmonitor.service": {
"name": "mdmonitor.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"modprobe@.service": {
"name": "modprobe@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"modprobe@configfs.service": {
"name": "modprobe@configfs.service",
"source": "systemd",
"state": "stopped",
"status": "inactive"
},
"modprobe@dm_mod.service": {
"name": "modprobe@dm_mod.service",
"source": "systemd",
"state": "stopped",
"status": "inactive"
},
"modprobe@drm.service": {
"name": "modprobe@drm.service",
"source": "systemd",
"state": "stopped",
"status": "inactive"
},
"modprobe@efi_pstore.service": {
"name": "modprobe@efi_pstore.service",
"source": "systemd",
"state": "stopped",
"status": "inactive"
},
"modprobe@fuse.service": {
"name": "modprobe@fuse.service",
"source": "systemd",
"state": "stopped",
"status": "inactive"
},
"modprobe@loop.service": {
"name": "modprobe@loop.service",
"source": "systemd",
"state": "stopped",
"status": "inactive"
},
"netavark-dhcp-proxy.service": {
"name": "netavark-dhcp-proxy.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"netavark-firewalld-reload.service": {
"name": "netavark-firewalld-reload.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"network.service": {
"name": "network.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"nfs-blkmap.service": {
"name": "nfs-blkmap.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"nfs-idmapd.service": {
"name": "nfs-idmapd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nfs-mountd.service": {
"name": "nfs-mountd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nfs-server.service": {
"name": "nfs-server.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"nfs-utils.service": {
"name": "nfs-utils.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nfsdcld.service": {
"name": "nfsdcld.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nftables.service": {
"name": "nftables.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"nis-domainname.service": {
"name": "nis-domainname.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"nm-priv-helper.service": {
"name": "nm-priv-helper.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"ntpd.service": {
"name": "ntpd.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"ntpdate.service": {
"name": "ntpdate.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"pam_namespace.service": {
"name": "pam_namespace.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"pcscd.service": {
"name": "pcscd.service",
"source": "systemd",
"state": "stopped",
"status": "indirect"
},
"plymouth-halt.service": {
"name": "plymouth-halt.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-kexec.service": {
"name": "plymouth-kexec.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-poweroff.service": {
"name": "plymouth-poweroff.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-quit-wait.service": {
"name": "plymouth-quit-wait.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-quit.service": {
"name": "plymouth-quit.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-read-write.service": {
"name": "plymouth-read-write.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-reboot.service": {
"name": "plymouth-reboot.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-start.service": {
"name": "plymouth-start.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-switch-root-initramfs.service": {
"name": "plymouth-switch-root-initramfs.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-switch-root.service": {
"name": "plymouth-switch-root.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"podman-auto-update.service": {
"name": "podman-auto-update.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"podman-clean-transient.service": {
"name": "podman-clean-transient.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"podman-kube@.service": {
"name": "podman-kube@.service",
"source": "systemd",
"state": "unknown",
"status": "disabled"
},
"podman-restart.service": {
"name": "podman-restart.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"podman.service": {
"name": "podman.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"polkit.service": {
"name": "polkit.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"quadlet-demo-mysql.service": {
"name": "quadlet-demo-mysql.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"quadlet-demo-network.service": {
"name": "quadlet-demo-network.service",
"source": "systemd",
"state": "stopped",
"status": "generated"
},
"quotaon.service": {
"name": "quotaon.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"raid-check.service": {
"name": "raid-check.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rc-local.service": {
"name": "rc-local.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rescue.service": {
"name": "rescue.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"restraintd.service": {
"name": "restraintd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"rngd.service": {
"name": "rngd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"rpc-gssd.service": {
"name": "rpc-gssd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rpc-statd-notify.service": {
"name": "rpc-statd-notify.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rpc-statd.service": {
"name": "rpc-statd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rpc-svcgssd.service": {
"name": "rpc-svcgssd.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"rpcbind.service": {
"name": "rpcbind.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"rpmdb-migrate.service": {
"name": "rpmdb-migrate.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"rpmdb-rebuild.service": {
"name": "rpmdb-rebuild.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"selinux-autorelabel-mark.service": {
"name": "selinux-autorelabel-mark.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"selinux-autorelabel.service": {
"name": "selinux-autorelabel.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"selinux-check-proper-disable.service": {
"name": "selinux-check-proper-disable.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"serial-getty@.service": {
"name": "serial-getty@.service",
"source": "systemd",
"state": "unknown",
"status": "indirect"
},
"serial-getty@ttyS0.service": {
"name": "serial-getty@ttyS0.service",
"source": "systemd",
"state": "running",
"status": "active"
},
"snapd.seeded.service": {
"name": "snapd.seeded.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"sntp.service": {
"name": "sntp.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"ssh-host-keys-migration.service": {
"name": "ssh-host-keys-migration.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"sshd-keygen.service": {
"name": "sshd-keygen.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"sshd-keygen@.service": {
"name": "sshd-keygen@.service",
"source": "systemd",
"state": "unknown",
"status": "disabled"
},
"sshd-keygen@ecdsa.service": {
"name": "sshd-keygen@ecdsa.service",
"source": "systemd",
"state": "stopped",
"status": "inactive"
},
"sshd-keygen@ed25519.service": {
"name": "sshd-keygen@ed25519.service",
"source": "systemd",
"state": "stopped",
"status": "inactive"
},
"sshd-keygen@rsa.service": {
"name": "sshd-keygen@rsa.service",
"source": "systemd",
"state": "stopped",
"status": "inactive"
},
"sshd.service": {
"name": "sshd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"sshd@.service": {
"name": "sshd@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"sssd-autofs.service": {
"name": "sssd-autofs.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-kcm.service": {
"name": "sssd-kcm.service",
"source": "systemd",
"state": "stopped",
"status": "indirect"
},
"sssd-nss.service": {
"name": "sssd-nss.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-pac.service": {
"name": "sssd-pac.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-pam.service": {
"name": "sssd-pam.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-ssh.service": {
"name": "sssd-ssh.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-sudo.service": {
"name": "sssd-sudo.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd.service": {
"name": "sssd.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"syslog.service": {
"name": "syslog.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"system-update-cleanup.service": {
"name": "system-update-cleanup.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-ask-password-console.service": {
"name": "systemd-ask-password-console.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-ask-password-plymouth.service": {
"name": "systemd-ask-password-plymouth.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-ask-password-wall.service": {
"name": "systemd-ask-password-wall.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-backlight@.service": {
"name": "systemd-backlight@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-battery-check.service": {
"name": "systemd-battery-check.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-binfmt.service": {
"name": "systemd-binfmt.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-bless-boot.service": {
"name": "systemd-bless-boot.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-boot-check-no-failures.service": {
"name": "systemd-boot-check-no-failures.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"systemd-boot-random-seed.service": {
"name": "systemd-boot-random-seed.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-boot-update.service": {
"name": "systemd-boot-update.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"systemd-confext.service": {
"name": "systemd-confext.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"systemd-coredump@.service": {
"name": "systemd-coredump@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-exit.service": {
"name": "systemd-exit.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-firstboot.service": {
"name": "systemd-firstboot.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-fsck-root.service": {
"name": "systemd-fsck-root.service",
"source": "systemd",
"state": "stopped",
"status": "enabled-runtime"
},
"systemd-fsck@.service": {
"name": "systemd-fsck@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-growfs-root.service": {
"name": "systemd-growfs-root.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-growfs@.service": {
"name": "systemd-growfs@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-halt.service": {
"name": "systemd-halt.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-hibernate.service": {
"name": "systemd-hibernate.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-homed-activate.service": {
"name": "systemd-homed-activate.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"systemd-homed.service": {
"name": "systemd-homed.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"systemd-hostnamed.service": {
"name": "systemd-hostnamed.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-hwdb-update.service": {
"name": "systemd-hwdb-update.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-hybrid-sleep.service": {
"name": "systemd-hybrid-sleep.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-initctl.service": {
"name": "systemd-initctl.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-journal-catalog-update.service": {
"name": "systemd-journal-catalog-update.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-journal-flush.service": {
"name": "systemd-journal-flush.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-journald.service": {
"name": "systemd-journald.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"systemd-journald@.service": {
"name": "systemd-journald@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-kexec.service": {
"name": "systemd-kexec.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-localed.service": {
"name": "systemd-localed.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-logind.service": {
"name": "systemd-logind.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"systemd-machine-id-commit.service": {
"name": "systemd-machine-id-commit.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-modules-load.service": {
"name": "systemd-modules-load.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-network-generator.service": {
"name": "systemd-network-generator.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"systemd-networkd-wait-online.service": {
"name": "systemd-networkd-wait-online.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"systemd-networkd-wait-online@.service": {
"name": "systemd-networkd-wait-online@.service",
"source": "systemd",
"state": "unknown",
"status": "disabled"
},
"systemd-networkd.service": {
"name": "systemd-networkd.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"systemd-oomd.service": {
"name": "systemd-oomd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"systemd-pcrfs-root.service": {
"name": "systemd-pcrfs-root.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-pcrfs@.service": {
"name": "systemd-pcrfs@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-pcrmachine.service": {
"name": "systemd-pcrmachine.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-pcrphase-initrd.service": {
"name": "systemd-pcrphase-initrd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-pcrphase-sysinit.service": {
"name": "systemd-pcrphase-sysinit.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-pcrphase.service": {
"name": "systemd-pcrphase.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-portabled.service": {
"name": "systemd-portabled.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-poweroff.service": {
"name": "systemd-poweroff.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-pstore.service": {
"name": "systemd-pstore.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"systemd-quotacheck.service": {
"name": "systemd-quotacheck.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-random-seed.service": {
"name": "systemd-random-seed.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-reboot.service": {
"name": "systemd-reboot.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-remount-fs.service": {
"name": "systemd-remount-fs.service",
"source": "systemd",
"state": "stopped",
"status": "enabled-runtime"
},
"systemd-repart.service": {
"name": "systemd-repart.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-resolved.service": {
"name": "systemd-resolved.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"systemd-rfkill.service": {
"name": "systemd-rfkill.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-soft-reboot.service": {
"name": "systemd-soft-reboot.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-suspend-then-hibernate.service": {
"name": "systemd-suspend-then-hibernate.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-suspend.service": {
"name": "systemd-suspend.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-sysctl.service": {
"name": "systemd-sysctl.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-sysext.service": {
"name": "systemd-sysext.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"systemd-sysupdate-reboot.service": {
"name": "systemd-sysupdate-reboot.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"systemd-sysupdate.service": {
"name": "systemd-sysupdate.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"systemd-sysusers.service": {
"name": "systemd-sysusers.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-time-wait-sync.service": {
"name": "systemd-time-wait-sync.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"systemd-timedated.service": {
"name": "systemd-timedated.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-timesyncd.service": {
"name": "systemd-timesyncd.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"systemd-tmpfiles-clean.service": {
"name": "systemd-tmpfiles-clean.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-tmpfiles-setup-dev-early.service": {
"name": "systemd-tmpfiles-setup-dev-early.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-tmpfiles-setup-dev.service": {
"name": "systemd-tmpfiles-setup-dev.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-tmpfiles-setup.service": {
"name": "systemd-tmpfiles-setup.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-udev-settle.service": {
"name": "systemd-udev-settle.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-udev-trigger.service": {
"name": "systemd-udev-trigger.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-udevd.service": {
"name": "systemd-udevd.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"systemd-update-done.service": {
"name": "systemd-update-done.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-update-utmp-runlevel.service": {
"name": "systemd-update-utmp-runlevel.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-update-utmp.service": {
"name": "systemd-update-utmp.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-user-sessions.service": {
"name": "systemd-user-sessions.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-userdbd.service": {
"name": "systemd-userdbd.service",
"source": "systemd",
"state": "running",
"status": "indirect"
},
"systemd-vconsole-setup.service": {
"name": "systemd-vconsole-setup.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-volatile-root.service": {
"name": "systemd-volatile-root.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-zram-setup@.service": {
"name": "systemd-zram-setup@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-zram-setup@zram0.service": {
"name": "systemd-zram-setup@zram0.service",
"source": "systemd",
"state": "stopped",
"status": "active"
},
"udisks2.service": {
"name": "udisks2.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"unbound-anchor.service": {
"name": "unbound-anchor.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"user-runtime-dir@.service": {
"name": "user-runtime-dir@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"user-runtime-dir@0.service": {
"name": "user-runtime-dir@0.service",
"source": "systemd",
"state": "stopped",
"status": "active"
},
"user@.service": {
"name": "user@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"user@0.service": {
"name": "user@0.service",
"source": "systemd",
"state": "running",
"status": "active"
}
}
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Create and update quadlets] ***********
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:114
Wednesday 10 July 2024 05:15:32 +0000 (0:00:02.298) 0:03:36.845 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_state != \"absent\"",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14
Wednesday 10 July 2024 05:15:32 +0000 (0:00:00.022) 0:03:36.868 ********
ok: [sut] => {
"ansible_facts": {
"__podman_quadlet_file_src": "quadlet-demo.network",
"__podman_quadlet_spec": {},
"__podman_quadlet_str": "[Network]\nSubnet=192.168.30.0/24\nGateway=192.168.30.1\nLabel=app=wordpress",
"__podman_quadlet_template_src": ""
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25
Wednesday 10 July 2024 05:15:32 +0000 (0:00:00.031) 0:03:36.899 ********
ok: [sut] => {
"ansible_facts": {
"__podman_continue_if_pull_fails": false,
"__podman_pull_image": true,
"__podman_state": "absent",
"__podman_systemd_unit_scope": "",
"__podman_user": "root"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] *****
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35
Wednesday 10 July 2024 05:15:32 +0000 (0:00:00.027) 0:03:36.926 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_quadlet_file_src",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48
Wednesday 10 July 2024 05:15:32 +0000 (0:00:00.021) 0:03:36.948 ********
ok: [sut] => {
"ansible_facts": {
"__podman_quadlet_name": "quadlet-demo",
"__podman_quadlet_type": "network",
"__podman_rootless": false
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Check user and group information] *****
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57
Wednesday 10 July 2024 05:15:32 +0000 (0:00:00.032) 0:03:36.981 ********
included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for sut
TASK [fedora.linux_system_roles.podman : Get user information] *****************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2
Wednesday 10 July 2024 05:15:32 +0000 (0:00:00.041) 0:03:37.022 ********
skipping: [sut] => {
"changed": false,
"false_condition": "'getent_passwd' not in ansible_facts or __podman_user not in ansible_facts['getent_passwd']",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Fail if user does not exist] **********
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9
Wednesday 10 July 2024 05:15:32 +0000 (0:00:00.025) 0:03:37.048 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not ansible_facts[\"getent_passwd\"][__podman_user]",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set group for podman user] ************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16
Wednesday 10 July 2024 05:15:32 +0000 (0:00:00.026) 0:03:37.074 ********
ok: [sut] => {
"ansible_facts": {
"__podman_group": "0"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Get group information] ****************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:28
Wednesday 10 July 2024 05:15:32 +0000 (0:00:00.031) 0:03:37.105 ********
ok: [sut] => {
"ansible_facts": {
"getent_group": {
"root": [
"x",
"0",
""
]
}
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set group name] ***********************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:35
Wednesday 10 July 2024 05:15:32 +0000 (0:00:00.237) 0:03:37.343 ********
ok: [sut] => {
"ansible_facts": {
"__podman_group_name": "root"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : See if getsubids exists] **************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:39
Wednesday 10 July 2024 05:15:32 +0000 (0:00:00.029) 0:03:37.372 ********
ok: [sut] => {
"changed": false,
"stat": {
"atime": 1720588427.5164855,
"attr_flags": "e",
"attributes": [
"extents"
],
"block_size": 4096,
"blocks": 32,
"charset": "binary",
"checksum": "72ba28a3be19c467e89ab90f5d21f99b605159a7",
"ctime": 1720588377.3007717,
"dev": 51714,
"device_type": 0,
"executable": true,
"exists": true,
"gid": 0,
"gr_name": "root",
"inode": 203603,
"isblk": false,
"ischr": false,
"isdir": false,
"isfifo": false,
"isgid": false,
"islnk": false,
"isreg": true,
"issock": false,
"isuid": false,
"mimetype": "application/x-pie-executable",
"mode": "0755",
"mtime": 1696291200.0,
"nlink": 1,
"path": "/usr/bin/getsubids",
"pw_name": "root",
"readable": true,
"rgrp": true,
"roth": true,
"rusr": true,
"size": 16032,
"uid": 0,
"version": "2092964490",
"wgrp": false,
"woth": false,
"writeable": true,
"wusr": true,
"xgrp": true,
"xoth": true,
"xusr": true
}
}
TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:50
Wednesday 10 July 2024 05:15:32 +0000 (0:00:00.238) 0:03:37.611 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_user not in [\"root\", \"0\"]",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Check group with getsubids] ***********
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:55
Wednesday 10 July 2024 05:15:33 +0000 (0:00:00.061) 0:03:37.673 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_user not in [\"root\", \"0\"]",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ******
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:60
Wednesday 10 July 2024 05:15:33 +0000 (0:00:00.022) 0:03:37.695 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_user not in [\"root\", \"0\"]",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Get subuid file] **********************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:74
Wednesday 10 July 2024 05:15:33 +0000 (0:00:00.020) 0:03:37.716 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_stat_getsubids.stat.exists",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Get subgid file] **********************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:79
Wednesday 10 July 2024 05:15:33 +0000 (0:00:00.020) 0:03:37.737 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_stat_getsubids.stat.exists",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ******
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:84
Wednesday 10 July 2024 05:15:33 +0000 (0:00:00.021) 0:03:37.758 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_stat_getsubids.stat.exists",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ******
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:94
Wednesday 10 July 2024 05:15:33 +0000 (0:00:00.020) 0:03:37.779 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_stat_getsubids.stat.exists",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] *****
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:101
Wednesday 10 July 2024 05:15:33 +0000 (0:00:00.019) 0:03:37.799 ********
skipping: [sut] => {
"changed": false,
"false_condition": "not __podman_stat_getsubids.stat.exists",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62
Wednesday 10 July 2024 05:15:33 +0000 (0:00:00.020) 0:03:37.819 ********
ok: [sut] => {
"ansible_facts": {
"__podman_activate_systemd_unit": true,
"__podman_images_found": [],
"__podman_kube_yamls_raw": "",
"__podman_service_name": "quadlet-demo-network.service",
"__podman_systemd_scope": "system",
"__podman_user_home_dir": "/root",
"__podman_xdg_runtime_dir": "/run/user/0"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73
Wednesday 10 July 2024 05:15:33 +0000 (0:00:00.039) 0:03:37.858 ********
ok: [sut] => {
"ansible_facts": {
"__podman_quadlet_path": "/etc/containers/systemd"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Get kube yaml contents] ***************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77
Wednesday 10 July 2024 05:15:33 +0000 (0:00:00.023) 0:03:37.882 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_state != \"absent\"",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87
Wednesday 10 July 2024 05:15:33 +0000 (0:00:00.019) 0:03:37.901 ********
ok: [sut] => {
"ansible_facts": {
"__podman_images": [],
"__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo.network",
"__podman_volumes": []
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 6] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:103
Wednesday 10 July 2024 05:15:33 +0000 (0:00:00.048) 0:03:37.950 ********
ok: [sut] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : Cleanup quadlets] *********************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:110
Wednesday 10 July 2024 05:15:33 +0000 (0:00:00.025) 0:03:37.976 ********
included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml for sut
TASK [fedora.linux_system_roles.podman : Stat XDG_RUNTIME_DIR] *****************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:4
Wednesday 10 July 2024 05:15:33 +0000 (0:00:00.047) 0:03:38.023 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_rootless | bool",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Stop and disable service] *************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:12
Wednesday 10 July 2024 05:15:33 +0000 (0:00:00.061) 0:03:38.085 ********
changed: [sut] => {
"changed": true,
"enabled": false,
"failed_when_result": false,
"name": "quadlet-demo-network.service",
"state": "stopped",
"status": {
"AccessSELinuxContext": "system_u:object_r:systemd_unit_file_t:s0",
"ActiveEnterTimestamp": "Wed 2024-07-10 05:13:58 UTC",
"ActiveEnterTimestampMonotonic": "959271314",
"ActiveExitTimestampMonotonic": "0",
"ActiveState": "active",
"After": "basic.target -.mount systemd-journald.socket system.slice sysinit.target",
"AllowIsolate": "no",
"AssertResult": "yes",
"AssertTimestamp": "Wed 2024-07-10 05:13:58 UTC",
"AssertTimestampMonotonic": "959223551",
"Before": "shutdown.target",
"BlockIOAccounting": "no",
"BlockIOWeight": "[not set]",
"CPUAccounting": "yes",
"CPUAffinityFromNUMA": "no",
"CPUQuotaPerSecUSec": "infinity",
"CPUQuotaPeriodUSec": "infinity",
"CPUSchedulingPolicy": "0",
"CPUSchedulingPriority": "0",
"CPUSchedulingResetOnFork": "no",
"CPUShares": "[not set]",
"CPUUsageNSec": "37260000",
"CPUWeight": "[not set]",
"CacheDirectoryMode": "0755",
"CanFreeze": "yes",
"CanIsolate": "no",
"CanReload": "no",
"CanStart": "yes",
"CanStop": "yes",
"CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore",
"CleanResult": "success",
"CollectMode": "inactive",
"ConditionResult": "yes",
"ConditionTimestamp": "Wed 2024-07-10 05:13:58 UTC",
"ConditionTimestampMonotonic": "959223548",
"ConfigurationDirectoryMode": "0755",
"Conflicts": "shutdown.target",
"ControlGroupId": "0",
"ControlPID": "0",
"CoredumpFilter": "0x33",
"DefaultDependencies": "yes",
"DefaultMemoryLow": "0",
"DefaultMemoryMin": "0",
"DefaultStartupMemoryLow": "0",
"Delegate": "no",
"Description": "quadlet-demo-network.service",
"DevicePolicy": "auto",
"DropInPaths": "/usr/lib/systemd/system/service.d/10-timeout-abort.conf",
"DynamicUser": "no",
"ExecMainCode": "1",
"ExecMainExitTimestamp": "Wed 2024-07-10 05:13:58 UTC",
"ExecMainExitTimestampMonotonic": "959271098",
"ExecMainPID": "7247",
"ExecMainStartTimestamp": "Wed 2024-07-10 05:13:58 UTC",
"ExecMainStartTimestampMonotonic": "959231102",
"ExecMainStatus": "0",
"ExecStart": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman network create --ignore --subnet=192.168.30.0/24 --gateway=192.168.30.1 --label app=wordpress systemd-quadlet-demo ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"ExecStartEx": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman network create --ignore --subnet=192.168.30.0/24 --gateway=192.168.30.1 --label app=wordpress systemd-quadlet-demo ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"ExitType": "main",
"ExtensionImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent",
"FailureAction": "none",
"FileDescriptorStoreMax": "0",
"FileDescriptorStorePreserve": "restart",
"FinalKillSignal": "9",
"FragmentPath": "/run/systemd/generator/quadlet-demo-network.service",
"FreezerState": "running",
"GID": "[not set]",
"GuessMainPID": "yes",
"IOAccounting": "no",
"IOReadBytes": "18446744073709551615",
"IOReadOperations": "18446744073709551615",
"IOSchedulingClass": "2",
"IOSchedulingPriority": "4",
"IOWeight": "[not set]",
"IOWriteBytes": "18446744073709551615",
"IOWriteOperations": "18446744073709551615",
"IPAccounting": "no",
"IPEgressBytes": "[no data]",
"IPEgressPackets": "[no data]",
"IPIngressBytes": "[no data]",
"IPIngressPackets": "[no data]",
"Id": "quadlet-demo-network.service",
"IgnoreOnIsolate": "no",
"IgnoreSIGPIPE": "yes",
"InactiveEnterTimestampMonotonic": "0",
"InactiveExitTimestamp": "Wed 2024-07-10 05:13:58 UTC",
"InactiveExitTimestampMonotonic": "959231435",
"InvocationID": "a85c69adee0c4357a9d61aadc1100ae1",
"JobRunningTimeoutUSec": "infinity",
"JobTimeoutAction": "none",
"JobTimeoutUSec": "infinity",
"KeyringMode": "private",
"KillMode": "control-group",
"KillSignal": "15",
"LimitAS": "infinity",
"LimitASSoft": "infinity",
"LimitCORE": "infinity",
"LimitCORESoft": "infinity",
"LimitCPU": "infinity",
"LimitCPUSoft": "infinity",
"LimitDATA": "infinity",
"LimitDATASoft": "infinity",
"LimitFSIZE": "infinity",
"LimitFSIZESoft": "infinity",
"LimitLOCKS": "infinity",
"LimitLOCKSSoft": "infinity",
"LimitMEMLOCK": "8388608",
"LimitMEMLOCKSoft": "8388608",
"LimitMSGQUEUE": "819200",
"LimitMSGQUEUESoft": "819200",
"LimitNICE": "0",
"LimitNICESoft": "0",
"LimitNOFILE": "524288",
"LimitNOFILESoft": "1024",
"LimitNPROC": "14756",
"LimitNPROCSoft": "14756",
"LimitRSS": "infinity",
"LimitRSSSoft": "infinity",
"LimitRTPRIO": "0",
"LimitRTPRIOSoft": "0",
"LimitRTTIME": "infinity",
"LimitRTTIMESoft": "infinity",
"LimitSIGPENDING": "14756",
"LimitSIGPENDINGSoft": "14756",
"LimitSTACK": "infinity",
"LimitSTACKSoft": "8388608",
"LoadState": "loaded",
"LockPersonality": "no",
"LogLevelMax": "-1",
"LogRateLimitBurst": "0",
"LogRateLimitIntervalUSec": "0",
"LogsDirectoryMode": "0755",
"MainPID": "0",
"ManagedOOMMemoryPressure": "auto",
"ManagedOOMMemoryPressureLimit": "0",
"ManagedOOMPreference": "none",
"ManagedOOMSwap": "auto",
"MemoryAccounting": "yes",
"MemoryAvailable": "infinity",
"MemoryCurrent": "[not set]",
"MemoryDenyWriteExecute": "no",
"MemoryHigh": "infinity",
"MemoryKSM": "no",
"MemoryLimit": "infinity",
"MemoryLow": "0",
"MemoryMax": "infinity",
"MemoryMin": "0",
"MemoryPressureThresholdUSec": "200ms",
"MemoryPressureWatch": "auto",
"MemorySwapMax": "infinity",
"MemoryZSwapMax": "infinity",
"MountAPIVFS": "no",
"MountImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent",
"NFileDescriptorStore": "0",
"NRestarts": "0",
"NUMAPolicy": "n/a",
"Names": "quadlet-demo-network.service",
"NeedDaemonReload": "no",
"Nice": "0",
"NoNewPrivileges": "no",
"NonBlocking": "no",
"NotifyAccess": "none",
"OOMPolicy": "stop",
"OOMScoreAdjust": "0",
"OnFailureJobMode": "replace",
"OnSuccessJobMode": "fail",
"Perpetual": "no",
"PrivateDevices": "no",
"PrivateIPC": "no",
"PrivateMounts": "no",
"PrivateNetwork": "no",
"PrivateTmp": "no",
"PrivateUsers": "no",
"ProcSubset": "all",
"ProtectClock": "no",
"ProtectControlGroups": "no",
"ProtectHome": "no",
"ProtectHostname": "no",
"ProtectKernelLogs": "no",
"ProtectKernelModules": "no",
"ProtectKernelTunables": "no",
"ProtectProc": "default",
"ProtectSystem": "no",
"RefuseManualStart": "no",
"RefuseManualStop": "no",
"ReloadResult": "success",
"ReloadSignal": "1",
"RemainAfterExit": "yes",
"RemoveIPC": "no",
"Requires": "-.mount system.slice sysinit.target",
"RequiresMountsFor": "/run/containers",
"Restart": "no",
"RestartKillSignal": "15",
"RestartMaxDelayUSec": "infinity",
"RestartMode": "normal",
"RestartSteps": "0",
"RestartUSec": "100ms",
"RestartUSecNext": "100ms",
"RestrictNamespaces": "no",
"RestrictRealtime": "no",
"RestrictSUIDSGID": "no",
"Result": "success",
"RootDirectoryStartOnly": "no",
"RootEphemeral": "no",
"RootImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent",
"RuntimeDirectoryMode": "0755",
"RuntimeDirectoryPreserve": "no",
"RuntimeMaxUSec": "infinity",
"RuntimeRandomizedExtraUSec": "0",
"SameProcessGroup": "no",
"SecureBits": "0",
"SendSIGHUP": "no",
"SendSIGKILL": "yes",
"Slice": "system.slice",
"StandardError": "inherit",
"StandardInput": "null",
"StandardOutput": "journal",
"StartLimitAction": "none",
"StartLimitBurst": "5",
"StartLimitIntervalUSec": "10s",
"StartupBlockIOWeight": "[not set]",
"StartupCPUShares": "[not set]",
"StartupCPUWeight": "[not set]",
"StartupIOWeight": "[not set]",
"StartupMemoryHigh": "18446744073709551615",
"StartupMemoryLow": "0",
"StartupMemoryMax": "18446744073709551615",
"StartupMemorySwapMax": "18446744073709551615",
"StartupMemoryZSwapMax": "18446744073709551615",
"StateChangeTimestamp": "Wed 2024-07-10 05:13:58 UTC",
"StateChangeTimestampMonotonic": "959271314",
"StateDirectoryMode": "0755",
"StatusErrno": "0",
"StopWhenUnneeded": "no",
"SubState": "exited",
"SuccessAction": "none",
"SyslogFacility": "3",
"SyslogIdentifier": "quadlet-demo-network",
"SyslogLevel": "6",
"SyslogLevelPrefix": "yes",
"SyslogPriority": "30",
"SystemCallErrorNumber": "2147483646",
"TTYReset": "no",
"TTYVHangup": "no",
"TTYVTDisallocate": "no",
"TasksAccounting": "yes",
"TasksCurrent": "[not set]",
"TasksMax": "4426",
"TimeoutAbortUSec": "45s",
"TimeoutCleanUSec": "infinity",
"TimeoutStartFailureMode": "terminate",
"TimeoutStartUSec": "infinity",
"TimeoutStopFailureMode": "abort",
"TimeoutStopUSec": "45s",
"TimerSlackNSec": "50000",
"Transient": "no",
"Type": "oneshot",
"UID": "[not set]",
"UMask": "0022",
"UnitFilePreset": "disabled",
"UnitFileState": "generated",
"UtmpMode": "init",
"WatchdogSignal": "6",
"WatchdogTimestampMonotonic": "0",
"WatchdogUSec": "0"
}
}
TASK [fedora.linux_system_roles.podman : Remove quadlet file] ******************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:33
Wednesday 10 July 2024 05:15:34 +0000 (0:00:00.832) 0:03:38.918 ********
changed: [sut] => {
"changed": true,
"path": "/etc/containers/systemd/quadlet-demo.network",
"state": "absent"
}
TASK [fedora.linux_system_roles.podman : Manage linger] ************************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:39
Wednesday 10 July 2024 05:15:34 +0000 (0:00:00.241) 0:03:39.159 ********
included: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for sut
TASK [fedora.linux_system_roles.podman : Enable linger if needed] **************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12
Wednesday 10 July 2024 05:15:34 +0000 (0:00:00.035) 0:03:39.195 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_rootless | bool",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18
Wednesday 10 July 2024 05:15:34 +0000 (0:00:00.021) 0:03:39.216 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_rootless | bool",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22
Wednesday 10 July 2024 05:15:34 +0000 (0:00:00.020) 0:03:39.237 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_rootless | bool",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Reload systemctl] *********************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:47
Wednesday 10 July 2024 05:15:34 +0000 (0:00:00.020) 0:03:39.258 ********
ok: [sut] => {
"changed": false,
"name": null,
"status": {}
}
TASK [fedora.linux_system_roles.podman : Prune images no longer in use] ********
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:56
Wednesday 10 July 2024 05:15:35 +0000 (0:00:00.788) 0:03:40.046 ********
changed: [sut] => {
"changed": true,
"cmd": [
"podman",
"image",
"prune",
"--all",
"-f"
],
"delta": "0:00:00.038657",
"end": "2024-07-10 05:15:35.619304",
"rc": 0,
"start": "2024-07-10 05:15:35.580647"
}
TASK [fedora.linux_system_roles.podman : Remove networks no longer in use] *****
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:64
Wednesday 10 July 2024 05:15:35 +0000 (0:00:00.275) 0:03:40.322 ********
changed: [sut] => {
"changed": true,
"cmd": [
"podman",
"network",
"prune",
"-f"
],
"delta": "0:00:00.035443",
"end": "2024-07-10 05:15:35.890721",
"rc": 0,
"start": "2024-07-10 05:15:35.855278"
}
STDOUT:
systemd-quadlet-demo
TASK [fedora.linux_system_roles.podman : Remove volumes no longer in use] ******
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:73
Wednesday 10 July 2024 05:15:35 +0000 (0:00:00.271) 0:03:40.594 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_quadlet_type == \"volume\"",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : For testing and debugging - images] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:82
Wednesday 10 July 2024 05:15:35 +0000 (0:00:00.023) 0:03:40.617 ********
ok: [sut] => {
"changed": false,
"cmd": [
"podman",
"images",
"-n"
],
"delta": "0:00:00.033808",
"end": "2024-07-10 05:15:36.188924",
"rc": 0,
"start": "2024-07-10 05:15:36.155116"
}
TASK [fedora.linux_system_roles.podman : For testing and debugging - volumes] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:92
Wednesday 10 July 2024 05:15:36 +0000 (0:00:00.272) 0:03:40.889 ********
ok: [sut] => {
"changed": false,
"cmd": [
"podman",
"volume",
"ls",
"-n"
],
"delta": "0:00:00.031044",
"end": "2024-07-10 05:15:36.456793",
"rc": 0,
"start": "2024-07-10 05:15:36.425749"
}
TASK [fedora.linux_system_roles.podman : For testing and debugging - containers] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:102
Wednesday 10 July 2024 05:15:36 +0000 (0:00:00.269) 0:03:41.159 ********
ok: [sut] => {
"changed": false,
"cmd": [
"podman",
"ps",
"--noheading"
],
"delta": "0:00:00.036209",
"end": "2024-07-10 05:15:36.731388",
"rc": 0,
"start": "2024-07-10 05:15:36.695179"
}
TASK [fedora.linux_system_roles.podman : For testing and debugging - networks] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:112
Wednesday 10 July 2024 05:15:36 +0000 (0:00:00.317) 0:03:41.476 ********
ok: [sut] => {
"changed": false,
"cmd": [
"podman",
"network",
"ls",
"-n",
"-q"
],
"delta": "0:00:00.034849",
"end": "2024-07-10 05:15:37.049063",
"rc": 0,
"start": "2024-07-10 05:15:37.014214"
}
STDOUT:
podman
TASK [fedora.linux_system_roles.podman : For testing and debugging - secrets] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:122
Wednesday 10 July 2024 05:15:37 +0000 (0:00:00.276) 0:03:41.752 ********
ok: [sut] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : For testing and debugging - services] ***
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:133
Wednesday 10 July 2024 05:15:37 +0000 (0:00:00.268) 0:03:42.020 ********
ok: [sut] => {
"ansible_facts": {
"services": {
"NetworkManager-dispatcher.service": {
"name": "NetworkManager-dispatcher.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"NetworkManager-wait-online.service": {
"name": "NetworkManager-wait-online.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"NetworkManager.service": {
"name": "NetworkManager.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"arp-ethers.service": {
"name": "arp-ethers.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"auditd.service": {
"name": "auditd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"auth-rpcgss-module.service": {
"name": "auth-rpcgss-module.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"autovt@.service": {
"name": "autovt@.service",
"source": "systemd",
"state": "unknown",
"status": "alias"
},
"blk-availability.service": {
"name": "blk-availability.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"bluetooth.service": {
"name": "bluetooth.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"certmonger.service": {
"name": "certmonger.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"chrony-wait.service": {
"name": "chrony-wait.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"chronyd-restricted.service": {
"name": "chronyd-restricted.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"chronyd.service": {
"name": "chronyd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"cloud-config.service": {
"name": "cloud-config.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"cloud-final.service": {
"name": "cloud-final.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"cloud-init-hotplugd.service": {
"name": "cloud-init-hotplugd.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"cloud-init-local.service": {
"name": "cloud-init-local.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"cloud-init.service": {
"name": "cloud-init.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"console-getty.service": {
"name": "console-getty.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"container-getty@.service": {
"name": "container-getty@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"dbus-broker.service": {
"name": "dbus-broker.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"dbus-org.bluez.service": {
"name": "dbus-org.bluez.service",
"source": "systemd",
"state": "inactive",
"status": "alias"
},
"dbus-org.fedoraproject.FirewallD1.service": {
"name": "dbus-org.fedoraproject.FirewallD1.service",
"source": "systemd",
"state": "active",
"status": "alias"
},
"dbus-org.freedesktop.home1.service": {
"name": "dbus-org.freedesktop.home1.service",
"source": "systemd",
"state": "active",
"status": "alias"
},
"dbus-org.freedesktop.hostname1.service": {
"name": "dbus-org.freedesktop.hostname1.service",
"source": "systemd",
"state": "inactive",
"status": "alias"
},
"dbus-org.freedesktop.locale1.service": {
"name": "dbus-org.freedesktop.locale1.service",
"source": "systemd",
"state": "inactive",
"status": "alias"
},
"dbus-org.freedesktop.login1.service": {
"name": "dbus-org.freedesktop.login1.service",
"source": "systemd",
"state": "active",
"status": "alias"
},
"dbus-org.freedesktop.nm-dispatcher.service": {
"name": "dbus-org.freedesktop.nm-dispatcher.service",
"source": "systemd",
"state": "inactive",
"status": "alias"
},
"dbus-org.freedesktop.oom1.service": {
"name": "dbus-org.freedesktop.oom1.service",
"source": "systemd",
"state": "active",
"status": "alias"
},
"dbus-org.freedesktop.portable1.service": {
"name": "dbus-org.freedesktop.portable1.service",
"source": "systemd",
"state": "inactive",
"status": "alias"
},
"dbus-org.freedesktop.resolve1.service": {
"name": "dbus-org.freedesktop.resolve1.service",
"source": "systemd",
"state": "active",
"status": "alias"
},
"dbus-org.freedesktop.timedate1.service": {
"name": "dbus-org.freedesktop.timedate1.service",
"source": "systemd",
"state": "inactive",
"status": "alias"
},
"dbus.service": {
"name": "dbus.service",
"source": "systemd",
"state": "active",
"status": "alias"
},
"debug-shell.service": {
"name": "debug-shell.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"display-manager.service": {
"name": "display-manager.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"dm-event.service": {
"name": "dm-event.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dnf-makecache.service": {
"name": "dnf-makecache.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dnf-system-upgrade-cleanup.service": {
"name": "dnf-system-upgrade-cleanup.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dnf-system-upgrade.service": {
"name": "dnf-system-upgrade.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"dracut-cmdline.service": {
"name": "dracut-cmdline.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-initqueue.service": {
"name": "dracut-initqueue.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-mount.service": {
"name": "dracut-mount.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-mount.service": {
"name": "dracut-pre-mount.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-pivot.service": {
"name": "dracut-pre-pivot.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-trigger.service": {
"name": "dracut-pre-trigger.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-udev.service": {
"name": "dracut-pre-udev.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-shutdown-onfailure.service": {
"name": "dracut-shutdown-onfailure.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-shutdown.service": {
"name": "dracut-shutdown.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"ebtables.service": {
"name": "ebtables.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"emergency.service": {
"name": "emergency.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"firewalld.service": {
"name": "firewalld.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"fsidd.service": {
"name": "fsidd.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"fstrim.service": {
"name": "fstrim.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"fwupd-offline-update.service": {
"name": "fwupd-offline-update.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"fwupd-refresh.service": {
"name": "fwupd-refresh.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"fwupd.service": {
"name": "fwupd.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"getty@.service": {
"name": "getty@.service",
"source": "systemd",
"state": "unknown",
"status": "enabled"
},
"getty@tty1.service": {
"name": "getty@tty1.service",
"source": "systemd",
"state": "running",
"status": "active"
},
"grub-boot-indeterminate.service": {
"name": "grub-boot-indeterminate.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"grub2-systemd-integration.service": {
"name": "grub2-systemd-integration.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"gssproxy.service": {
"name": "gssproxy.service",
"source": "systemd",
"state": "running",
"status": "disabled"
},
"hv_kvp_daemon.service": {
"name": "hv_kvp_daemon.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"import-state.service": {
"name": "import-state.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"initrd-cleanup.service": {
"name": "initrd-cleanup.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"initrd-parse-etc.service": {
"name": "initrd-parse-etc.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"initrd-switch-root.service": {
"name": "initrd-switch-root.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"initrd-udevadm-cleanup-db.service": {
"name": "initrd-udevadm-cleanup-db.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"ip6tables.service": {
"name": "ip6tables.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"ipset.service": {
"name": "ipset.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"iptables.service": {
"name": "iptables.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"kmod-static-nodes.service": {
"name": "kmod-static-nodes.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"ldconfig.service": {
"name": "ldconfig.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"loadmodules.service": {
"name": "loadmodules.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"lvm2-lvmpolld.service": {
"name": "lvm2-lvmpolld.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"lvm2-monitor.service": {
"name": "lvm2-monitor.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"man-db-cache-update.service": {
"name": "man-db-cache-update.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"man-db-restart-cache-update.service": {
"name": "man-db-restart-cache-update.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"mdadm-grow-continue@.service": {
"name": "mdadm-grow-continue@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"mdadm-last-resort@.service": {
"name": "mdadm-last-resort@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"mdcheck_continue.service": {
"name": "mdcheck_continue.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"mdcheck_start.service": {
"name": "mdcheck_start.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"mdmon@.service": {
"name": "mdmon@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"mdmonitor-oneshot.service": {
"name": "mdmonitor-oneshot.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"mdmonitor.service": {
"name": "mdmonitor.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"modprobe@.service": {
"name": "modprobe@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"modprobe@configfs.service": {
"name": "modprobe@configfs.service",
"source": "systemd",
"state": "stopped",
"status": "inactive"
},
"modprobe@dm_mod.service": {
"name": "modprobe@dm_mod.service",
"source": "systemd",
"state": "stopped",
"status": "inactive"
},
"modprobe@drm.service": {
"name": "modprobe@drm.service",
"source": "systemd",
"state": "stopped",
"status": "inactive"
},
"modprobe@efi_pstore.service": {
"name": "modprobe@efi_pstore.service",
"source": "systemd",
"state": "stopped",
"status": "inactive"
},
"modprobe@fuse.service": {
"name": "modprobe@fuse.service",
"source": "systemd",
"state": "stopped",
"status": "inactive"
},
"modprobe@loop.service": {
"name": "modprobe@loop.service",
"source": "systemd",
"state": "stopped",
"status": "inactive"
},
"netavark-dhcp-proxy.service": {
"name": "netavark-dhcp-proxy.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"netavark-firewalld-reload.service": {
"name": "netavark-firewalld-reload.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"network.service": {
"name": "network.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"nfs-blkmap.service": {
"name": "nfs-blkmap.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"nfs-idmapd.service": {
"name": "nfs-idmapd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nfs-mountd.service": {
"name": "nfs-mountd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nfs-server.service": {
"name": "nfs-server.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"nfs-utils.service": {
"name": "nfs-utils.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nfsdcld.service": {
"name": "nfsdcld.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nftables.service": {
"name": "nftables.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"nis-domainname.service": {
"name": "nis-domainname.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"nm-priv-helper.service": {
"name": "nm-priv-helper.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"ntpd.service": {
"name": "ntpd.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"ntpdate.service": {
"name": "ntpdate.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"pam_namespace.service": {
"name": "pam_namespace.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"pcscd.service": {
"name": "pcscd.service",
"source": "systemd",
"state": "stopped",
"status": "indirect"
},
"plymouth-halt.service": {
"name": "plymouth-halt.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-kexec.service": {
"name": "plymouth-kexec.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-poweroff.service": {
"name": "plymouth-poweroff.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-quit-wait.service": {
"name": "plymouth-quit-wait.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-quit.service": {
"name": "plymouth-quit.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-read-write.service": {
"name": "plymouth-read-write.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-reboot.service": {
"name": "plymouth-reboot.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-start.service": {
"name": "plymouth-start.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-switch-root-initramfs.service": {
"name": "plymouth-switch-root-initramfs.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-switch-root.service": {
"name": "plymouth-switch-root.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"podman-auto-update.service": {
"name": "podman-auto-update.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"podman-clean-transient.service": {
"name": "podman-clean-transient.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"podman-kube@.service": {
"name": "podman-kube@.service",
"source": "systemd",
"state": "unknown",
"status": "disabled"
},
"podman-restart.service": {
"name": "podman-restart.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"podman.service": {
"name": "podman.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"polkit.service": {
"name": "polkit.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"quadlet-demo-mysql.service": {
"name": "quadlet-demo-mysql.service",
"source": "systemd",
"state": "stopped",
"status": "failed"
},
"quotaon.service": {
"name": "quotaon.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"raid-check.service": {
"name": "raid-check.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rc-local.service": {
"name": "rc-local.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rescue.service": {
"name": "rescue.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"restraintd.service": {
"name": "restraintd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"rngd.service": {
"name": "rngd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"rpc-gssd.service": {
"name": "rpc-gssd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rpc-statd-notify.service": {
"name": "rpc-statd-notify.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rpc-statd.service": {
"name": "rpc-statd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rpc-svcgssd.service": {
"name": "rpc-svcgssd.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"rpcbind.service": {
"name": "rpcbind.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"rpmdb-migrate.service": {
"name": "rpmdb-migrate.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"rpmdb-rebuild.service": {
"name": "rpmdb-rebuild.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"selinux-autorelabel-mark.service": {
"name": "selinux-autorelabel-mark.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"selinux-autorelabel.service": {
"name": "selinux-autorelabel.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"selinux-check-proper-disable.service": {
"name": "selinux-check-proper-disable.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"serial-getty@.service": {
"name": "serial-getty@.service",
"source": "systemd",
"state": "unknown",
"status": "indirect"
},
"serial-getty@ttyS0.service": {
"name": "serial-getty@ttyS0.service",
"source": "systemd",
"state": "running",
"status": "active"
},
"snapd.seeded.service": {
"name": "snapd.seeded.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"sntp.service": {
"name": "sntp.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"ssh-host-keys-migration.service": {
"name": "ssh-host-keys-migration.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"sshd-keygen.service": {
"name": "sshd-keygen.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"sshd-keygen@.service": {
"name": "sshd-keygen@.service",
"source": "systemd",
"state": "unknown",
"status": "disabled"
},
"sshd-keygen@ecdsa.service": {
"name": "sshd-keygen@ecdsa.service",
"source": "systemd",
"state": "stopped",
"status": "inactive"
},
"sshd-keygen@ed25519.service": {
"name": "sshd-keygen@ed25519.service",
"source": "systemd",
"state": "stopped",
"status": "inactive"
},
"sshd-keygen@rsa.service": {
"name": "sshd-keygen@rsa.service",
"source": "systemd",
"state": "stopped",
"status": "inactive"
},
"sshd.service": {
"name": "sshd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"sshd@.service": {
"name": "sshd@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"sssd-autofs.service": {
"name": "sssd-autofs.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-kcm.service": {
"name": "sssd-kcm.service",
"source": "systemd",
"state": "stopped",
"status": "indirect"
},
"sssd-nss.service": {
"name": "sssd-nss.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-pac.service": {
"name": "sssd-pac.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-pam.service": {
"name": "sssd-pam.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-ssh.service": {
"name": "sssd-ssh.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-sudo.service": {
"name": "sssd-sudo.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd.service": {
"name": "sssd.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"syslog.service": {
"name": "syslog.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"system-update-cleanup.service": {
"name": "system-update-cleanup.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-ask-password-console.service": {
"name": "systemd-ask-password-console.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-ask-password-plymouth.service": {
"name": "systemd-ask-password-plymouth.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-ask-password-wall.service": {
"name": "systemd-ask-password-wall.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-backlight@.service": {
"name": "systemd-backlight@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-battery-check.service": {
"name": "systemd-battery-check.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-binfmt.service": {
"name": "systemd-binfmt.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-bless-boot.service": {
"name": "systemd-bless-boot.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-boot-check-no-failures.service": {
"name": "systemd-boot-check-no-failures.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"systemd-boot-random-seed.service": {
"name": "systemd-boot-random-seed.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-boot-update.service": {
"name": "systemd-boot-update.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"systemd-confext.service": {
"name": "systemd-confext.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"systemd-coredump@.service": {
"name": "systemd-coredump@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-exit.service": {
"name": "systemd-exit.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-firstboot.service": {
"name": "systemd-firstboot.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-fsck-root.service": {
"name": "systemd-fsck-root.service",
"source": "systemd",
"state": "stopped",
"status": "enabled-runtime"
},
"systemd-fsck@.service": {
"name": "systemd-fsck@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-growfs-root.service": {
"name": "systemd-growfs-root.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-growfs@.service": {
"name": "systemd-growfs@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-halt.service": {
"name": "systemd-halt.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-hibernate.service": {
"name": "systemd-hibernate.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-homed-activate.service": {
"name": "systemd-homed-activate.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"systemd-homed.service": {
"name": "systemd-homed.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"systemd-hostnamed.service": {
"name": "systemd-hostnamed.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-hwdb-update.service": {
"name": "systemd-hwdb-update.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-hybrid-sleep.service": {
"name": "systemd-hybrid-sleep.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-initctl.service": {
"name": "systemd-initctl.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-journal-catalog-update.service": {
"name": "systemd-journal-catalog-update.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-journal-flush.service": {
"name": "systemd-journal-flush.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-journald.service": {
"name": "systemd-journald.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"systemd-journald@.service": {
"name": "systemd-journald@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-kexec.service": {
"name": "systemd-kexec.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-localed.service": {
"name": "systemd-localed.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-logind.service": {
"name": "systemd-logind.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"systemd-machine-id-commit.service": {
"name": "systemd-machine-id-commit.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-modules-load.service": {
"name": "systemd-modules-load.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-network-generator.service": {
"name": "systemd-network-generator.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"systemd-networkd-wait-online.service": {
"name": "systemd-networkd-wait-online.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"systemd-networkd-wait-online@.service": {
"name": "systemd-networkd-wait-online@.service",
"source": "systemd",
"state": "unknown",
"status": "disabled"
},
"systemd-networkd.service": {
"name": "systemd-networkd.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"systemd-oomd.service": {
"name": "systemd-oomd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"systemd-pcrfs-root.service": {
"name": "systemd-pcrfs-root.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-pcrfs@.service": {
"name": "systemd-pcrfs@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-pcrmachine.service": {
"name": "systemd-pcrmachine.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-pcrphase-initrd.service": {
"name": "systemd-pcrphase-initrd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-pcrphase-sysinit.service": {
"name": "systemd-pcrphase-sysinit.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-pcrphase.service": {
"name": "systemd-pcrphase.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-portabled.service": {
"name": "systemd-portabled.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-poweroff.service": {
"name": "systemd-poweroff.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-pstore.service": {
"name": "systemd-pstore.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"systemd-quotacheck.service": {
"name": "systemd-quotacheck.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-random-seed.service": {
"name": "systemd-random-seed.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-reboot.service": {
"name": "systemd-reboot.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-remount-fs.service": {
"name": "systemd-remount-fs.service",
"source": "systemd",
"state": "stopped",
"status": "enabled-runtime"
},
"systemd-repart.service": {
"name": "systemd-repart.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-resolved.service": {
"name": "systemd-resolved.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"systemd-rfkill.service": {
"name": "systemd-rfkill.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-soft-reboot.service": {
"name": "systemd-soft-reboot.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-suspend-then-hibernate.service": {
"name": "systemd-suspend-then-hibernate.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-suspend.service": {
"name": "systemd-suspend.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-sysctl.service": {
"name": "systemd-sysctl.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-sysext.service": {
"name": "systemd-sysext.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"systemd-sysupdate-reboot.service": {
"name": "systemd-sysupdate-reboot.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"systemd-sysupdate.service": {
"name": "systemd-sysupdate.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"systemd-sysusers.service": {
"name": "systemd-sysusers.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-time-wait-sync.service": {
"name": "systemd-time-wait-sync.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"systemd-timedated.service": {
"name": "systemd-timedated.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-timesyncd.service": {
"name": "systemd-timesyncd.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"systemd-tmpfiles-clean.service": {
"name": "systemd-tmpfiles-clean.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-tmpfiles-setup-dev-early.service": {
"name": "systemd-tmpfiles-setup-dev-early.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-tmpfiles-setup-dev.service": {
"name": "systemd-tmpfiles-setup-dev.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-tmpfiles-setup.service": {
"name": "systemd-tmpfiles-setup.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-udev-settle.service": {
"name": "systemd-udev-settle.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-udev-trigger.service": {
"name": "systemd-udev-trigger.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-udevd.service": {
"name": "systemd-udevd.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"systemd-update-done.service": {
"name": "systemd-update-done.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-update-utmp-runlevel.service": {
"name": "systemd-update-utmp-runlevel.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-update-utmp.service": {
"name": "systemd-update-utmp.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-user-sessions.service": {
"name": "systemd-user-sessions.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-userdbd.service": {
"name": "systemd-userdbd.service",
"source": "systemd",
"state": "running",
"status": "indirect"
},
"systemd-vconsole-setup.service": {
"name": "systemd-vconsole-setup.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-volatile-root.service": {
"name": "systemd-volatile-root.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-zram-setup@.service": {
"name": "systemd-zram-setup@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-zram-setup@zram0.service": {
"name": "systemd-zram-setup@zram0.service",
"source": "systemd",
"state": "stopped",
"status": "active"
},
"udisks2.service": {
"name": "udisks2.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"unbound-anchor.service": {
"name": "unbound-anchor.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"user-runtime-dir@.service": {
"name": "user-runtime-dir@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"user-runtime-dir@0.service": {
"name": "user-runtime-dir@0.service",
"source": "systemd",
"state": "stopped",
"status": "active"
},
"user@.service": {
"name": "user@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"user@0.service": {
"name": "user@0.service",
"source": "systemd",
"state": "running",
"status": "active"
}
}
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Create and update quadlets] ***********
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:114
Wednesday 10 July 2024 05:15:39 +0000 (0:00:02.220) 0:03:44.241 ********
skipping: [sut] => {
"changed": false,
"false_condition": "__podman_state != \"absent\"",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Cancel linger] ************************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:143
Wednesday 10 July 2024 05:15:39 +0000 (0:00:00.020) 0:03:44.262 ********
skipping: [sut] => {
"changed": false,
"skipped_reason": "No items in the list"
}
TASK [fedora.linux_system_roles.podman : Handle credential files - absent] *****
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:149
Wednesday 10 July 2024 05:15:39 +0000 (0:00:00.019) 0:03:44.282 ********
skipping: [sut] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : Handle certs.d files - absent] ********
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:158
Wednesday 10 July 2024 05:15:39 +0000 (0:00:00.019) 0:03:44.301 ********
skipping: [sut] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [Ensure no resources] *****************************************************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/tests/tests_quadlet_demo.yml:170
Wednesday 10 July 2024 05:15:39 +0000 (0:00:00.031) 0:03:44.332 ********
fatal: [sut]: FAILED! => {
"assertion": "ansible_facts[\"services\"] | dict2items | rejectattr(\"value.status\", \"match\", \"not-found\") | selectattr(\"key\", \"match\", \"quadlet-demo\") | list | length == 0",
"changed": false,
"evaluated_to": false
}
MSG:
Assertion failed
TASK [Debug] *******************************************************************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/tests/tests_quadlet_demo.yml:181
Wednesday 10 July 2024 05:15:39 +0000 (0:00:00.036) 0:03:44.369 ********
ok: [sut] => {
"changed": false,
"cmd": "exec 1>&2\nset -x\nset -o pipefail\nsystemctl list-units --plain -l --all | grep quadlet || :\nsystemctl list-unit-files --all | grep quadlet || :\nsystemctl list-units --plain --failed -l --all | grep quadlet || :\njournalctl -u quadlet-demo-mysql.service || :\nsystemctl status quadlet-demo-mysql.service || :\n",
"delta": "0:00:00.596940",
"end": "2024-07-10 05:15:40.494700",
"rc": 0,
"start": "2024-07-10 05:15:39.897760"
}
STDERR:
+ set -o pipefail
+ systemctl list-units --plain -l --all
+ grep quadlet
quadlet-demo-mysql.service not-found failed failed quadlet-demo-mysql.service
+ systemctl list-unit-files --all
+ grep quadlet
+ :
+ systemctl list-units --plain --failed -l --all
+ grep quadlet
quadlet-demo-mysql.service not-found failed failed quadlet-demo-mysql.service
+ journalctl -u quadlet-demo-mysql.service
Jul 10 05:14:12 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: Starting quadlet-demo-mysql.service...
Jul 10 05:14:12 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm podman[7928]: 2024-07-10 05:14:12.80841545 +0000 UTC m=+0.052126110 container create 21b1326f57855b147efcb07cbd10c5f7b7edbabb6693b7752105182196a3f742 (image=quay.io/linux-system-roles/mysql:5.6, name=quadlet-demo-mysql, PODMAN_SYSTEMD_UNIT=quadlet-demo-mysql.service)
Jul 10 05:14:12 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm podman[7928]: 2024-07-10 05:14:12.786708435 +0000 UTC m=+0.030419211 image pull dd3b2a5dcb48ff61113592ed5ddd762581be4387c7bc552375a2159422aa6bf5 quay.io/linux-system-roles/mysql:5.6
Jul 10 05:14:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm podman[7928]: 2024-07-10 05:14:13.195131115 +0000 UTC m=+0.438841882 container init 21b1326f57855b147efcb07cbd10c5f7b7edbabb6693b7752105182196a3f742 (image=quay.io/linux-system-roles/mysql:5.6, name=quadlet-demo-mysql, PODMAN_SYSTEMD_UNIT=quadlet-demo-mysql.service)
Jul 10 05:14:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: Started quadlet-demo-mysql.service.
Jul 10 05:14:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:13+00:00 [Note] [Entrypoint]: Entrypoint script for MySQL Server 5.6.51-1debian9 started.
Jul 10 05:14:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm podman[7928]: 2024-07-10 05:14:13.232698056 +0000 UTC m=+0.476408784 container start 21b1326f57855b147efcb07cbd10c5f7b7edbabb6693b7752105182196a3f742 (image=quay.io/linux-system-roles/mysql:5.6, name=quadlet-demo-mysql, PODMAN_SYSTEMD_UNIT=quadlet-demo-mysql.service)
Jul 10 05:14:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[7928]: 21b1326f57855b147efcb07cbd10c5f7b7edbabb6693b7752105182196a3f742
Jul 10 05:14:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:13+00:00 [Note] [Entrypoint]: Switching to dedicated user 'mysql'
Jul 10 05:14:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:13+00:00 [Note] [Entrypoint]: Entrypoint script for MySQL Server 5.6.51-1debian9 started.
Jul 10 05:14:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:13+00:00 [Note] [Entrypoint]: Initializing database files
Jul 10 05:14:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:13 0 [Warning] TIMESTAMP with implicit DEFAULT value is deprecated. Please use --explicit_defaults_for_timestamp server option (see documentation for more details).
Jul 10 05:14:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:13 0 [Note] Ignoring --secure-file-priv value as server is running with --bootstrap.
Jul 10 05:14:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:13 0 [Note] /usr/sbin/mysqld (mysqld 5.6.51) starting as process 42 ...
Jul 10 05:14:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:13 42 [Note] InnoDB: Using atomics to ref count buffer pool pages
Jul 10 05:14:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:13 42 [Note] InnoDB: The InnoDB memory heap is disabled
Jul 10 05:14:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:13 42 [Note] InnoDB: Mutexes and rw_locks use GCC atomic builtins
Jul 10 05:14:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:13 42 [Note] InnoDB: Memory barrier is not used
Jul 10 05:14:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:13 42 [Note] InnoDB: Compressed tables use zlib 1.2.11
Jul 10 05:14:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:13 42 [Note] InnoDB: Using Linux native AIO
Jul 10 05:14:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:13 42 [Note] InnoDB: Using CPU crc32 instructions
Jul 10 05:14:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:13 42 [Note] InnoDB: Initializing buffer pool, size = 128.0M
Jul 10 05:14:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:13 42 [Note] InnoDB: Completed initialization of buffer pool
Jul 10 05:14:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:13 42 [Note] InnoDB: The first specified data file ./ibdata1 did not exist: a new database to be created!
Jul 10 05:14:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:13 42 [Note] InnoDB: Setting file ./ibdata1 size to 12 MB
Jul 10 05:14:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:13 42 [Note] InnoDB: Database physically writes the file full: wait...
Jul 10 05:14:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:13 42 [Note] InnoDB: Setting log file ./ib_logfile101 size to 48 MB
Jul 10 05:14:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:13 42 [Note] InnoDB: Setting log file ./ib_logfile1 size to 48 MB
Jul 10 05:14:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:14 42 [Note] InnoDB: Renaming log file ./ib_logfile101 to ./ib_logfile0
Jul 10 05:14:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:14 42 [Warning] InnoDB: New log files created, LSN=45781
Jul 10 05:14:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:14 42 [Note] InnoDB: Doublewrite buffer not found: creating new
Jul 10 05:14:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:14 42 [Note] InnoDB: Doublewrite buffer created
Jul 10 05:14:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:14 42 [Note] InnoDB: 128 rollback segment(s) are active.
Jul 10 05:14:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:14 42 [Warning] InnoDB: Creating foreign key constraint system tables.
Jul 10 05:14:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:14 42 [Note] InnoDB: Foreign key constraint system tables created
Jul 10 05:14:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:14 42 [Note] InnoDB: Creating tablespace and datafile system tables.
Jul 10 05:14:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:14 42 [Note] InnoDB: Tablespace and datafile system tables created.
Jul 10 05:14:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:14 42 [Note] InnoDB: Waiting for purge to start
Jul 10 05:14:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:14 42 [Note] InnoDB: 5.6.51 started; log sequence number 0
Jul 10 05:14:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:14 42 [Note] RSA private key file not found: /var/lib/mysql//private_key.pem. Some authentication plugins will not work.
Jul 10 05:14:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:14 42 [Note] RSA public key file not found: /var/lib/mysql//public_key.pem. Some authentication plugins will not work.
Jul 10 05:14:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:14 42 [Note] Binlog end
Jul 10 05:14:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:14 42 [Note] InnoDB: FTS optimize thread exiting.
Jul 10 05:14:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:14 42 [Note] InnoDB: Starting shutdown...
Jul 10 05:14:16 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:16 42 [Note] InnoDB: Shutdown completed; log sequence number 1625977
Jul 10 05:14:16 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]:
Jul 10 05:14:16 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]:
Jul 10 05:14:16 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:16 0 [Warning] TIMESTAMP with implicit DEFAULT value is deprecated. Please use --explicit_defaults_for_timestamp server option (see documentation for more details).
Jul 10 05:14:16 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:16 0 [Note] Ignoring --secure-file-priv value as server is running with --bootstrap.
Jul 10 05:14:16 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:16 0 [Note] /usr/sbin/mysqld (mysqld 5.6.51) starting as process 65 ...
Jul 10 05:14:16 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:16 65 [Note] InnoDB: Using atomics to ref count buffer pool pages
Jul 10 05:14:16 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:16 65 [Note] InnoDB: The InnoDB memory heap is disabled
Jul 10 05:14:16 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:16 65 [Note] InnoDB: Mutexes and rw_locks use GCC atomic builtins
Jul 10 05:14:16 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:16 65 [Note] InnoDB: Memory barrier is not used
Jul 10 05:14:16 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:16 65 [Note] InnoDB: Compressed tables use zlib 1.2.11
Jul 10 05:14:16 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:16 65 [Note] InnoDB: Using Linux native AIO
Jul 10 05:14:16 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:16 65 [Note] InnoDB: Using CPU crc32 instructions
Jul 10 05:14:16 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:16 65 [Note] InnoDB: Initializing buffer pool, size = 128.0M
Jul 10 05:14:16 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:16 65 [Note] InnoDB: Completed initialization of buffer pool
Jul 10 05:14:16 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:16 65 [Note] InnoDB: Highest supported file format is Barracuda.
Jul 10 05:14:16 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:16 65 [Note] InnoDB: 128 rollback segment(s) are active.
Jul 10 05:14:16 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:16 65 [Note] InnoDB: Waiting for purge to start
Jul 10 05:14:16 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:16 65 [Note] InnoDB: 5.6.51 started; log sequence number 1625977
Jul 10 05:14:16 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:16 65 [Note] RSA private key file not found: /var/lib/mysql//private_key.pem. Some authentication plugins will not work.
Jul 10 05:14:16 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:16 65 [Note] RSA public key file not found: /var/lib/mysql//public_key.pem. Some authentication plugins will not work.
Jul 10 05:14:16 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:16 65 [Note] Binlog end
Jul 10 05:14:16 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:16 65 [Note] InnoDB: FTS optimize thread exiting.
Jul 10 05:14:16 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:16 65 [Note] InnoDB: Starting shutdown...
Jul 10 05:14:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:18 65 [Note] InnoDB: Shutdown completed; log sequence number 1625987
Jul 10 05:14:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]:
Jul 10 05:14:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]:
Jul 10 05:14:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]:
Jul 10 05:14:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]:
Jul 10 05:14:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: PLEASE REMEMBER TO SET A PASSWORD FOR THE MySQL root USER !
Jul 10 05:14:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: To do so, start the server, then issue the following commands:
Jul 10 05:14:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]:
Jul 10 05:14:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: /usr/bin/mysqladmin -u root password 'new-password'
Jul 10 05:14:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: /usr/bin/mysqladmin -u root -h 21b1326f5785 password 'new-password'
Jul 10 05:14:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]:
Jul 10 05:14:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: Alternatively you can run:
Jul 10 05:14:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]:
Jul 10 05:14:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: /usr/bin/mysql_secure_installation
Jul 10 05:14:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]:
Jul 10 05:14:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: which will also give you the option of removing the test
Jul 10 05:14:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: databases and anonymous user created by default. This is
Jul 10 05:14:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: strongly recommended for production servers.
Jul 10 05:14:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]:
Jul 10 05:14:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: See the manual for more instructions.
Jul 10 05:14:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]:
Jul 10 05:14:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: Please report any problems at http://bugs.mysql.com/
Jul 10 05:14:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]:
Jul 10 05:14:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: The latest information about MySQL is available on the web at
Jul 10 05:14:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]:
Jul 10 05:14:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: http://www.mysql.com
Jul 10 05:14:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]:
Jul 10 05:14:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: Support MySQL by buying support/licenses at http://shop.mysql.com
Jul 10 05:14:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]:
Jul 10 05:14:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: Note: new default config file not created.
Jul 10 05:14:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: Please make sure your config file is current
Jul 10 05:14:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]:
Jul 10 05:14:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: WARNING: Default config file /etc/mysql/my.cnf exists on the system
Jul 10 05:14:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: This file will be read by default by the MySQL server
Jul 10 05:14:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: If you do not want to use this, either remove it, or use the
Jul 10 05:14:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: --defaults-file argument to mysqld_safe when starting the server
Jul 10 05:14:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]:
Jul 10 05:14:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:18+00:00 [Note] [Entrypoint]: Database files initialized
Jul 10 05:14:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:18+00:00 [Note] [Entrypoint]: Starting temporary server
Jul 10 05:14:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:18+00:00 [Note] [Entrypoint]: Waiting for server startup
Jul 10 05:14:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:19 0 [Warning] TIMESTAMP with implicit DEFAULT value is deprecated. Please use --explicit_defaults_for_timestamp server option (see documentation for more details).
Jul 10 05:14:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:19 0 [Note] mysqld (mysqld 5.6.51) starting as process 90 ...
Jul 10 05:14:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:19 90 [Note] Plugin 'FEDERATED' is disabled.
Jul 10 05:14:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:19 90 [Note] InnoDB: Using atomics to ref count buffer pool pages
Jul 10 05:14:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:19 90 [Note] InnoDB: The InnoDB memory heap is disabled
Jul 10 05:14:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:19 90 [Note] InnoDB: Mutexes and rw_locks use GCC atomic builtins
Jul 10 05:14:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:19 90 [Note] InnoDB: Memory barrier is not used
Jul 10 05:14:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:19 90 [Note] InnoDB: Compressed tables use zlib 1.2.11
Jul 10 05:14:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:19 90 [Note] InnoDB: Using Linux native AIO
Jul 10 05:14:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:19 90 [Note] InnoDB: Using CPU crc32 instructions
Jul 10 05:14:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:19 90 [Note] InnoDB: Initializing buffer pool, size = 128.0M
Jul 10 05:14:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:19 90 [Note] InnoDB: Completed initialization of buffer pool
Jul 10 05:14:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:19 90 [Note] InnoDB: Highest supported file format is Barracuda.
Jul 10 05:14:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:19 90 [Note] InnoDB: 128 rollback segment(s) are active.
Jul 10 05:14:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:19 90 [Note] InnoDB: Waiting for purge to start
Jul 10 05:14:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:19 90 [Note] InnoDB: 5.6.51 started; log sequence number 1625987
Jul 10 05:14:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:19 90 [Warning] No existing UUID has been found, so we assume that this is the first time that this server has been started. Generating a new UUID: 42dc40c5-3e7b-11ef-9a49-2a8356b5d061.
Jul 10 05:14:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:19 90 [Note] RSA private key file not found: /var/lib/mysql//private_key.pem. Some authentication plugins will not work.
Jul 10 05:14:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:19 90 [Note] RSA public key file not found: /var/lib/mysql//public_key.pem. Some authentication plugins will not work.
Jul 10 05:14:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:19 90 [Warning] Insecure configuration for --pid-file: Location '/var/run/mysqld' in the path is accessible to all OS users. Consider choosing a different directory.
Jul 10 05:14:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:19 90 [Warning] 'user' entry 'root@21b1326f5785' ignored in --skip-name-resolve mode.
Jul 10 05:14:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:19 90 [Warning] 'user' entry '@21b1326f5785' ignored in --skip-name-resolve mode.
Jul 10 05:14:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:19 90 [Warning] 'proxies_priv' entry '@ root@21b1326f5785' ignored in --skip-name-resolve mode.
Jul 10 05:14:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:19 90 [Note] Event Scheduler: Loaded 0 events
Jul 10 05:14:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:19 90 [Note] mysqld: ready for connections.
Jul 10 05:14:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: Version: '5.6.51' socket: '/var/run/mysqld/mysqld.sock' port: 0 MySQL Community Server (GPL)
Jul 10 05:14:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:19+00:00 [Note] [Entrypoint]: Temporary server started.
Jul 10 05:14:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: Warning: Unable to load '/usr/share/zoneinfo/iso3166.tab' as time zone. Skipping it.
Jul 10 05:14:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: Warning: Unable to load '/usr/share/zoneinfo/leap-seconds.list' as time zone. Skipping it.
Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: Warning: Unable to load '/usr/share/zoneinfo/zone.tab' as time zone. Skipping it.
Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: Warning: Unable to load '/usr/share/zoneinfo/zone1970.tab' as time zone. Skipping it.
Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:21 90 [Warning] 'proxies_priv' entry '@ root@21b1326f5785' ignored in --skip-name-resolve mode.
Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]:
Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:21+00:00 [Note] [Entrypoint]: Stopping temporary server
Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:21 90 [Note] mysqld: Normal shutdown
Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]:
Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:21 90 [Note] Giving 0 client threads a chance to die gracefully
Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:21 90 [Note] Event Scheduler: Purging the queue. 0 events
Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:21 90 [Note] Shutting down slave threads
Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:21 90 [Note] Forcefully disconnecting 0 remaining clients
Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:21 90 [Note] Binlog end
Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:21 90 [Note] Shutting down plugin 'partition'
Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:21 90 [Note] Shutting down plugin 'PERFORMANCE_SCHEMA'
Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:21 90 [Note] Shutting down plugin 'INNODB_SYS_DATAFILES'
Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:21 90 [Note] Shutting down plugin 'INNODB_SYS_TABLESPACES'
Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:21 90 [Note] Shutting down plugin 'INNODB_SYS_FOREIGN_COLS'
Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:21 90 [Note] Shutting down plugin 'INNODB_SYS_FOREIGN'
Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:21 90 [Note] Shutting down plugin 'INNODB_SYS_FIELDS'
Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:21 90 [Note] Shutting down plugin 'INNODB_SYS_COLUMNS'
Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:21 90 [Note] Shutting down plugin 'INNODB_SYS_INDEXES'
Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:21 90 [Note] Shutting down plugin 'INNODB_SYS_TABLESTATS'
Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:21 90 [Note] Shutting down plugin 'INNODB_SYS_TABLES'
Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:21 90 [Note] Shutting down plugin 'INNODB_FT_INDEX_TABLE'
Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:21 90 [Note] Shutting down plugin 'INNODB_FT_INDEX_CACHE'
Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:21 90 [Note] Shutting down plugin 'INNODB_FT_CONFIG'
Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:21 90 [Note] Shutting down plugin 'INNODB_FT_BEING_DELETED'
Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:21 90 [Note] Shutting down plugin 'INNODB_FT_DELETED'
Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:21 90 [Note] Shutting down plugin 'INNODB_FT_DEFAULT_STOPWORD'
Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:21 90 [Note] Shutting down plugin 'INNODB_METRICS'
Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:21 90 [Note] Shutting down plugin 'INNODB_BUFFER_POOL_STATS'
Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:21 90 [Note] Shutting down plugin 'INNODB_BUFFER_PAGE_LRU'
Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:21 90 [Note] Shutting down plugin 'INNODB_BUFFER_PAGE'
Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:21 90 [Note] Shutting down plugin 'INNODB_CMP_PER_INDEX_RESET'
Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:21 90 [Note] Shutting down plugin 'INNODB_CMP_PER_INDEX'
Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:21 90 [Note] Shutting down plugin 'INNODB_CMPMEM_RESET'
Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:21 90 [Note] Shutting down plugin 'INNODB_CMPMEM'
Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:21 90 [Note] Shutting down plugin 'INNODB_CMP_RESET'
Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:21 90 [Note] Shutting down plugin 'INNODB_CMP'
Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:21 90 [Note] Shutting down plugin 'INNODB_LOCK_WAITS'
Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:21 90 [Note] Shutting down plugin 'INNODB_LOCKS'
Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:21 90 [Note] Shutting down plugin 'INNODB_TRX'
Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:21 90 [Note] Shutting down plugin 'InnoDB'
Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:21 90 [Note] InnoDB: FTS optimize thread exiting.
Jul 10 05:14:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:21 90 [Note] InnoDB: Starting shutdown...
Jul 10 05:14:23 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:23 90 [Note] InnoDB: Shutdown completed; log sequence number 1625997
Jul 10 05:14:23 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:23 90 [Note] Shutting down plugin 'BLACKHOLE'
Jul 10 05:14:23 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:23 90 [Note] Shutting down plugin 'ARCHIVE'
Jul 10 05:14:23 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:23 90 [Note] Shutting down plugin 'MRG_MYISAM'
Jul 10 05:14:23 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:23 90 [Note] Shutting down plugin 'MyISAM'
Jul 10 05:14:23 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:23 90 [Note] Shutting down plugin 'MEMORY'
Jul 10 05:14:23 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:23 90 [Note] Shutting down plugin 'CSV'
Jul 10 05:14:23 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:23 90 [Note] Shutting down plugin 'sha256_password'
Jul 10 05:14:23 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:23 90 [Note] Shutting down plugin 'mysql_old_password'
Jul 10 05:14:23 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:23 90 [Note] Shutting down plugin 'mysql_native_password'
Jul 10 05:14:23 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:23 90 [Note] Shutting down plugin 'binlog'
Jul 10 05:14:23 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:23 90 [Note] mysqld: Shutdown complete
Jul 10 05:14:23 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]:
Jul 10 05:14:23 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:23+00:00 [Note] [Entrypoint]: Temporary server stopped
Jul 10 05:14:23 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]:
Jul 10 05:14:23 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:23+00:00 [Note] [Entrypoint]: MySQL init process done. Ready for start up.
Jul 10 05:14:23 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]:
Jul 10 05:14:24 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:24 0 [Warning] TIMESTAMP with implicit DEFAULT value is deprecated. Please use --explicit_defaults_for_timestamp server option (see documentation for more details).
Jul 10 05:14:24 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:24 0 [Note] mysqld (mysqld 5.6.51) starting as process 1 ...
Jul 10 05:14:24 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:24 1 [Note] Plugin 'FEDERATED' is disabled.
Jul 10 05:14:24 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:24 1 [Note] InnoDB: Using atomics to ref count buffer pool pages
Jul 10 05:14:24 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:24 1 [Note] InnoDB: The InnoDB memory heap is disabled
Jul 10 05:14:24 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:24 1 [Note] InnoDB: Mutexes and rw_locks use GCC atomic builtins
Jul 10 05:14:24 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:24 1 [Note] InnoDB: Memory barrier is not used
Jul 10 05:14:24 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:24 1 [Note] InnoDB: Compressed tables use zlib 1.2.11
Jul 10 05:14:24 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:24 1 [Note] InnoDB: Using Linux native AIO
Jul 10 05:14:24 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:24 1 [Note] InnoDB: Using CPU crc32 instructions
Jul 10 05:14:24 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:24 1 [Note] InnoDB: Initializing buffer pool, size = 128.0M
Jul 10 05:14:24 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:24 1 [Note] InnoDB: Completed initialization of buffer pool
Jul 10 05:14:24 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:24 1 [Note] InnoDB: Highest supported file format is Barracuda.
Jul 10 05:14:24 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:24 1 [Note] InnoDB: 128 rollback segment(s) are active.
Jul 10 05:14:24 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:24 1 [Note] InnoDB: Waiting for purge to start
Jul 10 05:14:24 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:24 1 [Note] InnoDB: 5.6.51 started; log sequence number 1625997
Jul 10 05:14:24 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:24 1 [Note] RSA private key file not found: /var/lib/mysql//private_key.pem. Some authentication plugins will not work.
Jul 10 05:14:24 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:24 1 [Note] RSA public key file not found: /var/lib/mysql//public_key.pem. Some authentication plugins will not work.
Jul 10 05:14:24 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:24 1 [Note] Server hostname (bind-address): '*'; port: 3306
Jul 10 05:14:24 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:24 1 [Note] IPv6 is available.
Jul 10 05:14:24 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:24 1 [Note] - '::' resolves to '::';
Jul 10 05:14:24 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:24 1 [Note] Server socket created on IP: '::'.
Jul 10 05:14:24 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:24 1 [Warning] Insecure configuration for --pid-file: Location '/var/run/mysqld' in the path is accessible to all OS users. Consider choosing a different directory.
Jul 10 05:14:24 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:24 1 [Warning] 'proxies_priv' entry '@ root@21b1326f5785' ignored in --skip-name-resolve mode.
Jul 10 05:14:24 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:24 1 [Note] Event Scheduler: Loaded 0 events
Jul 10 05:14:24 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:24 1 [Note] mysqld: ready for connections.
Jul 10 05:14:24 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: Version: '5.6.51' socket: '/var/run/mysqld/mysqld.sock' port: 3306 MySQL Community Server (GPL)
Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm podman[11232]: 2024-07-10 05:15:14.395138604 +0000 UTC m=+0.362193900 container remove 21b1326f57855b147efcb07cbd10c5f7b7edbabb6693b7752105182196a3f742 (image=quay.io/linux-system-roles/mysql:5.6, name=quadlet-demo-mysql, PODMAN_SYSTEMD_UNIT=quadlet-demo-mysql.service)
Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: quadlet-demo-mysql.service: Main process exited, code=exited, status=137/n/a
Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: quadlet-demo-mysql.service: Failed with result 'exit-code'.
Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: quadlet-demo-mysql.service: Consumed 3.750s CPU time.
+ systemctl status quadlet-demo-mysql.service
× quadlet-demo-mysql.service
Loaded: not-found (Reason: Unit quadlet-demo-mysql.service not found.)
Active: failed (Result: exit-code) since Wed 2024-07-10 05:15:14 UTC; 26s ago
Duration: 1min 1.221s
Main PID: 8040 (code=exited, status=137)
CPU: 3.750s
Jul 10 05:14:24 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:24 1 [Note] Server socket created on IP: '::'.
Jul 10 05:14:24 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:24 1 [Warning] Insecure configuration for --pid-file: Location '/var/run/mysqld' in the path is accessible to all OS users. Consider choosing a different directory.
Jul 10 05:14:24 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:24 1 [Warning] 'proxies_priv' entry '@ root@21b1326f5785' ignored in --skip-name-resolve mode.
Jul 10 05:14:24 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:24 1 [Note] Event Scheduler: Loaded 0 events
Jul 10 05:14:24 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: 2024-07-10 05:14:24 1 [Note] mysqld: ready for connections.
Jul 10 05:14:24 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo-mysql[8040]: Version: '5.6.51' socket: '/var/run/mysqld/mysqld.sock' port: 3306 MySQL Community Server (GPL)
Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm podman[11232]: 2024-07-10 05:15:14.395138604 +0000 UTC m=+0.362193900 container remove 21b1326f57855b147efcb07cbd10c5f7b7edbabb6693b7752105182196a3f742 (image=quay.io/linux-system-roles/mysql:5.6, name=quadlet-demo-mysql, PODMAN_SYSTEMD_UNIT=quadlet-demo-mysql.service)
Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: quadlet-demo-mysql.service: Main process exited, code=exited, status=137/n/a
Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: quadlet-demo-mysql.service: Failed with result 'exit-code'.
Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: quadlet-demo-mysql.service: Consumed 3.750s CPU time.
+ :
TASK [Get journald] ************************************************************
task path: /WORKDIR/git-handle-quadlet-cleanup35__kigx/tests/tests_quadlet_demo.yml:193
Wednesday 10 July 2024 05:15:40 +0000 (0:00:00.828) 0:03:45.197 ********
fatal: [sut]: FAILED! => {
"changed": false,
"cmd": [
"journalctl",
"-ex"
],
"delta": "0:00:00.036562",
"end": "2024-07-10 05:15:40.762072",
"failed_when_result": true,
"rc": 0,
"start": "2024-07-10 05:15:40.725510"
}
STDOUT:
Jul 10 05:14:54 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:14:54 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:14:54 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:14:54 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=10021 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success'
Jul 10 05:14:54 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:14:54 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:14:55 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:14:55 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:14:55 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=10045 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success'
Jul 10 05:14:55 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:14:55 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:14:55 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:14:55 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:14:55 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=10069 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success'
Jul 10 05:14:55 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:14:55 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:14:56 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:14:56 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:14:56 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=10093 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success'
Jul 10 05:14:56 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[10109]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None
Jul 10 05:14:56 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:14:56 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:14:56 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:14:56 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:14:56 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=10111 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success'
Jul 10 05:14:56 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[10127]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1
Jul 10 05:14:56 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:14:56 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=10130 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success'
Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[10146]: ansible-systemd Invoked with name=quadlet-demo.service scope=system state=stopped enabled=False force=True daemon_reload=False daemon_reexec=False no_block=False masked=None
Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: Reloading requested from client PID 10149 ('systemctl') (unit session-6.scope)...
Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: Reloading...
Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: /run/systemd/transient/libpod-dce8c2f6a9a9de70b22786996e3a7fd320907983dffeb24910cd53361f6a0144.scope:14: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon.
Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: /run/systemd/transient/libpod-af5eaff28e9fad6d7bd929af0607de207b40ae227252858bca105cbe91dcf058.scope:14: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon.
Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: Reloading finished in 407 ms.
Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=315 op=LOAD
Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=292 op=UNLOAD
Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=316 op=LOAD
Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=293 op=UNLOAD
Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=317 op=LOAD
Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=318 op=LOAD
Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=294 op=UNLOAD
Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=295 op=UNLOAD
Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=319 op=LOAD
Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=296 op=UNLOAD
Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=320 op=LOAD
Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=297 op=UNLOAD
Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=321 op=LOAD
Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=322 op=LOAD
Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=298 op=UNLOAD
Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=299 op=UNLOAD
Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=323 op=LOAD
Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=300 op=UNLOAD
Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=324 op=LOAD
Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=325 op=LOAD
Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=301 op=UNLOAD
Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=302 op=UNLOAD
Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=326 op=LOAD
Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=303 op=UNLOAD
Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=327 op=LOAD
Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=304 op=UNLOAD
Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=328 op=LOAD
Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=305 op=UNLOAD
Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=329 op=LOAD
Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=330 op=LOAD
Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=306 op=UNLOAD
Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=307 op=UNLOAD
Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=331 op=LOAD
Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=308 op=UNLOAD
Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=332 op=LOAD
Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=333 op=LOAD
Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=309 op=UNLOAD
Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=310 op=UNLOAD
Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: Stopping quadlet-demo.service...
░░ Subject: A stop job for unit quadlet-demo.service has begun execution
░░ Defined-By: systemd
░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel
░░
░░ A stop job for unit quadlet-demo.service has begun execution.
░░
░░ The job identifier is 3523.
Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: libpod-9de2222e9853a1d7c250e11a1e4975a817569c5e9b9c202c5e84f41b7332711a.scope: Deactivated successfully.
░░ Subject: Unit succeeded
░░ Defined-By: systemd
░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel
░░
░░ The unit libpod-9de2222e9853a1d7c250e11a1e4975a817569c5e9b9c202c5e84f41b7332711a.scope has successfully entered the 'dead' state.
Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=311 op=UNLOAD
Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm podman[10200]: 2024-07-10 05:14:57.965328709 +0000 UTC m=+0.031740190 container died 9de2222e9853a1d7c250e11a1e4975a817569c5e9b9c202c5e84f41b7332711a (image=localhost/podman-pause:4.9.4-1711445992, name=a96f3a51b8d1-service, PODMAN_SYSTEMD_UNIT=quadlet-demo.service)
Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: var-lib-containers-storage-overlay-2b7d88ad00bd6e4b28025299b957f2c60bd7968d96b676e2b876667ee5d7bc54-merged.mount: Deactivated successfully.
░░ Subject: Unit succeeded
░░ Defined-By: systemd
░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel
░░
░░ The unit var-lib-containers-storage-overlay-2b7d88ad00bd6e4b28025299b957f2c60bd7968d96b676e2b876667ee5d7bc54-merged.mount has successfully entered the 'dead' state.
Jul 10 05:14:57 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-9de2222e9853a1d7c250e11a1e4975a817569c5e9b9c202c5e84f41b7332711a-userdata-shm.mount: Deactivated successfully.
░░ Subject: Unit succeeded
░░ Defined-By: systemd
░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel
░░
░░ The unit var-lib-containers-storage-overlay\x2dcontainers-9de2222e9853a1d7c250e11a1e4975a817569c5e9b9c202c5e84f41b7332711a-userdata-shm.mount has successfully entered the 'dead' state.
Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm podman[10200]: 2024-07-10 05:14:58.00693557 +0000 UTC m=+0.073347244 container cleanup 9de2222e9853a1d7c250e11a1e4975a817569c5e9b9c202c5e84f41b7332711a (image=localhost/podman-pause:4.9.4-1711445992, name=a96f3a51b8d1-service, PODMAN_SYSTEMD_UNIT=quadlet-demo.service)
Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm podman[10208]: 2024-07-10 05:14:58.050049719 +0000 UTC m=+0.028193783 pod stop da66443b0d18df866aa4578267e9465681155ec0f758ec8d84214580c2123234 (image=, name=quadlet-demo)
Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: libpod-ee4bc05257fc811692ed315d4bc9e3e8815ce0788437369ad146bb547ed5565f.scope: Deactivated successfully.
░░ Subject: Unit succeeded
░░ Defined-By: systemd
░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel
░░
░░ The unit libpod-ee4bc05257fc811692ed315d4bc9e3e8815ce0788437369ad146bb547ed5565f.scope has successfully entered the 'dead' state.
Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=312 op=UNLOAD
Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm podman[10208]: 2024-07-10 05:14:58.081849733 +0000 UTC m=+0.059993869 container died ee4bc05257fc811692ed315d4bc9e3e8815ce0788437369ad146bb547ed5565f (image=localhost/podman-pause:4.9.4-1711445992, name=da66443b0d18-infra, PODMAN_SYSTEMD_UNIT=quadlet-demo.service, io.buildah.version=1.33.7)
Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: libpod-dce8c2f6a9a9de70b22786996e3a7fd320907983dffeb24910cd53361f6a0144.scope: Deactivated successfully.
░░ Subject: Unit succeeded
░░ Defined-By: systemd
░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel
░░
░░ The unit libpod-dce8c2f6a9a9de70b22786996e3a7fd320907983dffeb24910cd53361f6a0144.scope has successfully entered the 'dead' state.
Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=314 op=UNLOAD
Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm podman[10208]: 2024-07-10 05:14:58.113820223 +0000 UTC m=+0.091964394 container died dce8c2f6a9a9de70b22786996e3a7fd320907983dffeb24910cd53361f6a0144 (image=quay.io/linux-system-roles/envoyproxy:v1.25.0, name=quadlet-demo-envoy, PODMAN_SYSTEMD_UNIT=quadlet-demo.service)
Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: libpod-af5eaff28e9fad6d7bd929af0607de207b40ae227252858bca105cbe91dcf058.scope: Deactivated successfully.
░░ Subject: Unit succeeded
░░ Defined-By: systemd
░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel
░░
░░ The unit libpod-af5eaff28e9fad6d7bd929af0607de207b40ae227252858bca105cbe91dcf058.scope has successfully entered the 'dead' state.
Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: libpod-af5eaff28e9fad6d7bd929af0607de207b40ae227252858bca105cbe91dcf058.scope: Consumed 1.176s CPU time.
░░ Subject: Resources consumed by unit runtime
░░ Defined-By: systemd
░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel
░░
░░ The unit libpod-af5eaff28e9fad6d7bd929af0607de207b40ae227252858bca105cbe91dcf058.scope completed and consumed the indicated resources.
Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=313 op=UNLOAD
Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm kernel: podman1: port 2(veth1) entered disabled state
Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm kernel: veth1 (unregistering): left allmulticast mode
Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm kernel: veth1 (unregistering): left promiscuous mode
Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm kernel: podman1: port 2(veth1) entered disabled state
Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: ANOM_PROMISCUOUS dev=veth1 prom=0 old_prom=256 auid=4294967295 uid=0 gid=0 ses=4294967295
Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm podman[10208]: 2024-07-10 05:14:58.144771474 +0000 UTC m=+0.122915565 container died af5eaff28e9fad6d7bd929af0607de207b40ae227252858bca105cbe91dcf058 (image=quay.io/linux-system-roles/wordpress:4.8-apache, name=quadlet-demo-wordpress, PODMAN_SYSTEMD_UNIT=quadlet-demo.service)
Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[10247]: NETFILTER_CFG table=nat:42 family=2 entries=1 op=nft_unregister_rule pid=10247 subj=system_u:system_r:iptables_t:s0 comm="iptables"
Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[10250]: NETFILTER_CFG table=nat:43 family=2 entries=1 op=nft_unregister_rule pid=10250 subj=system_u:system_r:iptables_t:s0 comm="iptables"
Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm podman[10208]: 2024-07-10 05:14:58.218821937 +0000 UTC m=+0.196965972 container cleanup dce8c2f6a9a9de70b22786996e3a7fd320907983dffeb24910cd53361f6a0144 (image=quay.io/linux-system-roles/envoyproxy:v1.25.0, name=quadlet-demo-envoy, pod_id=da66443b0d18df866aa4578267e9465681155ec0f758ec8d84214580c2123234, PODMAN_SYSTEMD_UNIT=quadlet-demo.service)
Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[10254]: NETFILTER_CFG table=nat:44 family=2 entries=1 op=nft_unregister_rule pid=10254 subj=system_u:system_r:iptables_t:s0 comm="iptables"
Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm podman[10208]: 2024-07-10 05:14:58.224499975 +0000 UTC m=+0.202643952 container cleanup af5eaff28e9fad6d7bd929af0607de207b40ae227252858bca105cbe91dcf058 (image=quay.io/linux-system-roles/wordpress:4.8-apache, name=quadlet-demo-wordpress, pod_id=da66443b0d18df866aa4578267e9465681155ec0f758ec8d84214580c2123234, PODMAN_SYSTEMD_UNIT=quadlet-demo.service)
Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[10256]: NETFILTER_CFG table=nat:45 family=2 entries=1 op=nft_unregister_rule pid=10256 subj=system_u:system_r:iptables_t:s0 comm="iptables"
Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[10259]: NETFILTER_CFG table=nat:46 family=2 entries=1 op=nft_unregister_rule pid=10259 subj=system_u:system_r:iptables_t:s0 comm="iptables"
Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[10263]: NETFILTER_CFG table=nat:47 family=2 entries=1 op=nft_unregister_rule pid=10263 subj=system_u:system_r:iptables_t:s0 comm="iptables"
Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[10265]: NETFILTER_CFG table=nat:48 family=2 entries=1 op=nft_unregister_rule pid=10265 subj=system_u:system_r:iptables_t:s0 comm="iptables"
Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[10267]: NETFILTER_CFG table=nat:49 family=2 entries=1 op=nft_unregister_rule pid=10267 subj=system_u:system_r:iptables_t:s0 comm="iptables"
Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm podman[10208]: 2024-07-10 05:14:58.300925855 +0000 UTC m=+0.279069830 container cleanup ee4bc05257fc811692ed315d4bc9e3e8815ce0788437369ad146bb547ed5565f (image=localhost/podman-pause:4.9.4-1711445992, name=da66443b0d18-infra, pod_id=da66443b0d18df866aa4578267e9465681155ec0f758ec8d84214580c2123234, io.buildah.version=1.33.7, PODMAN_SYSTEMD_UNIT=quadlet-demo.service)
Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: Removed slice machine-libpod_pod_da66443b0d18df866aa4578267e9465681155ec0f758ec8d84214580c2123234.slice - cgroup machine-libpod_pod_da66443b0d18df866aa4578267e9465681155ec0f758ec8d84214580c2123234.slice.
░░ Subject: A stop job for unit machine-libpod_pod_da66443b0d18df866aa4578267e9465681155ec0f758ec8d84214580c2123234.slice has finished
░░ Defined-By: systemd
░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel
░░
░░ A stop job for unit machine-libpod_pod_da66443b0d18df866aa4578267e9465681155ec0f758ec8d84214580c2123234.slice has finished.
░░
░░ The job identifier is 3525 and the job result is done.
Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: machine-libpod_pod_da66443b0d18df866aa4578267e9465681155ec0f758ec8d84214580c2123234.slice: Consumed 1.276s CPU time.
░░ Subject: Resources consumed by unit runtime
░░ Defined-By: systemd
░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel
░░
░░ The unit machine-libpod_pod_da66443b0d18df866aa4578267e9465681155ec0f758ec8d84214580c2123234.slice completed and consumed the indicated resources.
Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm podman[10208]: 2024-07-10 05:14:58.311544683 +0000 UTC m=+0.289688767 pod stop da66443b0d18df866aa4578267e9465681155ec0f758ec8d84214580c2123234 (image=, name=quadlet-demo)
Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: machine-libpod_pod_da66443b0d18df866aa4578267e9465681155ec0f758ec8d84214580c2123234.slice: Failed to open /run/systemd/transient/machine-libpod_pod_da66443b0d18df866aa4578267e9465681155ec0f758ec8d84214580c2123234.slice: No such file or directory
Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo[10208]: time="2024-07-10T05:14:58Z" level=error msg="Checking if infra needs to be stopped: removing pod da66443b0d18df866aa4578267e9465681155ec0f758ec8d84214580c2123234 cgroup: Unit machine-libpod_pod_da66443b0d18df866aa4578267e9465681155ec0f758ec8d84214580c2123234.slice not loaded."
Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm podman[10208]: 2024-07-10 05:14:58.381480436 +0000 UTC m=+0.359624442 container remove dce8c2f6a9a9de70b22786996e3a7fd320907983dffeb24910cd53361f6a0144 (image=quay.io/linux-system-roles/envoyproxy:v1.25.0, name=quadlet-demo-envoy, pod_id=da66443b0d18df866aa4578267e9465681155ec0f758ec8d84214580c2123234, PODMAN_SYSTEMD_UNIT=quadlet-demo.service)
Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm podman[10208]: 2024-07-10 05:14:58.400712486 +0000 UTC m=+0.378856510 container remove af5eaff28e9fad6d7bd929af0607de207b40ae227252858bca105cbe91dcf058 (image=quay.io/linux-system-roles/wordpress:4.8-apache, name=quadlet-demo-wordpress, pod_id=da66443b0d18df866aa4578267e9465681155ec0f758ec8d84214580c2123234, PODMAN_SYSTEMD_UNIT=quadlet-demo.service)
Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm podman[10208]: 2024-07-10 05:14:58.42903182 +0000 UTC m=+0.407175829 container remove ee4bc05257fc811692ed315d4bc9e3e8815ce0788437369ad146bb547ed5565f (image=localhost/podman-pause:4.9.4-1711445992, name=da66443b0d18-infra, pod_id=da66443b0d18df866aa4578267e9465681155ec0f758ec8d84214580c2123234, PODMAN_SYSTEMD_UNIT=quadlet-demo.service, io.buildah.version=1.33.7)
Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: machine-libpod_pod_da66443b0d18df866aa4578267e9465681155ec0f758ec8d84214580c2123234.slice: Failed to open /run/systemd/transient/machine-libpod_pod_da66443b0d18df866aa4578267e9465681155ec0f758ec8d84214580c2123234.slice: No such file or directory
Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm podman[10208]: 2024-07-10 05:14:58.437391183 +0000 UTC m=+0.415535161 pod remove da66443b0d18df866aa4578267e9465681155ec0f758ec8d84214580c2123234 (image=, name=quadlet-demo)
Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm podman[10208]: 2024-07-10 05:14:58.456671982 +0000 UTC m=+0.434815999 container remove 9de2222e9853a1d7c250e11a1e4975a817569c5e9b9c202c5e84f41b7332711a (image=localhost/podman-pause:4.9.4-1711445992, name=a96f3a51b8d1-service, PODMAN_SYSTEMD_UNIT=quadlet-demo.service)
Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo[10208]: Pods stopped:
Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo[10208]: da66443b0d18df866aa4578267e9465681155ec0f758ec8d84214580c2123234
Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo[10208]: Pods removed:
Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo[10208]: Error: removing pod da66443b0d18df866aa4578267e9465681155ec0f758ec8d84214580c2123234 cgroup: removing pod da66443b0d18df866aa4578267e9465681155ec0f758ec8d84214580c2123234 cgroup: Unit machine-libpod_pod_da66443b0d18df866aa4578267e9465681155ec0f758ec8d84214580c2123234.slice not loaded.
Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo[10208]: Secrets removed:
Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo[10208]: Error: %!s()
Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo[10208]: Volumes removed:
Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo[10208]: time="2024-07-10T05:14:58Z" level=error msg="Checking whether service of container 9de2222e9853a1d7c250e11a1e4975a817569c5e9b9c202c5e84f41b7332711a can be stopped: no container with ID 9de2222e9853a1d7c250e11a1e4975a817569c5e9b9c202c5e84f41b7332711a found in database: no such container"
Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm quadlet-demo[10208]: time="2024-07-10T05:14:58Z" level=error msg="Checking whether service of container 9de2222e9853a1d7c250e11a1e4975a817569c5e9b9c202c5e84f41b7332711a can be stopped: no container with ID 9de2222e9853a1d7c250e11a1e4975a817569c5e9b9c202c5e84f41b7332711a found in database: no such container"
Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: quadlet-demo.service: Deactivated successfully.
░░ Subject: Unit succeeded
░░ Defined-By: systemd
░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel
░░
░░ The unit quadlet-demo.service has successfully entered the 'dead' state.
Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: Stopped quadlet-demo.service.
░░ Subject: A stop job for unit quadlet-demo.service has finished
░░ Defined-By: systemd
░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel
░░
░░ A stop job for unit quadlet-demo.service has finished.
░░
░░ The job identifier is 3523 and the job result is done.
Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:init_t:s0 msg='unit=quadlet-demo comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success'
Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=10271 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success'
Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[10287]: ansible-file Invoked with path=/etc/containers/systemd/quadlet-demo.kube state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None
Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=10288 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success'
Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: tmp-crun.HDFMzq.mount: Deactivated successfully.
░░ Subject: Unit succeeded
░░ Defined-By: systemd
░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel
░░
░░ The unit tmp-crun.HDFMzq.mount has successfully entered the 'dead' state.
Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: var-lib-containers-storage-overlay-3016041ec40d0362707a4f866fdaa64e34e3c1e1999e0d6d8bd640ecbb918ce8-merged.mount: Deactivated successfully.
░░ Subject: Unit succeeded
░░ Defined-By: systemd
░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel
░░
░░ The unit var-lib-containers-storage-overlay-3016041ec40d0362707a4f866fdaa64e34e3c1e1999e0d6d8bd640ecbb918ce8-merged.mount has successfully entered the 'dead' state.
Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: var-lib-containers-storage-overlay-859ab31e620e0a4681f0f8f3ef5d32c5fe4e820f6d74606719739fdbfd66173e-merged.mount: Deactivated successfully.
░░ Subject: Unit succeeded
░░ Defined-By: systemd
░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel
░░
░░ The unit var-lib-containers-storage-overlay-859ab31e620e0a4681f0f8f3ef5d32c5fe4e820f6d74606719739fdbfd66173e-merged.mount has successfully entered the 'dead' state.
Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: var-lib-containers-storage-overlay-bac024e2d3f38a9994be6ed0f1964c96696ceaf11d0399758a993d5026385fae-merged.mount: Deactivated successfully.
░░ Subject: Unit succeeded
░░ Defined-By: systemd
░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel
░░
░░ The unit var-lib-containers-storage-overlay-bac024e2d3f38a9994be6ed0f1964c96696ceaf11d0399758a993d5026385fae-merged.mount has successfully entered the 'dead' state.
Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-ee4bc05257fc811692ed315d4bc9e3e8815ce0788437369ad146bb547ed5565f-userdata-shm.mount: Deactivated successfully.
░░ Subject: Unit succeeded
░░ Defined-By: systemd
░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel
░░
░░ The unit var-lib-containers-storage-overlay\x2dcontainers-ee4bc05257fc811692ed315d4bc9e3e8815ce0788437369ad146bb547ed5565f-userdata-shm.mount has successfully entered the 'dead' state.
Jul 10 05:14:58 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: run-netns-netns\x2dc9c42157\x2d2a14\x2df2d5\x2d22c0\x2d3e88aa7c59f3.mount: Deactivated successfully.
░░ Subject: Unit succeeded
░░ Defined-By: systemd
░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel
░░
░░ The unit run-netns-netns\x2dc9c42157\x2d2a14\x2df2d5\x2d22c0\x2d3e88aa7c59f3.mount has successfully entered the 'dead' state.
Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[10304]: ansible-systemd Invoked with daemon_reload=True scope=system daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None
Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: Reloading requested from client PID 10305 ('systemctl') (unit session-6.scope)...
Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: Reloading...
Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: Reloading finished in 403 ms.
Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=334 op=LOAD
Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=315 op=UNLOAD
Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=335 op=LOAD
Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=316 op=UNLOAD
Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=336 op=LOAD
Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=337 op=LOAD
Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=317 op=UNLOAD
Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=318 op=UNLOAD
Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=338 op=LOAD
Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=319 op=UNLOAD
Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=339 op=LOAD
Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=320 op=UNLOAD
Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=340 op=LOAD
Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=341 op=LOAD
Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=321 op=UNLOAD
Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=322 op=UNLOAD
Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=342 op=LOAD
Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=323 op=UNLOAD
Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=343 op=LOAD
Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=344 op=LOAD
Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=324 op=UNLOAD
Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=325 op=UNLOAD
Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=345 op=LOAD
Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=326 op=UNLOAD
Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=346 op=LOAD
Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=327 op=UNLOAD
Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=347 op=LOAD
Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=328 op=UNLOAD
Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=348 op=LOAD
Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=349 op=LOAD
Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=329 op=UNLOAD
Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=330 op=UNLOAD
Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=350 op=LOAD
Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=331 op=UNLOAD
Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=351 op=LOAD
Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=352 op=LOAD
Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=332 op=UNLOAD
Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=333 op=UNLOAD
Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=10355 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success'
Jul 10 05:14:59 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[10371]: ansible-ansible.legacy.command Invoked with _raw_params=podman image prune --all -f _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Jul 10 05:15:00 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm podman[10372]: 2024-07-10 05:15:00.298256034 +0000 UTC m=+0.411657053 image untag fcf3e41b8864a14d75a6d0627d3d02154e28a153aa57e8baa392cd744ffa0d0b quay.io/linux-system-roles/wordpress:4.8-apache
Jul 10 05:15:00 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm podman[10372]: 2024-07-10 05:14:59.912089645 +0000 UTC m=+0.025490984 image remove fcf3e41b8864a14d75a6d0627d3d02154e28a153aa57e8baa392cd744ffa0d0b
Jul 10 05:15:00 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm podman[10372]: 2024-07-10 05:15:00.410901119 +0000 UTC m=+0.524302167 image untag 5af2585e22ed1562885d9407efab74010090427be79048c2cd6a226517cc1e1d quay.io/linux-system-roles/envoyproxy:v1.25.0
Jul 10 05:15:00 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm podman[10372]: 2024-07-10 05:15:00.29826683 +0000 UTC m=+0.411667812 image remove 5af2585e22ed1562885d9407efab74010090427be79048c2cd6a226517cc1e1d
Jul 10 05:15:00 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm podman[10372]: 2024-07-10 05:15:00.417719687 +0000 UTC m=+0.531120696 image untag 6c68feb81ac7fb684f7b4d483de9bffaff110fa95445c248359b77bb5d3b2295 localhost/podman-pause:4.9.4-1711445992
Jul 10 05:15:00 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm podman[10372]: 2024-07-10 05:15:00.410910715 +0000 UTC m=+0.524311688 image remove 6c68feb81ac7fb684f7b4d483de9bffaff110fa95445c248359b77bb5d3b2295
Jul 10 05:15:00 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:00 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:00 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:00 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:00 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=10380 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success'
Jul 10 05:15:00 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[10396]: ansible-ansible.legacy.command Invoked with _raw_params=podman images -n _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Jul 10 05:15:00 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:00 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:00 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:00 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:00 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=10404 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success'
Jul 10 05:15:00 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[10420]: ansible-ansible.legacy.command Invoked with _raw_params=podman volume ls -n _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Jul 10 05:15:01 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:01 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:01 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:01 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:01 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=10428 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success'
Jul 10 05:15:01 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[10444]: ansible-ansible.legacy.command Invoked with _raw_params=podman ps --noheading _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Jul 10 05:15:01 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:01 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:01 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:01 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:01 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=10453 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success'
Jul 10 05:15:01 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[10469]: ansible-ansible.legacy.command Invoked with _raw_params=podman network ls -n -q _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Jul 10 05:15:01 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:01 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:01 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:01 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:01 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=10477 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success'
Jul 10 05:15:01 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:01 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:01 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:01 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:01 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=10501 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success'
Jul 10 05:15:02 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[10517]: ansible-service_facts Invoked
Jul 10 05:15:04 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:04 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:04 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:04 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:04 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=10630 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success'
Jul 10 05:15:04 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[10646]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None
Jul 10 05:15:04 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:04 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:04 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:04 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:04 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=10648 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success'
Jul 10 05:15:05 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[10664]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1
Jul 10 05:15:05 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:05 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:05 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:05 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:05 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=10667 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success'
Jul 10 05:15:05 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[10683]: ansible-file Invoked with path=/etc/containers/systemd/quadlet-demo.yml state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None
Jul 10 05:15:05 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:05 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:05 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:05 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:05 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=10684 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success'
Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[10700]: ansible-systemd Invoked with daemon_reload=True scope=system daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None
Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: Reloading requested from client PID 10701 ('systemctl') (unit session-6.scope)...
Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: Reloading...
Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: Reloading finished in 387 ms.
Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=353 op=LOAD
Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=334 op=UNLOAD
Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=354 op=LOAD
Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=335 op=UNLOAD
Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=355 op=LOAD
Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=356 op=LOAD
Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=336 op=UNLOAD
Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=337 op=UNLOAD
Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=357 op=LOAD
Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=338 op=UNLOAD
Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=358 op=LOAD
Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=339 op=UNLOAD
Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=359 op=LOAD
Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=360 op=LOAD
Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=340 op=UNLOAD
Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=341 op=UNLOAD
Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=361 op=LOAD
Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=342 op=UNLOAD
Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=362 op=LOAD
Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=363 op=LOAD
Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=343 op=UNLOAD
Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=344 op=UNLOAD
Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=364 op=LOAD
Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=345 op=UNLOAD
Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=365 op=LOAD
Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=346 op=UNLOAD
Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=366 op=LOAD
Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=347 op=UNLOAD
Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=367 op=LOAD
Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=368 op=LOAD
Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=348 op=UNLOAD
Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=349 op=UNLOAD
Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=369 op=LOAD
Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=350 op=UNLOAD
Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=370 op=LOAD
Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=371 op=LOAD
Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=351 op=UNLOAD
Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=352 op=UNLOAD
Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=10751 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success'
Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[10767]: ansible-ansible.legacy.command Invoked with _raw_params=podman image prune --all -f _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:06 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:07 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:07 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:07 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=10775 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success'
Jul 10 05:15:07 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[10791]: ansible-ansible.legacy.command Invoked with _raw_params=podman images -n _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Jul 10 05:15:07 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:07 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:07 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:07 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:07 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=10800 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success'
Jul 10 05:15:07 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[10816]: ansible-ansible.legacy.command Invoked with _raw_params=podman volume ls -n _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Jul 10 05:15:07 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:07 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:07 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:07 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:07 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=10824 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success'
Jul 10 05:15:07 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[10840]: ansible-ansible.legacy.command Invoked with _raw_params=podman ps --noheading _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Jul 10 05:15:07 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:07 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:07 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:07 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:07 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=10848 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success'
Jul 10 05:15:08 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[10864]: ansible-ansible.legacy.command Invoked with _raw_params=podman network ls -n -q _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Jul 10 05:15:08 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:08 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:08 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:08 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:08 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=10872 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success'
Jul 10 05:15:08 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:08 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:08 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:08 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:08 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=10896 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success'
Jul 10 05:15:08 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[10912]: ansible-service_facts Invoked
Jul 10 05:15:10 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:10 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:10 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:10 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:10 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=11025 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success'
Jul 10 05:15:11 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[11041]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None
Jul 10 05:15:11 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:11 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:11 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:11 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:11 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=11043 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success'
Jul 10 05:15:11 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[11059]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1
Jul 10 05:15:11 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:11 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:11 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:11 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:11 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=11062 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success'
Jul 10 05:15:12 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[11078]: ansible-file Invoked with path=/etc/containers/systemd/envoy-proxy-configmap.yml state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None
Jul 10 05:15:12 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:12 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:12 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:12 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:12 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=11079 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success'
Jul 10 05:15:12 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[11095]: ansible-systemd Invoked with daemon_reload=True scope=system daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None
Jul 10 05:15:12 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: Reloading requested from client PID 11096 ('systemctl') (unit session-6.scope)...
Jul 10 05:15:12 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: Reloading...
Jul 10 05:15:12 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: Reloading finished in 384 ms.
Jul 10 05:15:12 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=372 op=LOAD
Jul 10 05:15:12 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=353 op=UNLOAD
Jul 10 05:15:12 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=373 op=LOAD
Jul 10 05:15:12 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=354 op=UNLOAD
Jul 10 05:15:12 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=374 op=LOAD
Jul 10 05:15:12 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=375 op=LOAD
Jul 10 05:15:12 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=355 op=UNLOAD
Jul 10 05:15:12 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=356 op=UNLOAD
Jul 10 05:15:12 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=376 op=LOAD
Jul 10 05:15:12 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=357 op=UNLOAD
Jul 10 05:15:12 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=377 op=LOAD
Jul 10 05:15:12 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=358 op=UNLOAD
Jul 10 05:15:12 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=378 op=LOAD
Jul 10 05:15:12 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=379 op=LOAD
Jul 10 05:15:12 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=359 op=UNLOAD
Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=360 op=UNLOAD
Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=380 op=LOAD
Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=361 op=UNLOAD
Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=381 op=LOAD
Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=382 op=LOAD
Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=362 op=UNLOAD
Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=363 op=UNLOAD
Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=383 op=LOAD
Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=364 op=UNLOAD
Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=384 op=LOAD
Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=365 op=UNLOAD
Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=385 op=LOAD
Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=366 op=UNLOAD
Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=386 op=LOAD
Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=387 op=LOAD
Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=367 op=UNLOAD
Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=368 op=UNLOAD
Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=388 op=LOAD
Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=369 op=UNLOAD
Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=389 op=LOAD
Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=390 op=LOAD
Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=370 op=UNLOAD
Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=371 op=UNLOAD
Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=11146 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success'
Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[11162]: ansible-ansible.legacy.command Invoked with _raw_params=podman image prune --all -f _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=11171 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success'
Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[11187]: ansible-ansible.legacy.command Invoked with _raw_params=podman images -n _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=11195 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success'
Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[11211]: ansible-ansible.legacy.command Invoked with _raw_params=podman volume ls -n _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:init_t:s0 msg='unit=21b1326f57855b147efcb07cbd10c5f7b7edbabb6693b7752105182196a3f742 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success'
Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:13 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm podman[11212]: 2024-07-10 05:15:13.983498841 +0000 UTC m=+0.068110829 container health_status 21b1326f57855b147efcb07cbd10c5f7b7edbabb6693b7752105182196a3f742 (image=quay.io/linux-system-roles/mysql:5.6, name=quadlet-demo-mysql, health_status=unhealthy, PODMAN_SYSTEMD_UNIT=quadlet-demo-mysql.service)
Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=11228 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success'
Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: 21b1326f57855b147efcb07cbd10c5f7b7edbabb6693b7752105182196a3f742.service: Main process exited, code=exited, status=125/n/a
░░ Subject: Unit process exited
░░ Defined-By: systemd
░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel
░░
░░ An ExecStart= process belonging to unit 21b1326f57855b147efcb07cbd10c5f7b7edbabb6693b7752105182196a3f742.service has exited.
░░
░░ The process' exit code is 'exited' and its exit status is 125.
Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: 21b1326f57855b147efcb07cbd10c5f7b7edbabb6693b7752105182196a3f742.service: Failed with result 'exit-code'.
░░ Subject: Unit failed
░░ Defined-By: systemd
░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel
░░
░░ The unit 21b1326f57855b147efcb07cbd10c5f7b7edbabb6693b7752105182196a3f742.service has entered the 'failed' state with result 'exit-code'.
Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:init_t:s0 msg='unit=21b1326f57855b147efcb07cbd10c5f7b7edbabb6693b7752105182196a3f742 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed'
Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: 21b1326f57855b147efcb07cbd10c5f7b7edbabb6693b7752105182196a3f742.timer: Deactivated successfully.
░░ Subject: Unit succeeded
░░ Defined-By: systemd
░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel
░░
░░ The unit 21b1326f57855b147efcb07cbd10c5f7b7edbabb6693b7752105182196a3f742.timer has successfully entered the 'dead' state.
Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: Stopped 21b1326f57855b147efcb07cbd10c5f7b7edbabb6693b7752105182196a3f742.timer - /usr/bin/podman healthcheck run 21b1326f57855b147efcb07cbd10c5f7b7edbabb6693b7752105182196a3f742.
░░ Subject: A stop job for unit 21b1326f57855b147efcb07cbd10c5f7b7edbabb6693b7752105182196a3f742.timer has finished
░░ Defined-By: systemd
░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel
░░
░░ A stop job for unit 21b1326f57855b147efcb07cbd10c5f7b7edbabb6693b7752105182196a3f742.timer has finished.
░░
░░ The job identifier is 3611 and the job result is done.
Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: 21b1326f57855b147efcb07cbd10c5f7b7edbabb6693b7752105182196a3f742.service: Failed to open /run/systemd/transient/21b1326f57855b147efcb07cbd10c5f7b7edbabb6693b7752105182196a3f742.service: No such file or directory
Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: run-r9a5a2edf77764b9e940f4c2cad5029cc.scope: Deactivated successfully.
░░ Subject: Unit succeeded
░░ Defined-By: systemd
░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel
░░
░░ The unit run-r9a5a2edf77764b9e940f4c2cad5029cc.scope has successfully entered the 'dead' state.
Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm kernel: podman1: port 1(veth0) entered disabled state
Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm kernel: veth0 (unregistering): left allmulticast mode
Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm kernel: veth0 (unregistering): left promiscuous mode
Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm kernel: podman1: port 1(veth0) entered disabled state
Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: ANOM_PROMISCUOUS dev=veth0 prom=0 old_prom=256 auid=4294967295 uid=0 gid=0 ses=4294967295
Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm NetworkManager[657]: [1720588514.1347] device (podman1): state change: activated -> unmanaged (reason 'unmanaged', sys-iface-state: 'removed')
Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[11260]: NETFILTER_CFG table=nat:50 family=2 entries=1 op=nft_unregister_rule pid=11260 subj=system_u:system_r:iptables_t:s0 comm="iptables"
Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[11266]: NETFILTER_CFG table=nat:51 family=2 entries=1 op=nft_unregister_rule pid=11266 subj=system_u:system_r:iptables_t:s0 comm="iptables"
Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: Starting NetworkManager-dispatcher.service - Network Manager Script Dispatcher Service...
░░ Subject: A start job for unit NetworkManager-dispatcher.service has begun execution
░░ Defined-By: systemd
░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel
░░
░░ A start job for unit NetworkManager-dispatcher.service has begun execution.
░░
░░ The job identifier is 3613.
Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[11268]: NETFILTER_CFG table=nat:52 family=2 entries=1 op=nft_unregister_rule pid=11268 subj=system_u:system_r:iptables_t:s0 comm="iptables"
Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[11270]: NETFILTER_CFG table=filter:53 family=2 entries=1 op=nft_unregister_rule pid=11270 subj=system_u:system_r:iptables_t:s0 comm="iptables"
Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[11276]: NETFILTER_CFG table=filter:54 family=2 entries=1 op=nft_unregister_rule pid=11276 subj=system_u:system_r:iptables_t:s0 comm="iptables"
Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: Started NetworkManager-dispatcher.service - Network Manager Script Dispatcher Service.
░░ Subject: A start job for unit NetworkManager-dispatcher.service has finished successfully
░░ Defined-By: systemd
░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel
░░
░░ A start job for unit NetworkManager-dispatcher.service has finished successfully.
░░
░░ The job identifier is 3613.
Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:init_t:s0 msg='unit=NetworkManager-dispatcher comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success'
Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[11283]: NETFILTER_CFG table=filter:55 family=2 entries=1 op=nft_unregister_rule pid=11283 subj=system_u:system_r:iptables_t:s0 comm="iptables"
Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[11285]: NETFILTER_CFG table=filter:56 family=2 entries=1 op=nft_unregister_rule pid=11285 subj=system_u:system_r:iptables_t:s0 comm="iptables"
Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[11287]: NETFILTER_CFG table=nat:57 family=2 entries=1 op=nft_unregister_chain pid=11287 subj=system_u:system_r:iptables_t:s0 comm="iptables"
Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[11252]: ansible-ansible.legacy.command Invoked with _raw_params=podman ps --noheading _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[6763]: NETFILTER_CFG table=firewalld:58 family=1 entries=46 op=nft_unregister_rule pid=6763 subj=system_u:system_r:firewalld_t:s0 comm="firewalld"
Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: run-netns-netns\x2de51cc4f9\x2d239a\x2d1eda\x2d2116\x2d2ee6febc9b1b.mount: Deactivated successfully.
░░ Subject: Unit succeeded
░░ Defined-By: systemd
░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel
░░
░░ The unit run-netns-netns\x2de51cc4f9\x2d239a\x2d1eda\x2d2116\x2d2ee6febc9b1b.mount has successfully entered the 'dead' state.
Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=253 op=UNLOAD
Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: var-lib-containers-storage-overlay-365acc50bbd6897f3aabf4b51e9c692de2d02cbffe3822f80e928de6dd72014d-merged.mount: Deactivated successfully.
░░ Subject: Unit succeeded
░░ Defined-By: systemd
░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel
░░
░░ The unit var-lib-containers-storage-overlay-365acc50bbd6897f3aabf4b51e9c692de2d02cbffe3822f80e928de6dd72014d-merged.mount has successfully entered the 'dead' state.
Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-21b1326f57855b147efcb07cbd10c5f7b7edbabb6693b7752105182196a3f742-userdata-shm.mount: Deactivated successfully.
░░ Subject: Unit succeeded
░░ Defined-By: systemd
░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel
░░
░░ The unit var-lib-containers-storage-overlay\x2dcontainers-21b1326f57855b147efcb07cbd10c5f7b7edbabb6693b7752105182196a3f742-userdata-shm.mount has successfully entered the 'dead' state.
Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm podman[11232]: 2024-07-10 05:15:14.395138604 +0000 UTC m=+0.362193900 container remove 21b1326f57855b147efcb07cbd10c5f7b7edbabb6693b7752105182196a3f742 (image=quay.io/linux-system-roles/mysql:5.6, name=quadlet-demo-mysql, PODMAN_SYSTEMD_UNIT=quadlet-demo-mysql.service)
Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully.
░░ Subject: Unit succeeded
░░ Defined-By: systemd
░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel
░░
░░ The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state.
Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: quadlet-demo-mysql.service: Main process exited, code=exited, status=137/n/a
░░ Subject: Unit process exited
░░ Defined-By: systemd
░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel
░░
░░ An ExecStart= process belonging to unit quadlet-demo-mysql.service has exited.
░░
░░ The process' exit code is 'exited' and its exit status is 137.
Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: quadlet-demo-mysql.service: Failed with result 'exit-code'.
░░ Subject: Unit failed
░░ Defined-By: systemd
░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel
░░
░░ The unit quadlet-demo-mysql.service has entered the 'failed' state with result 'exit-code'.
Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:init_t:s0 msg='unit=quadlet-demo-mysql comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed'
Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: quadlet-demo-mysql.service: Consumed 3.750s CPU time.
░░ Subject: Resources consumed by unit runtime
░░ Defined-By: systemd
░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel
░░
░░ The unit quadlet-demo-mysql.service completed and consumed the indicated resources.
Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=11307 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success'
Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[11323]: ansible-ansible.legacy.command Invoked with _raw_params=podman network ls -n -q _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=11331 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success'
Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:14 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=11356 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success'
Jul 10 05:15:15 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[11372]: ansible-service_facts Invoked
Jul 10 05:15:17 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:17 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:17 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:17 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:17 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=11484 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success'
Jul 10 05:15:17 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[11500]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None
Jul 10 05:15:17 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:17 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:17 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:17 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:17 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=11502 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success'
Jul 10 05:15:17 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[11518]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1
Jul 10 05:15:17 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:17 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=11521 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success'
Jul 10 05:15:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[11537]: ansible-systemd Invoked with name=quadlet-demo-mysql.service scope=system state=stopped enabled=False force=True daemon_reload=False daemon_reexec=False no_block=False masked=None
Jul 10 05:15:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: Reloading requested from client PID 11540 ('systemctl') (unit session-6.scope)...
Jul 10 05:15:18 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: Reloading...
Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: Reloading finished in 380 ms.
Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=391 op=LOAD
Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=372 op=UNLOAD
Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=392 op=LOAD
Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=373 op=UNLOAD
Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=393 op=LOAD
Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=394 op=LOAD
Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=374 op=UNLOAD
Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=375 op=UNLOAD
Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=395 op=LOAD
Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=376 op=UNLOAD
Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=396 op=LOAD
Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=377 op=UNLOAD
Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=397 op=LOAD
Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=398 op=LOAD
Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=378 op=UNLOAD
Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=379 op=UNLOAD
Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=399 op=LOAD
Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=380 op=UNLOAD
Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=400 op=LOAD
Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=401 op=LOAD
Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=381 op=UNLOAD
Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=382 op=UNLOAD
Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=402 op=LOAD
Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=383 op=UNLOAD
Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=403 op=LOAD
Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=384 op=UNLOAD
Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=404 op=LOAD
Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=385 op=UNLOAD
Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=405 op=LOAD
Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=406 op=LOAD
Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=386 op=UNLOAD
Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=387 op=UNLOAD
Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=407 op=LOAD
Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=388 op=UNLOAD
Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=408 op=LOAD
Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=409 op=LOAD
Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=389 op=UNLOAD
Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=390 op=UNLOAD
Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=11591 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success'
Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[11607]: ansible-file Invoked with path=/etc/containers/systemd/quadlet-demo-mysql.container state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None
Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:19 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=11608 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success'
Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[11624]: ansible-systemd Invoked with daemon_reload=True scope=system daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None
Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: Reloading requested from client PID 11625 ('systemctl') (unit session-6.scope)...
Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: Reloading...
Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: Reloading finished in 372 ms.
Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=410 op=LOAD
Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=391 op=UNLOAD
Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=411 op=LOAD
Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=392 op=UNLOAD
Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=412 op=LOAD
Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=413 op=LOAD
Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=393 op=UNLOAD
Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=394 op=UNLOAD
Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=414 op=LOAD
Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=395 op=UNLOAD
Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=415 op=LOAD
Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=396 op=UNLOAD
Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=416 op=LOAD
Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=417 op=LOAD
Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=397 op=UNLOAD
Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=398 op=UNLOAD
Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=418 op=LOAD
Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=399 op=UNLOAD
Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=419 op=LOAD
Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=420 op=LOAD
Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=400 op=UNLOAD
Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=401 op=UNLOAD
Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=421 op=LOAD
Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=402 op=UNLOAD
Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=422 op=LOAD
Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=403 op=UNLOAD
Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=423 op=LOAD
Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=404 op=UNLOAD
Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=424 op=LOAD
Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=425 op=LOAD
Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=405 op=UNLOAD
Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=406 op=UNLOAD
Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=426 op=LOAD
Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=407 op=UNLOAD
Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=427 op=LOAD
Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=428 op=LOAD
Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=408 op=UNLOAD
Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=409 op=UNLOAD
Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=11676 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success'
Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[11692]: ansible-ansible.legacy.command Invoked with _raw_params=podman image prune --all -f _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm podman[11693]: 2024-07-10 05:15:20.967347867 +0000 UTC m=+0.212576529 image untag dd3b2a5dcb48ff61113592ed5ddd762581be4387c7bc552375a2159422aa6bf5 quay.io/linux-system-roles/mysql:5.6
Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm podman[11693]: 2024-07-10 05:15:20.779292153 +0000 UTC m=+0.024520898 image remove dd3b2a5dcb48ff61113592ed5ddd762581be4387c7bc552375a2159422aa6bf5
Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully.
░░ Subject: Unit succeeded
░░ Defined-By: systemd
░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel
░░
░░ The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state.
Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:20 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=11701 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success'
Jul 10 05:15:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[11717]: ansible-ansible.legacy.command Invoked with _raw_params=podman images -n _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Jul 10 05:15:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully.
░░ Subject: Unit succeeded
░░ Defined-By: systemd
░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel
░░
░░ The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state.
Jul 10 05:15:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=11726 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success'
Jul 10 05:15:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[11742]: ansible-ansible.legacy.command Invoked with _raw_params=podman volume ls -n _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Jul 10 05:15:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=11750 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success'
Jul 10 05:15:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully.
░░ Subject: Unit succeeded
░░ Defined-By: systemd
░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel
░░
░░ The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state.
Jul 10 05:15:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[11766]: ansible-ansible.legacy.command Invoked with _raw_params=podman ps --noheading _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Jul 10 05:15:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully.
░░ Subject: Unit succeeded
░░ Defined-By: systemd
░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel
░░
░░ The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state.
Jul 10 05:15:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:21 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=11774 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success'
Jul 10 05:15:22 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[11790]: ansible-ansible.legacy.command Invoked with _raw_params=podman network ls -n -q _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Jul 10 05:15:22 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully.
░░ Subject: Unit succeeded
░░ Defined-By: systemd
░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel
░░
░░ The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state.
Jul 10 05:15:22 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:22 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:22 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:22 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:22 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=11798 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success'
Jul 10 05:15:22 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:22 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:22 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:22 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:22 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=11823 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success'
Jul 10 05:15:22 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[11839]: ansible-service_facts Invoked
Jul 10 05:15:24 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: NetworkManager-dispatcher.service: Deactivated successfully.
░░ Subject: Unit succeeded
░░ Defined-By: systemd
░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel
░░
░░ The unit NetworkManager-dispatcher.service has successfully entered the 'dead' state.
Jul 10 05:15:24 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:init_t:s0 msg='unit=NetworkManager-dispatcher comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success'
Jul 10 05:15:24 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:24 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:24 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:24 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:24 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=11952 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success'
Jul 10 05:15:25 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[11968]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None
Jul 10 05:15:25 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:25 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:25 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:25 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:25 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=11970 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success'
Jul 10 05:15:25 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[11986]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1
Jul 10 05:15:25 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:25 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:25 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:25 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:25 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=11989 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success'
Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[12005]: ansible-systemd Invoked with name=quadlet-demo-mysql-volume.service scope=system state=stopped enabled=False force=True daemon_reload=False daemon_reexec=False no_block=False masked=None
Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: Reloading requested from client PID 12008 ('systemctl') (unit session-6.scope)...
Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: Reloading...
Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: Reloading finished in 370 ms.
Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=429 op=LOAD
Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=410 op=UNLOAD
Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=430 op=LOAD
Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=411 op=UNLOAD
Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=431 op=LOAD
Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=432 op=LOAD
Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=412 op=UNLOAD
Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=413 op=UNLOAD
Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=433 op=LOAD
Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=414 op=UNLOAD
Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=434 op=LOAD
Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=415 op=UNLOAD
Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=435 op=LOAD
Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=436 op=LOAD
Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=416 op=UNLOAD
Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=417 op=UNLOAD
Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=437 op=LOAD
Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=418 op=UNLOAD
Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=438 op=LOAD
Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=439 op=LOAD
Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=419 op=UNLOAD
Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=420 op=UNLOAD
Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=440 op=LOAD
Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=421 op=UNLOAD
Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=441 op=LOAD
Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=422 op=UNLOAD
Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=442 op=LOAD
Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=423 op=UNLOAD
Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=443 op=LOAD
Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=444 op=LOAD
Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=424 op=UNLOAD
Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=425 op=UNLOAD
Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=445 op=LOAD
Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=426 op=UNLOAD
Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=446 op=LOAD
Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=447 op=LOAD
Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=427 op=UNLOAD
Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=428 op=UNLOAD
Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: quadlet-demo-mysql-volume.service: Deactivated successfully.
░░ Subject: Unit succeeded
░░ Defined-By: systemd
░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel
░░
░░ The unit quadlet-demo-mysql-volume.service has successfully entered the 'dead' state.
Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: Stopped quadlet-demo-mysql-volume.service.
░░ Subject: A stop job for unit quadlet-demo-mysql-volume.service has finished
░░ Defined-By: systemd
░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel
░░
░░ A stop job for unit quadlet-demo-mysql-volume.service has finished.
░░
░░ The job identifier is 3699 and the job result is done.
Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:init_t:s0 msg='unit=quadlet-demo-mysql-volume comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success'
Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=12059 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success'
Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[12075]: ansible-file Invoked with path=/etc/containers/systemd/quadlet-demo-mysql.volume state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None
Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:26 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=12076 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success'
Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[12092]: ansible-systemd Invoked with daemon_reload=True scope=system daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None
Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: Reloading requested from client PID 12093 ('systemctl') (unit session-6.scope)...
Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: Reloading...
Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: Reloading finished in 363 ms.
Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=448 op=LOAD
Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=429 op=UNLOAD
Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=449 op=LOAD
Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=430 op=UNLOAD
Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=450 op=LOAD
Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=451 op=LOAD
Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=431 op=UNLOAD
Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=432 op=UNLOAD
Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=452 op=LOAD
Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=433 op=UNLOAD
Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=453 op=LOAD
Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=434 op=UNLOAD
Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=454 op=LOAD
Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=455 op=LOAD
Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=435 op=UNLOAD
Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=436 op=UNLOAD
Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=456 op=LOAD
Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=437 op=UNLOAD
Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=457 op=LOAD
Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=458 op=LOAD
Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=438 op=UNLOAD
Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=439 op=UNLOAD
Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=459 op=LOAD
Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=440 op=UNLOAD
Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=460 op=LOAD
Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=441 op=UNLOAD
Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=461 op=LOAD
Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=442 op=UNLOAD
Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=462 op=LOAD
Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=463 op=LOAD
Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=443 op=UNLOAD
Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=444 op=UNLOAD
Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=464 op=LOAD
Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=445 op=UNLOAD
Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=465 op=LOAD
Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=466 op=LOAD
Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=446 op=UNLOAD
Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=447 op=UNLOAD
Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:27 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=12143 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success'
Jul 10 05:15:28 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[12159]: ansible-ansible.legacy.command Invoked with _raw_params=podman image prune --all -f _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Jul 10 05:15:28 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:28 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:28 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:28 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:28 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=12167 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success'
Jul 10 05:15:28 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[12183]: ansible-ansible.legacy.command Invoked with _raw_params=podman volume prune -f _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Jul 10 05:15:28 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm podman[12184]: 2024-07-10 05:15:28.410743708 +0000 UTC m=+0.051639497 volume remove systemd-quadlet-demo-mysql
Jul 10 05:15:28 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm podman[12184]: 2024-07-10 05:15:28.411190479 +0000 UTC m=+0.052086323 volume prune systemd-quadlet-demo-mysql
Jul 10 05:15:28 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm podman[12184]: 2024-07-10 05:15:28.45686457 +0000 UTC m=+0.097760616 volume remove wp-pv-claim
Jul 10 05:15:28 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm podman[12184]: 2024-07-10 05:15:28.456915574 +0000 UTC m=+0.097811420 volume prune wp-pv-claim
Jul 10 05:15:28 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm podman[12184]: 2024-07-10 05:15:28.470779425 +0000 UTC m=+0.111675245 volume remove envoy-proxy-config
Jul 10 05:15:28 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm podman[12184]: 2024-07-10 05:15:28.47083971 +0000 UTC m=+0.111735606 volume prune envoy-proxy-config
Jul 10 05:15:28 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm podman[12184]: 2024-07-10 05:15:28.477537993 +0000 UTC m=+0.118433739 volume remove envoy-certificates
Jul 10 05:15:28 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm podman[12184]: 2024-07-10 05:15:28.477585664 +0000 UTC m=+0.118481540 volume prune envoy-certificates
Jul 10 05:15:28 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully.
░░ Subject: Unit succeeded
░░ Defined-By: systemd
░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel
░░
░░ The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state.
Jul 10 05:15:28 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:28 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:28 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:28 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:28 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=12192 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success'
Jul 10 05:15:28 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[12208]: ansible-ansible.legacy.command Invoked with _raw_params=podman images -n _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Jul 10 05:15:28 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:28 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:28 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:28 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:28 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=12217 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success'
Jul 10 05:15:28 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[12233]: ansible-ansible.legacy.command Invoked with _raw_params=podman volume ls -n _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Jul 10 05:15:29 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:29 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:29 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:29 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:29 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=12242 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success'
Jul 10 05:15:29 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[12258]: ansible-ansible.legacy.command Invoked with _raw_params=podman ps --noheading _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Jul 10 05:15:29 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully.
░░ Subject: Unit succeeded
░░ Defined-By: systemd
░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel
░░
░░ The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state.
Jul 10 05:15:29 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:29 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:29 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:29 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:29 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=12267 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success'
Jul 10 05:15:29 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[12283]: ansible-ansible.legacy.command Invoked with _raw_params=podman network ls -n -q _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Jul 10 05:15:29 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully.
░░ Subject: Unit succeeded
░░ Defined-By: systemd
░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel
░░
░░ The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state.
Jul 10 05:15:29 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:29 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:29 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:29 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:29 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=12291 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success'
Jul 10 05:15:29 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:29 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:29 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:29 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:29 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=12315 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success'
Jul 10 05:15:30 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[12331]: ansible-service_facts Invoked
Jul 10 05:15:32 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:32 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:32 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:32 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:32 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=12444 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success'
Jul 10 05:15:32 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[12460]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None
Jul 10 05:15:32 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:32 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:32 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:32 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:32 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=12462 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success'
Jul 10 05:15:32 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[12478]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1
Jul 10 05:15:32 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:32 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:33 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:33 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:33 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=12481 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success'
Jul 10 05:15:33 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[12497]: ansible-systemd Invoked with name=quadlet-demo-network.service scope=system state=stopped enabled=False force=True daemon_reload=False daemon_reexec=False no_block=False masked=None
Jul 10 05:15:33 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: Reloading requested from client PID 12500 ('systemctl') (unit session-6.scope)...
Jul 10 05:15:33 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: Reloading...
Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: Reloading finished in 361 ms.
Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=467 op=LOAD
Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=448 op=UNLOAD
Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=468 op=LOAD
Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=449 op=UNLOAD
Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=469 op=LOAD
Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=470 op=LOAD
Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=450 op=UNLOAD
Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=451 op=UNLOAD
Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=471 op=LOAD
Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=452 op=UNLOAD
Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=472 op=LOAD
Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=453 op=UNLOAD
Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=473 op=LOAD
Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=474 op=LOAD
Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=454 op=UNLOAD
Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=455 op=UNLOAD
Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=475 op=LOAD
Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=456 op=UNLOAD
Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=476 op=LOAD
Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=477 op=LOAD
Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=457 op=UNLOAD
Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=458 op=UNLOAD
Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=478 op=LOAD
Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=459 op=UNLOAD
Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=479 op=LOAD
Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=460 op=UNLOAD
Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=480 op=LOAD
Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=461 op=UNLOAD
Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=481 op=LOAD
Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=482 op=LOAD
Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=462 op=UNLOAD
Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=463 op=UNLOAD
Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=483 op=LOAD
Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=464 op=UNLOAD
Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=484 op=LOAD
Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=485 op=LOAD
Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=465 op=UNLOAD
Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=466 op=UNLOAD
Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:init_t:s0 msg='unit=quadlet-demo-network comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success'
Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: quadlet-demo-network.service: Deactivated successfully.
░░ Subject: Unit succeeded
░░ Defined-By: systemd
░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel
░░
░░ The unit quadlet-demo-network.service has successfully entered the 'dead' state.
Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: Stopped quadlet-demo-network.service.
░░ Subject: A stop job for unit quadlet-demo-network.service has finished
░░ Defined-By: systemd
░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel
░░
░░ A stop job for unit quadlet-demo-network.service has finished.
░░
░░ The job identifier is 3700 and the job result is done.
Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=12551 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success'
Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[12567]: ansible-file Invoked with path=/etc/containers/systemd/quadlet-demo.network state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None
Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=12568 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success'
Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[12584]: ansible-systemd Invoked with daemon_reload=True scope=system daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None
Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: Reloading requested from client PID 12585 ('systemctl') (unit session-6.scope)...
Jul 10 05:15:34 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: Reloading...
Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: Reloading finished in 363 ms.
Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=486 op=LOAD
Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=467 op=UNLOAD
Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=487 op=LOAD
Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=468 op=UNLOAD
Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=488 op=LOAD
Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=489 op=LOAD
Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=469 op=UNLOAD
Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=470 op=UNLOAD
Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=490 op=LOAD
Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=471 op=UNLOAD
Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=491 op=LOAD
Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=472 op=UNLOAD
Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=492 op=LOAD
Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=493 op=LOAD
Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=473 op=UNLOAD
Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=474 op=UNLOAD
Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=494 op=LOAD
Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=475 op=UNLOAD
Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=495 op=LOAD
Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=496 op=LOAD
Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=476 op=UNLOAD
Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=477 op=UNLOAD
Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=497 op=LOAD
Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=478 op=UNLOAD
Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=498 op=LOAD
Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=479 op=UNLOAD
Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=499 op=LOAD
Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=480 op=UNLOAD
Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=500 op=LOAD
Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=501 op=LOAD
Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=481 op=UNLOAD
Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=482 op=UNLOAD
Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=502 op=LOAD
Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=483 op=UNLOAD
Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=503 op=LOAD
Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=504 op=LOAD
Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=484 op=UNLOAD
Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit: BPF prog-id=485 op=UNLOAD
Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=12636 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success'
Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[12652]: ansible-ansible.legacy.command Invoked with _raw_params=podman image prune --all -f _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully.
░░ Subject: Unit succeeded
░░ Defined-By: systemd
░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel
░░
░░ The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state.
Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=12660 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success'
Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[12676]: ansible-ansible.legacy.command Invoked with _raw_params=podman network prune -f _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:35 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=12684 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success'
Jul 10 05:15:36 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[12700]: ansible-ansible.legacy.command Invoked with _raw_params=podman images -n _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Jul 10 05:15:36 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:36 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:36 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:36 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:36 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=12708 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success'
Jul 10 05:15:36 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[12724]: ansible-ansible.legacy.command Invoked with _raw_params=podman volume ls -n _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Jul 10 05:15:36 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:36 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:36 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:36 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:36 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=12732 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success'
Jul 10 05:15:36 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully.
░░ Subject: Unit succeeded
░░ Defined-By: systemd
░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel
░░
░░ The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state.
Jul 10 05:15:36 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[12748]: ansible-ansible.legacy.command Invoked with _raw_params=podman ps --noheading _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Jul 10 05:15:36 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully.
░░ Subject: Unit succeeded
░░ Defined-By: systemd
░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel
░░
░░ The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state.
Jul 10 05:15:36 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:36 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:36 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:36 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:36 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=12756 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success'
Jul 10 05:15:37 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[12772]: ansible-ansible.legacy.command Invoked with _raw_params=podman network ls -n -q _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Jul 10 05:15:37 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully.
░░ Subject: Unit succeeded
░░ Defined-By: systemd
░░ Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel
░░
░░ The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state.
Jul 10 05:15:37 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:37 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:37 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:37 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:37 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=12780 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success'
Jul 10 05:15:37 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:37 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:37 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:37 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:37 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=12804 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success'
Jul 10 05:15:37 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[12820]: ansible-service_facts Invoked
Jul 10 05:15:39 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:39 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:39 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:39 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:39 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=12933 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success'
Jul 10 05:15:39 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[12949]: ansible-ansible.legacy.command Invoked with _raw_params=exec 1>&2
set -x
set -o pipefail
systemctl list-units --plain -l --all | grep quadlet || :
systemctl list-unit-files --all | grep quadlet || :
systemctl list-units --plain --failed -l --all | grep quadlet || :
journalctl -u quadlet-demo-mysql.service || :
systemctl status quadlet-demo-mysql.service || :
_uses_shell=True expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Jul 10 05:15:40 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_END pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:40 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGOUT pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:40 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_LOGIN pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:40 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: USER_START pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=login id=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=ssh res=success'
Jul 10 05:15:40 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm audit[1002]: CRYPTO_KEY_USER pid=1002 uid=0 auid=0 ses=6 subj=system_u:system_r:sshd_t:s0-s0:c0.c1023 msg='op=destroy kind=server fp=SHA256:cd:a3:5c:ba:5c:28:94:39:99:e4:a9:5c:a3:e4:8b:fa:aa:4d:51:71:c7:0c:90:ce:1e:88:db:95:4a:a4:a6:d3 direction=? spid=12959 suid=0 exe="/usr/sbin/sshd" hostname=? addr=10.29.163.169 terminal=? res=success'
Jul 10 05:15:40 4c6d2b04-94a5-48bb-bd26-d49357f99041.testing-farm python3[12975]: ansible-ansible.legacy.command Invoked with _raw_params=journalctl -ex _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
to retry, use: --limit @/tmp/tests_quadlet_demo.retry
PLAY RECAP *********************************************************************
sut : ok=374 changed=49 unreachable=0 failed=1 skipped=353 rescued=1 ignored=0
Wednesday 10 July 2024 05:15:40 +0000 (0:00:00.263) 0:03:45.461 ********
===============================================================================
fedora.linux_system_roles.podman : Ensure required packages are installed -- 94.89s
/WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:14
fedora.linux_system_roles.podman : Ensure container images are present -- 18.02s
/WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:18
fedora.linux_system_roles.podman : Ensure container images are present --- 6.86s
/WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:18
fedora.linux_system_roles.certificate : Ensure certificate role dependencies are installed --- 4.83s
/WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:5
fedora.linux_system_roles.certificate : Ensure provider packages are installed --- 4.47s
/WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:23
fedora.linux_system_roles.podman : For testing and debugging - services --- 2.30s
/WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:133
fedora.linux_system_roles.podman : For testing and debugging - services --- 2.30s
/WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:133
fedora.linux_system_roles.firewall : Install firewalld ------------------ 2.29s
/WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:19
fedora.linux_system_roles.podman : For testing and debugging - services --- 2.22s
/WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:133
fedora.linux_system_roles.podman : For testing and debugging - services --- 2.21s
/WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:133
fedora.linux_system_roles.podman : For testing and debugging - services --- 2.19s
/WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:133
fedora.linux_system_roles.podman : For testing and debugging - services --- 2.18s
/WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:133
fedora.linux_system_roles.firewall : Install firewalld ------------------ 2.18s
/WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:19
fedora.linux_system_roles.podman : Stop and disable service ------------- 1.42s
/WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:12
fedora.linux_system_roles.podman : Start service ------------------------ 1.34s
/WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:110
fedora.linux_system_roles.podman : Reload systemctl --------------------- 1.28s
/WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:82
fedora.linux_system_roles.firewall : Configure firewall ----------------- 1.17s
/WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:71
Check web --------------------------------------------------------------- 1.11s
/WORKDIR/git-handle-quadlet-cleanup35__kigx/tests/tests_quadlet_demo.yml:104 --
fedora.linux_system_roles.firewall : Enable and start firewalld service --- 1.09s
/WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:28
fedora.linux_system_roles.certificate : Ensure provider service is running --- 1.01s
/WORKDIR/git-handle-quadlet-cleanup35__kigx/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:90
---^---^---^---^---^---
# STDERR:
---v---v---v---v---v---
[DEPRECATION WARNING]: ANSIBLE_COLLECTIONS_PATHS option, does not fit var
naming standard, use the singular form ANSIBLE_COLLECTIONS_PATH instead. This
feature will be removed from ansible-core in version 2.19. Deprecation warnings
can be disabled by setting deprecation_warnings=False in ansible.cfg.
[WARNING]: Using a variable for a task's 'args' is unsafe in some situations
(see
https://docs.ansible.com/ansible/devel/reference_appendices/faq.html#argsplat-
unsafe)
---^---^---^---^---^---