# STDOUT: ---v---v---v---v---v--- ansible-playbook [core 2.16.0] config file = /etc/ansible/ansible.cfg configured module search path = ['/home/jenkins/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules'] ansible python module location = /opt/ansible-2.16/lib/python3.11/site-packages/ansible ansible collection location = /WORKDIR/git-fix-linger-handlingih6nvedg/.collection executable location = /opt/ansible-2.16/bin/ansible-playbook python version = 3.11.5 (main, Sep 7 2023, 00:00:00) [GCC 11.4.1 20230605 (Red Hat 11.4.1-2)] (/opt/ansible-2.16/bin/python) jinja version = 3.1.2 libyaml = True Using /etc/ansible/ansible.cfg as config file Skipping callback 'debug', as we already have a stdout callback. Skipping callback 'default', as we already have a stdout callback. Skipping callback 'minimal', as we already have a stdout callback. Skipping callback 'oneline', as we already have a stdout callback. PLAYBOOK: tests_quadlet_demo.yml *********************************************** 2 plays in /WORKDIR/git-fix-linger-handlingih6nvedg/tests/tests_quadlet_demo.yml PLAY [all] ********************************************************************* TASK [Include vault variables] ************************************************* task path: /WORKDIR/git-fix-linger-handlingih6nvedg/tests/tests_quadlet_demo.yml:4 Tuesday 06 February 2024 16:17:13 +0000 (0:00:00.013) 0:00:00.013 ****** ok: [sut] => { "ansible_facts": { "mysql_container_root_password": { "__ansible_vault": "$ANSIBLE_VAULT;1.1;AES256\n61333932373230333539663035366431326163363166363036323963623131363530326231303634\n6635326161643165363366323062333334363730376631660a393566366139353861656364656661\n38653463363837336639363032646433666361646535366137303464623261313663643336306465\n6264663730656337310a343962353137386238383064646533366433333437303566656433386233\n34343235326665646661623131643335313236313131353661386338343366316261643634653633\n3832313034366536616531323963333234326461353130303532\n" } }, "ansible_included_var_files": [ "/WORKDIR/git-fix-linger-handlingih6nvedg/tests/vars/vault-variables.yml" ], "changed": false } PLAY [Deploy the quadlet demo app] ********************************************* TASK [Gathering Facts] ********************************************************* task path: /WORKDIR/git-fix-linger-handlingih6nvedg/tests/tests_quadlet_demo.yml:9 Tuesday 06 February 2024 16:17:13 +0000 (0:00:00.014) 0:00:00.027 ****** ok: [sut] TASK [Generate certificates] *************************************************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/tests/tests_quadlet_demo.yml:39 Tuesday 06 February 2024 16:17:14 +0000 (0:00:00.841) 0:00:00.868 ****** TASK [fedora.linux_system_roles.certificate : Set version specific variables] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:2 Tuesday 06 February 2024 16:17:14 +0000 (0:00:00.024) 0:00:00.892 ****** included: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/set_vars.yml for sut TASK [fedora.linux_system_roles.certificate : Ensure ansible_facts used by role] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/set_vars.yml:2 Tuesday 06 February 2024 16:17:14 +0000 (0:00:00.014) 0:00:00.907 ****** skipping: [sut] => { "changed": false, "false_condition": "__certificate_required_facts | difference(ansible_facts.keys() | list) | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.certificate : Check if system is ostree] ******* task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/set_vars.yml:10 Tuesday 06 February 2024 16:17:14 +0000 (0:00:00.020) 0:00:00.927 ****** ok: [sut] => { "changed": false, "stat": { "exists": false } } TASK [fedora.linux_system_roles.certificate : Set flag to indicate system is ostree] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/set_vars.yml:15 Tuesday 06 February 2024 16:17:14 +0000 (0:00:00.255) 0:00:01.183 ****** ok: [sut] => { "ansible_facts": { "__certificate_is_ostree": false }, "changed": false } TASK [fedora.linux_system_roles.certificate : Set platform/version specific variables] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/set_vars.yml:19 Tuesday 06 February 2024 16:17:14 +0000 (0:00:00.017) 0:00:01.200 ****** skipping: [sut] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "false_condition": "__vars_file is file", "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=CentOS.yml) => { "ansible_loop_var": "item", "changed": false, "false_condition": "__vars_file is file", "item": "CentOS.yml", "skip_reason": "Conditional result was False" } ok: [sut] => (item=CentOS_9.yml) => { "ansible_facts": { "__certificate_certmonger_packages": [ "certmonger", "python3-packaging" ] }, "ansible_included_var_files": [ "/WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/vars/CentOS_9.yml" ], "ansible_loop_var": "item", "changed": false, "item": "CentOS_9.yml" } ok: [sut] => (item=CentOS_9.yml) => { "ansible_facts": { "__certificate_certmonger_packages": [ "certmonger", "python3-packaging" ] }, "ansible_included_var_files": [ "/WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/vars/CentOS_9.yml" ], "ansible_loop_var": "item", "changed": false, "item": "CentOS_9.yml" } TASK [fedora.linux_system_roles.certificate : Ensure certificate role dependencies are installed] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:5 Tuesday 06 February 2024 16:17:14 +0000 (0:00:00.028) 0:00:01.229 ****** changed: [sut] => { "changed": true, "rc": 0, "results": [ "Installed: python3-pyasn1-0.4.8-6.el9.noarch", "Installed: python3-cffi-1.14.5-5.el9.x86_64", "Installed: python3-ply-3.11-14.el9.noarch", "Installed: python3-pycparser-2.20-6.el9.noarch", "Installed: python3-cryptography-36.0.1-4.el9.x86_64" ] } TASK [fedora.linux_system_roles.certificate : Ensure provider packages are installed] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:23 Tuesday 06 February 2024 16:17:20 +0000 (0:00:05.943) 0:00:07.172 ****** changed: [sut] => (item=certmonger) => { "__certificate_provider": "certmonger", "ansible_loop_var": "__certificate_provider", "changed": true, "rc": 0, "results": [ "Installed: certmonger-0.79.17-1.el9.x86_64", "Installed: python3-packaging-20.9-5.el9.noarch", "Installed: python3-pyparsing-2.4.7-9.el9.noarch" ] } TASK [fedora.linux_system_roles.certificate : Ensure pre-scripts hooks directory exists] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:35 Tuesday 06 February 2024 16:17:24 +0000 (0:00:03.319) 0:00:10.491 ****** changed: [sut] => (item=certmonger) => { "__certificate_provider": "certmonger", "ansible_loop_var": "__certificate_provider", "changed": true, "gid": 0, "group": "root", "mode": "0700", "owner": "root", "path": "/etc/certmonger//pre-scripts", "secontext": "unconfined_u:object_r:etc_t:s0", "size": 6, "state": "directory", "uid": 0 } TASK [fedora.linux_system_roles.certificate : Ensure post-scripts hooks directory exists] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:61 Tuesday 06 February 2024 16:17:24 +0000 (0:00:00.286) 0:00:10.778 ****** changed: [sut] => (item=certmonger) => { "__certificate_provider": "certmonger", "ansible_loop_var": "__certificate_provider", "changed": true, "gid": 0, "group": "root", "mode": "0700", "owner": "root", "path": "/etc/certmonger//post-scripts", "secontext": "unconfined_u:object_r:etc_t:s0", "size": 6, "state": "directory", "uid": 0 } TASK [fedora.linux_system_roles.certificate : Ensure provider service is running] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:90 Tuesday 06 February 2024 16:17:24 +0000 (0:00:00.215) 0:00:10.994 ****** changed: [sut] => (item=certmonger) => { "__certificate_provider": "certmonger", "ansible_loop_var": "__certificate_provider", "changed": true, "enabled": true, "name": "certmonger", "state": "started", "status": { "AccessSELinuxContext": "system_u:object_r:certmonger_unit_file_t:s0", "ActiveEnterTimestampMonotonic": "0", "ActiveExitTimestampMonotonic": "0", "ActiveState": "inactive", "After": "systemd-journald.socket dbus.socket basic.target system.slice network.target syslog.target dbus-broker.service sysinit.target", "AllowIsolate": "no", "AssertResult": "no", "AssertTimestampMonotonic": "0", "Before": "shutdown.target", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "BusName": "org.fedorahosted.certmonger", "CPUAccounting": "yes", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "[not set]", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "no", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore", "CleanResult": "success", "CollectMode": "inactive", "ConditionResult": "no", "ConditionTimestampMonotonic": "0", "ConfigurationDirectoryMode": "0755", "Conflicts": "shutdown.target", "ControlGroupId": "0", "ControlPID": "0", "CoredumpFilter": "0x33", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "Delegate": "no", "Description": "Certificate monitoring and PKI enrollment", "DevicePolicy": "auto", "DynamicUser": "no", "EnvironmentFiles": "/etc/sysconfig/certmonger (ignore_errors=yes)", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "0", "ExecMainStartTimestampMonotonic": "0", "ExecMainStatus": "0", "ExecStart": "{ path=/usr/sbin/certmonger ; argv[]=/usr/sbin/certmonger -S -p /run/certmonger.pid -n $OPTS ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStartEx": "{ path=/usr/sbin/certmonger ; argv[]=/usr/sbin/certmonger -S -p /run/certmonger.pid -n $OPTS ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExitType": "main", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FinalKillSignal": "9", "FragmentPath": "/usr/lib/systemd/system/certmonger.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOReadBytes": "18446744073709551615", "IOReadOperations": "18446744073709551615", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "IOWeight": "[not set]", "IOWriteBytes": "18446744073709551615", "IOWriteOperations": "18446744073709551615", "IPAccounting": "no", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "Id": "certmonger.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestampMonotonic": "0", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "control-group", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "0", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "8388608", "LimitMEMLOCKSoft": "8388608", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "524288", "LimitNOFILESoft": "1024", "LimitNPROC": "13974", "LimitNPROCSoft": "13974", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "13974", "LimitSIGPENDINGSoft": "13974", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "0", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "ManagedOOMSwap": "auto", "MemoryAccounting": "yes", "MemoryAvailable": "infinity", "MemoryCurrent": "[not set]", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemorySwapMax": "infinity", "MountAPIVFS": "no", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAPolicy": "n/a", "Names": "certmonger.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMPolicy": "stop", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "OnSuccessJobMode": "fail", "PIDFile": "/run/certmonger.pid", "PartOf": "dbus-broker.service", "Perpetual": "no", "PrivateDevices": "no", "PrivateIPC": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProcSubset": "all", "ProtectClock": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectHostname": "no", "ProtectKernelLogs": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectProc": "default", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "ReloadResult": "success", "ReloadSignal": "1", "RemainAfterExit": "no", "RemoveIPC": "no", "Requires": "sysinit.target system.slice dbus.socket", "Restart": "no", "RestartKillSignal": "15", "RestartUSec": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "Slice": "system.slice", "StandardError": "inherit", "StandardInput": "null", "StandardOutput": "journal", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StateChangeTimestampMonotonic": "0", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "dead", "SuccessAction": "none", "SyslogFacility": "3", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "2147483646", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "[not set]", "TasksMax": "22359", "TimeoutAbortUSec": "1min 30s", "TimeoutCleanUSec": "infinity", "TimeoutStartFailureMode": "terminate", "TimeoutStartUSec": "1min 30s", "TimeoutStopFailureMode": "terminate", "TimeoutStopUSec": "1min 30s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "dbus", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "disabled", "UnitFileState": "disabled", "UtmpMode": "init", "WatchdogSignal": "6", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "infinity" } } TASK [fedora.linux_system_roles.certificate : Ensure certificate requests] ***** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:101 Tuesday 06 February 2024 16:17:25 +0000 (0:00:00.759) 0:00:11.753 ****** changed: [sut] => (item={'name': 'quadlet_demo', 'dns': ['localhost'], 'ca': 'self-sign'}) => { "ansible_loop_var": "item", "changed": true, "item": { "ca": "self-sign", "dns": [ "localhost" ], "name": "quadlet_demo" } } MSG: Certificate requested (new). TASK [fedora.linux_system_roles.certificate : Slurp the contents of the files] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:152 Tuesday 06 February 2024 16:17:26 +0000 (0:00:00.925) 0:00:12.678 ****** ok: [sut] => (item=['cert', {'name': 'quadlet_demo', 'dns': ['localhost'], 'ca': 'self-sign'}]) => { "ansible_loop_var": "item", "changed": false, "content": "LS0tLS1CRUdJTiBDRVJUSUZJQ0FURS0tLS0tCk1JSURnakNDQW1xZ0F3SUJBZ0lRT0syRHpnYjFUa1crSUExc0FtcEtYREFOQmdrcWhraUc5dzBCQVFzRkFEQlEKTVNBd0hnWURWUVFEREJkTWIyTmhiQ0JUYVdkdWFXNW5JRUYxZEdodmNtbDBlVEVzTUNvR0ExVUVBd3dqTXpoaApaRGd6WTJVdE1EWm1OVFJsTkRVdFltVXlNREJrTm1NdE1ESTJZVFJoTldJd0hoY05NalF3TWpBMk1UWXhOekkyCldoY05NalV3TWpBMk1UWXhOekkxV2pBVU1SSXdFQVlEVlFRREV3bHNiMk5oYkdodmMzUXdnZ0VpTUEwR0NTcUcKU0liM0RRRUJBUVVBQTRJQkR3QXdnZ0VLQW9JQkFRRE9nR1BWY29CZHhoZzlvdXZTZDJxdVE2cFlOdDVDV2tjNQpDZ2dTOHNNa3pCTm5aRUdPd1paR2Vmd1o0NTRaM1ZqR2o1R1FPbEhLZlpxUERycUR3VENLVEZRRFFIbmZwc2N1Ci9yUXdVQmNKSll4ZzAyeTZkWHpkZ1cxNmQxWjc0RFFaZmlrTVdsVFFzd3RaRkpaNjYzaWlWa20vY3JSK3ZIczUKT3ovTkhDcnJzT2FYSTdOQ0tvOTllT2p6T25yRzBGbG0yeU0rTFYwNldhcWtTbGNXYTV0VGlCejdBcU95bFdFLwpBQ3dyUG1Da0NUMU44d2F5ZUYzNVhUTEtaTXgrUVIzSWRkbjZMbWdrZUcwZ1RSd2RoRXNNMUl4RHVhK1F6VkhPCjEycThHcjFYd1d6MmNRZjZhdXBlbFA3dVR4SXcxbGdlNmdHVGtOYzh1ZHcxNnNUM2tJc1RBZ01CQUFHamdaTXcKZ1pBd0N3WURWUjBQQkFRREFnV2dNQlFHQTFVZEVRUU5NQXVDQ1d4dlkyRnNhRzl6ZERBZEJnTlZIU1VFRmpBVQpCZ2dyQmdFRkJRY0RBUVlJS3dZQkJRVUhBd0l3REFZRFZSMFRBUUgvQkFJd0FEQWRCZ05WSFE0RUZnUVV5ZUxXCm51aHUyRk9yaXVwZ3ByT0tvd2EvRUc0d0h3WURWUjBqQkJnd0ZvQVV2OFlxZ3Z5MExkQXp3QUV2cHV2azBLazUKZjE4d0RRWUpLb1pJaHZjTkFRRUxCUUFEZ2dFQkFBS0VrZWgzc09VUklsMkNJdzBrblhnYUN0ZXVBeUVvM0FaLwpadzdlczRCeU1HdVVjSHEyaVBqQU84NW55ejhxTG5PUUtLV0xObTl4MVEvdzluS2o1NXA4dnErSFNueWJvcFh3CkN4MHNWRWxQOHQ1UjFtMG1UK0pGbG5lYnZxSDV5Z0FIMmdiUWh3a2wxZ05jSWMwanNnU0VzWERnK0JMbXpRSG8KM0pyaVROelkzcXpFOTlTNlBkM0llemNya3hKcSs3WVRvSnNBNGkyWEppdmNMb21DR3lWOCtObi9WRVVzRTJXVQpaQjNiU0d0bG1yWTNEZS9md28zVEpTU2lHK2R0OUtLcUd1dWNwYmxvWkR4ajVXOE9ib2g4NHR0M2QvQzdwYm5qClR3RkxVYjYyN096UFNMcVdiYzdCZEFwaGFZZDQwOVJydWFweUtXY3h4U203a1Q4RGRFbz0KLS0tLS1FTkQgQ0VSVElGSUNBVEUtLS0tLQo=", "encoding": "base64", "item": [ "cert", { "ca": "self-sign", "dns": [ "localhost" ], "name": "quadlet_demo" } ], "source": "/etc/pki/tls/certs/quadlet_demo.crt" } ok: [sut] => (item=['key', {'name': 'quadlet_demo', 'dns': ['localhost'], 'ca': 'self-sign'}]) => { "ansible_loop_var": "item", "changed": false, "content": "LS0tLS1CRUdJTiBQUklWQVRFIEtFWS0tLS0tCk1JSUV2UUlCQURBTkJna3Foa2lHOXcwQkFRRUZBQVNDQktjd2dnU2pBZ0VBQW9JQkFRRE9nR1BWY29CZHhoZzkKb3V2U2QycXVRNnBZTnQ1Q1drYzVDZ2dTOHNNa3pCTm5aRUdPd1paR2Vmd1o0NTRaM1ZqR2o1R1FPbEhLZlpxUApEcnFEd1RDS1RGUURRSG5mcHNjdS9yUXdVQmNKSll4ZzAyeTZkWHpkZ1cxNmQxWjc0RFFaZmlrTVdsVFFzd3RaCkZKWjY2M2lpVmttL2NyUit2SHM1T3ovTkhDcnJzT2FYSTdOQ0tvOTllT2p6T25yRzBGbG0yeU0rTFYwNldhcWsKU2xjV2E1dFRpQno3QXFPeWxXRS9BQ3dyUG1Da0NUMU44d2F5ZUYzNVhUTEtaTXgrUVIzSWRkbjZMbWdrZUcwZwpUUndkaEVzTTFJeER1YStRelZITzEycThHcjFYd1d6MmNRZjZhdXBlbFA3dVR4SXcxbGdlNmdHVGtOYzh1ZHcxCjZzVDNrSXNUQWdNQkFBRUNnZ0VBSVlWZzcyQzNhc0dCNU51Q0dYZXVGeHkvMWJNT2NFbjBYdE44alFyZ1Y2UlQKRUFrTHhVOHloRDFBQ2JuVHk5SU5HNnBwa0RhUFdXWDdSWGE0R3pRc20zNlpjcSt3T1BoN1Erc3Z0OW9BYTBrdgptUFlwSTdsM3JBTmJFTEszbE1OSDRjMkRmMUJZalVwTE1Nb3hSVUhkSXNrTURxcGhsMk1rUTUyVkZVZ1p4WWwyCnJ3REZUc25pa3BNYU1nVXFKM0RFd0NTUjQ1d3J4V2lCaE9tckRjcmdaRDJpSHNsclZHc3lhWG9lY000bVBSK0UKdldjc25YY1h6MWtqbWxhQXFHMUJXekVMMWhqRlpaQUw4WGEyQ3h3VjVTNVZmQmZGN1FaeVZtdXBJMU45YVZvaAo5WGcwdGxIb01uUVA2ZWVLMThPU094MlBqWnd2OFVORHRrd0tWNCtWVFFLQmdRRGZMUVZCUjF5VUdxSW9HRGFiClR2OXFlcG5SRytUM3o2Q0FHclZRTVhMNi9hQm1xZVpjdkxFaEJTRnlsek0xK3Fvd1RUejAvdkFvaFVmSG1aTjUKMjR2T1Rtby9OMlI1YjF3YkhWcGRxUS9yemxJT0ViR0NoT3hMNHNGK1BJTTlJMW9PL0FBK3MxSnNtQkhyckZ6Zgp6cm8xcW1wTXpESUFkMmxRRGdpTTFkVVpsUUtCZ1FEczM0elVTaldzMU9WWFpPUDhDbnc4cjdRZGZaY29zK0hBClJsZ1dPVTIxNTl5b2g4VVVnK1BTWWtLSzMyNjZmckE3Z2VrU1JvVHE0Y2RQUS9nKzVLVDUvZ1l3aDJqUDZQNDAKWnEwRkNaOFNtMzl1MGhLRjJJR2plQ3Z0czYrdzQ1anJJZS9aY3FMQ2tIN0NUd04rRjREQnBKaEljK0JQdzNIYQpIc1VBdGdONEJ3S0JnUURiL3NvdVZrVXBiYk9wYUd6ejNnU2psdVZubjJLdThjbmFYZm1iYXBOWVhZbXZpZzJlCkh4ZFFZSGxmTnpMeEkvNStETFZwSUtIbW43TnBTN2QzcVBjQTBJV2xlVm9lNXFwWnhDNmZJeGVoc0Vsc3V2NUwKbGZKRXYrdFRJQjJTdC9WL3plUUQ2c3IyWjFWTEJ5YXFMR09jb1RlV0tyZWpoQktaS3F3OUFGREhuUUtCZ0hzcgphekZnUFRUWHAwR2dacVJDVjRPKzRXdU1TTE9UaGNuVFNpNlVNWUpyUkdRbDVTYmJ4ZlU1dnR3emZvSmZLZ2huCmJ1aGYzdkc3NFVXb1ZFZHY0UkZWTTZhOFlwWHAwM1V3QnJ2ZDJ1ZDBEL2x5SnN1T1h5cS9iQ2c0RFIxdjI2NDQKZ1YyU3JHdDdjMDVhZzdoblJENXM5ZnFHdHBjRFRBcGRnWmJ6RGk2MUFvR0FNODhpSk9yVzZkYjA2UW5hbnVvcwpReGVWWXBJWE03djdUUFpicVpuL2ZzRFZaak5YQW5kQUNUTEN3ZXpFZTF1M3U1b0hzZXNZZGVHNWd5V05mQnQrCkpqb29sSlQ1cDdVbEI3aCtML2ZpYjNPVGJzL3pRbnV4QUE0U05vWk5xblRUQ1ZUREFSRGovd0p5d3FsakFhRXkKb05rYUhiSzI4MkVWeEJYbmwvTjY5Z3c9Ci0tLS0tRU5EIFBSSVZBVEUgS0VZLS0tLS0K", "encoding": "base64", "item": [ "key", { "ca": "self-sign", "dns": [ "localhost" ], "name": "quadlet_demo" } ], "source": "/etc/pki/tls/private/quadlet_demo.key" } ok: [sut] => (item=['ca', {'name': 'quadlet_demo', 'dns': ['localhost'], 'ca': 'self-sign'}]) => { "ansible_loop_var": "item", "changed": false, "content": "LS0tLS1CRUdJTiBDRVJUSUZJQ0FURS0tLS0tCk1JSURnakNDQW1xZ0F3SUJBZ0lRT0syRHpnYjFUa1crSUExc0FtcEtYREFOQmdrcWhraUc5dzBCQVFzRkFEQlEKTVNBd0hnWURWUVFEREJkTWIyTmhiQ0JUYVdkdWFXNW5JRUYxZEdodmNtbDBlVEVzTUNvR0ExVUVBd3dqTXpoaApaRGd6WTJVdE1EWm1OVFJsTkRVdFltVXlNREJrTm1NdE1ESTJZVFJoTldJd0hoY05NalF3TWpBMk1UWXhOekkyCldoY05NalV3TWpBMk1UWXhOekkxV2pBVU1SSXdFQVlEVlFRREV3bHNiMk5oYkdodmMzUXdnZ0VpTUEwR0NTcUcKU0liM0RRRUJBUVVBQTRJQkR3QXdnZ0VLQW9JQkFRRE9nR1BWY29CZHhoZzlvdXZTZDJxdVE2cFlOdDVDV2tjNQpDZ2dTOHNNa3pCTm5aRUdPd1paR2Vmd1o0NTRaM1ZqR2o1R1FPbEhLZlpxUERycUR3VENLVEZRRFFIbmZwc2N1Ci9yUXdVQmNKSll4ZzAyeTZkWHpkZ1cxNmQxWjc0RFFaZmlrTVdsVFFzd3RaRkpaNjYzaWlWa20vY3JSK3ZIczUKT3ovTkhDcnJzT2FYSTdOQ0tvOTllT2p6T25yRzBGbG0yeU0rTFYwNldhcWtTbGNXYTV0VGlCejdBcU95bFdFLwpBQ3dyUG1Da0NUMU44d2F5ZUYzNVhUTEtaTXgrUVIzSWRkbjZMbWdrZUcwZ1RSd2RoRXNNMUl4RHVhK1F6VkhPCjEycThHcjFYd1d6MmNRZjZhdXBlbFA3dVR4SXcxbGdlNmdHVGtOYzh1ZHcxNnNUM2tJc1RBZ01CQUFHamdaTXcKZ1pBd0N3WURWUjBQQkFRREFnV2dNQlFHQTFVZEVRUU5NQXVDQ1d4dlkyRnNhRzl6ZERBZEJnTlZIU1VFRmpBVQpCZ2dyQmdFRkJRY0RBUVlJS3dZQkJRVUhBd0l3REFZRFZSMFRBUUgvQkFJd0FEQWRCZ05WSFE0RUZnUVV5ZUxXCm51aHUyRk9yaXVwZ3ByT0tvd2EvRUc0d0h3WURWUjBqQkJnd0ZvQVV2OFlxZ3Z5MExkQXp3QUV2cHV2azBLazUKZjE4d0RRWUpLb1pJaHZjTkFRRUxCUUFEZ2dFQkFBS0VrZWgzc09VUklsMkNJdzBrblhnYUN0ZXVBeUVvM0FaLwpadzdlczRCeU1HdVVjSHEyaVBqQU84NW55ejhxTG5PUUtLV0xObTl4MVEvdzluS2o1NXA4dnErSFNueWJvcFh3CkN4MHNWRWxQOHQ1UjFtMG1UK0pGbG5lYnZxSDV5Z0FIMmdiUWh3a2wxZ05jSWMwanNnU0VzWERnK0JMbXpRSG8KM0pyaVROelkzcXpFOTlTNlBkM0llemNya3hKcSs3WVRvSnNBNGkyWEppdmNMb21DR3lWOCtObi9WRVVzRTJXVQpaQjNiU0d0bG1yWTNEZS9md28zVEpTU2lHK2R0OUtLcUd1dWNwYmxvWkR4ajVXOE9ib2g4NHR0M2QvQzdwYm5qClR3RkxVYjYyN096UFNMcVdiYzdCZEFwaGFZZDQwOVJydWFweUtXY3h4U203a1Q4RGRFbz0KLS0tLS1FTkQgQ0VSVElGSUNBVEUtLS0tLQo=", "encoding": "base64", "item": [ "ca", { "ca": "self-sign", "dns": [ "localhost" ], "name": "quadlet_demo" } ], "source": "/etc/pki/tls/certs/quadlet_demo.crt" } TASK [fedora.linux_system_roles.certificate : Create return data] ************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:160 Tuesday 06 February 2024 16:17:27 +0000 (0:00:00.624) 0:00:13.303 ****** ok: [sut] => { "ansible_facts": { "certificate_test_certs": { "quadlet_demo": { "ca": "/etc/pki/tls/certs/quadlet_demo.crt", "ca_content": "-----BEGIN CERTIFICATE-----\nMIIDgjCCAmqgAwIBAgIQOK2Dzgb1TkW+IA1sAmpKXDANBgkqhkiG9w0BAQsFADBQ\nMSAwHgYDVQQDDBdMb2NhbCBTaWduaW5nIEF1dGhvcml0eTEsMCoGA1UEAwwjMzhh\nZDgzY2UtMDZmNTRlNDUtYmUyMDBkNmMtMDI2YTRhNWIwHhcNMjQwMjA2MTYxNzI2\nWhcNMjUwMjA2MTYxNzI1WjAUMRIwEAYDVQQDEwlsb2NhbGhvc3QwggEiMA0GCSqG\nSIb3DQEBAQUAA4IBDwAwggEKAoIBAQDOgGPVcoBdxhg9ouvSd2quQ6pYNt5CWkc5\nCggS8sMkzBNnZEGOwZZGefwZ454Z3VjGj5GQOlHKfZqPDrqDwTCKTFQDQHnfpscu\n/rQwUBcJJYxg02y6dXzdgW16d1Z74DQZfikMWlTQswtZFJZ663iiVkm/crR+vHs5\nOz/NHCrrsOaXI7NCKo99eOjzOnrG0Flm2yM+LV06WaqkSlcWa5tTiBz7AqOylWE/\nACwrPmCkCT1N8wayeF35XTLKZMx+QR3Iddn6LmgkeG0gTRwdhEsM1IxDua+QzVHO\n12q8Gr1XwWz2cQf6aupelP7uTxIw1lge6gGTkNc8udw16sT3kIsTAgMBAAGjgZMw\ngZAwCwYDVR0PBAQDAgWgMBQGA1UdEQQNMAuCCWxvY2FsaG9zdDAdBgNVHSUEFjAU\nBggrBgEFBQcDAQYIKwYBBQUHAwIwDAYDVR0TAQH/BAIwADAdBgNVHQ4EFgQUyeLW\nnuhu2FOriupgprOKowa/EG4wHwYDVR0jBBgwFoAUv8Yqgvy0LdAzwAEvpuvk0Kk5\nf18wDQYJKoZIhvcNAQELBQADggEBAAKEkeh3sOURIl2CIw0knXgaCteuAyEo3AZ/\nZw7es4ByMGuUcHq2iPjAO85nyz8qLnOQKKWLNm9x1Q/w9nKj55p8vq+HSnybopXw\nCx0sVElP8t5R1m0mT+JFlnebvqH5ygAH2gbQhwkl1gNcIc0jsgSEsXDg+BLmzQHo\n3JriTNzY3qzE99S6Pd3IezcrkxJq+7YToJsA4i2XJivcLomCGyV8+Nn/VEUsE2WU\nZB3bSGtlmrY3De/fwo3TJSSiG+dt9KKqGuucpbloZDxj5W8Oboh84tt3d/C7pbnj\nTwFLUb627OzPSLqWbc7BdAphaYd409RruapyKWcxxSm7kT8DdEo=\n-----END CERTIFICATE-----\n", "cert": "/etc/pki/tls/certs/quadlet_demo.crt", "cert_content": "-----BEGIN CERTIFICATE-----\nMIIDgjCCAmqgAwIBAgIQOK2Dzgb1TkW+IA1sAmpKXDANBgkqhkiG9w0BAQsFADBQ\nMSAwHgYDVQQDDBdMb2NhbCBTaWduaW5nIEF1dGhvcml0eTEsMCoGA1UEAwwjMzhh\nZDgzY2UtMDZmNTRlNDUtYmUyMDBkNmMtMDI2YTRhNWIwHhcNMjQwMjA2MTYxNzI2\nWhcNMjUwMjA2MTYxNzI1WjAUMRIwEAYDVQQDEwlsb2NhbGhvc3QwggEiMA0GCSqG\nSIb3DQEBAQUAA4IBDwAwggEKAoIBAQDOgGPVcoBdxhg9ouvSd2quQ6pYNt5CWkc5\nCggS8sMkzBNnZEGOwZZGefwZ454Z3VjGj5GQOlHKfZqPDrqDwTCKTFQDQHnfpscu\n/rQwUBcJJYxg02y6dXzdgW16d1Z74DQZfikMWlTQswtZFJZ663iiVkm/crR+vHs5\nOz/NHCrrsOaXI7NCKo99eOjzOnrG0Flm2yM+LV06WaqkSlcWa5tTiBz7AqOylWE/\nACwrPmCkCT1N8wayeF35XTLKZMx+QR3Iddn6LmgkeG0gTRwdhEsM1IxDua+QzVHO\n12q8Gr1XwWz2cQf6aupelP7uTxIw1lge6gGTkNc8udw16sT3kIsTAgMBAAGjgZMw\ngZAwCwYDVR0PBAQDAgWgMBQGA1UdEQQNMAuCCWxvY2FsaG9zdDAdBgNVHSUEFjAU\nBggrBgEFBQcDAQYIKwYBBQUHAwIwDAYDVR0TAQH/BAIwADAdBgNVHQ4EFgQUyeLW\nnuhu2FOriupgprOKowa/EG4wHwYDVR0jBBgwFoAUv8Yqgvy0LdAzwAEvpuvk0Kk5\nf18wDQYJKoZIhvcNAQELBQADggEBAAKEkeh3sOURIl2CIw0knXgaCteuAyEo3AZ/\nZw7es4ByMGuUcHq2iPjAO85nyz8qLnOQKKWLNm9x1Q/w9nKj55p8vq+HSnybopXw\nCx0sVElP8t5R1m0mT+JFlnebvqH5ygAH2gbQhwkl1gNcIc0jsgSEsXDg+BLmzQHo\n3JriTNzY3qzE99S6Pd3IezcrkxJq+7YToJsA4i2XJivcLomCGyV8+Nn/VEUsE2WU\nZB3bSGtlmrY3De/fwo3TJSSiG+dt9KKqGuucpbloZDxj5W8Oboh84tt3d/C7pbnj\nTwFLUb627OzPSLqWbc7BdAphaYd409RruapyKWcxxSm7kT8DdEo=\n-----END CERTIFICATE-----\n", "key": "/etc/pki/tls/private/quadlet_demo.key", "key_content": "-----BEGIN PRIVATE KEY-----\nMIIEvQIBADANBgkqhkiG9w0BAQEFAASCBKcwggSjAgEAAoIBAQDOgGPVcoBdxhg9\nouvSd2quQ6pYNt5CWkc5CggS8sMkzBNnZEGOwZZGefwZ454Z3VjGj5GQOlHKfZqP\nDrqDwTCKTFQDQHnfpscu/rQwUBcJJYxg02y6dXzdgW16d1Z74DQZfikMWlTQswtZ\nFJZ663iiVkm/crR+vHs5Oz/NHCrrsOaXI7NCKo99eOjzOnrG0Flm2yM+LV06Waqk\nSlcWa5tTiBz7AqOylWE/ACwrPmCkCT1N8wayeF35XTLKZMx+QR3Iddn6LmgkeG0g\nTRwdhEsM1IxDua+QzVHO12q8Gr1XwWz2cQf6aupelP7uTxIw1lge6gGTkNc8udw1\n6sT3kIsTAgMBAAECggEAIYVg72C3asGB5NuCGXeuFxy/1bMOcEn0XtN8jQrgV6RT\nEAkLxU8yhD1ACbnTy9ING6ppkDaPWWX7RXa4GzQsm36Zcq+wOPh7Q+svt9oAa0kv\nmPYpI7l3rANbELK3lMNH4c2Df1BYjUpLMMoxRUHdIskMDqphl2MkQ52VFUgZxYl2\nrwDFTsnikpMaMgUqJ3DEwCSR45wrxWiBhOmrDcrgZD2iHslrVGsyaXoecM4mPR+E\nvWcsnXcXz1kjmlaAqG1BWzEL1hjFZZAL8Xa2CxwV5S5VfBfF7QZyVmupI1N9aVoh\n9Xg0tlHoMnQP6eeK18OSOx2PjZwv8UNDtkwKV4+VTQKBgQDfLQVBR1yUGqIoGDab\nTv9qepnRG+T3z6CAGrVQMXL6/aBmqeZcvLEhBSFylzM1+qowTTz0/vAohUfHmZN5\n24vOTmo/N2R5b1wbHVpdqQ/rzlIOEbGChOxL4sF+PIM9I1oO/AA+s1JsmBHrrFzf\nzro1qmpMzDIAd2lQDgiM1dUZlQKBgQDs34zUSjWs1OVXZOP8Cnw8r7QdfZcos+HA\nRlgWOU2159yoh8UUg+PSYkKK3266frA7gekSRoTq4cdPQ/g+5KT5/gYwh2jP6P40\nZq0FCZ8Sm39u0hKF2IGjeCvts6+w45jrIe/ZcqLCkH7CTwN+F4DBpJhIc+BPw3Ha\nHsUAtgN4BwKBgQDb/souVkUpbbOpaGzz3gSjluVnn2Ku8cnaXfmbapNYXYmvig2e\nHxdQYHlfNzLxI/5+DLVpIKHmn7NpS7d3qPcA0IWleVoe5qpZxC6fIxehsElsuv5L\nlfJEv+tTIB2St/V/zeQD6sr2Z1VLByaqLGOcoTeWKrejhBKZKqw9AFDHnQKBgHsr\nazFgPTTXp0GgZqRCV4O+4WuMSLOThcnTSi6UMYJrRGQl5SbbxfU5vtwzfoJfKghn\nbuhf3vG74UWoVEdv4RFVM6a8YpXp03UwBrvd2ud0D/lyJsuOXyq/bCg4DR1v2644\ngV2SrGt7c05ag7hnRD5s9fqGtpcDTApdgZbzDi61AoGAM88iJOrW6db06Qnanuos\nQxeVYpIXM7v7TPZbqZn/fsDVZjNXAndACTLCwezEe1u3u5oHsesYdeG5gyWNfBt+\nJjoolJT5p7UlB7h+L/fib3OTbs/zQnuxAA4SNoZNqnTTCVTDARDj/wJywqljAaEy\noNkaHbK282EVxBXnl/N69gw=\n-----END PRIVATE KEY-----\n" } } }, "changed": false } TASK [fedora.linux_system_roles.certificate : Stop tracking certificates] ****** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:176 Tuesday 06 February 2024 16:17:27 +0000 (0:00:00.025) 0:00:13.329 ****** ok: [sut] => (item={'cert': '/etc/pki/tls/certs/quadlet_demo.crt', 'cert_content': '-----BEGIN CERTIFICATE-----\nMIIDgjCCAmqgAwIBAgIQOK2Dzgb1TkW+IA1sAmpKXDANBgkqhkiG9w0BAQsFADBQ\nMSAwHgYDVQQDDBdMb2NhbCBTaWduaW5nIEF1dGhvcml0eTEsMCoGA1UEAwwjMzhh\nZDgzY2UtMDZmNTRlNDUtYmUyMDBkNmMtMDI2YTRhNWIwHhcNMjQwMjA2MTYxNzI2\nWhcNMjUwMjA2MTYxNzI1WjAUMRIwEAYDVQQDEwlsb2NhbGhvc3QwggEiMA0GCSqG\nSIb3DQEBAQUAA4IBDwAwggEKAoIBAQDOgGPVcoBdxhg9ouvSd2quQ6pYNt5CWkc5\nCggS8sMkzBNnZEGOwZZGefwZ454Z3VjGj5GQOlHKfZqPDrqDwTCKTFQDQHnfpscu\n/rQwUBcJJYxg02y6dXzdgW16d1Z74DQZfikMWlTQswtZFJZ663iiVkm/crR+vHs5\nOz/NHCrrsOaXI7NCKo99eOjzOnrG0Flm2yM+LV06WaqkSlcWa5tTiBz7AqOylWE/\nACwrPmCkCT1N8wayeF35XTLKZMx+QR3Iddn6LmgkeG0gTRwdhEsM1IxDua+QzVHO\n12q8Gr1XwWz2cQf6aupelP7uTxIw1lge6gGTkNc8udw16sT3kIsTAgMBAAGjgZMw\ngZAwCwYDVR0PBAQDAgWgMBQGA1UdEQQNMAuCCWxvY2FsaG9zdDAdBgNVHSUEFjAU\nBggrBgEFBQcDAQYIKwYBBQUHAwIwDAYDVR0TAQH/BAIwADAdBgNVHQ4EFgQUyeLW\nnuhu2FOriupgprOKowa/EG4wHwYDVR0jBBgwFoAUv8Yqgvy0LdAzwAEvpuvk0Kk5\nf18wDQYJKoZIhvcNAQELBQADggEBAAKEkeh3sOURIl2CIw0knXgaCteuAyEo3AZ/\nZw7es4ByMGuUcHq2iPjAO85nyz8qLnOQKKWLNm9x1Q/w9nKj55p8vq+HSnybopXw\nCx0sVElP8t5R1m0mT+JFlnebvqH5ygAH2gbQhwkl1gNcIc0jsgSEsXDg+BLmzQHo\n3JriTNzY3qzE99S6Pd3IezcrkxJq+7YToJsA4i2XJivcLomCGyV8+Nn/VEUsE2WU\nZB3bSGtlmrY3De/fwo3TJSSiG+dt9KKqGuucpbloZDxj5W8Oboh84tt3d/C7pbnj\nTwFLUb627OzPSLqWbc7BdAphaYd409RruapyKWcxxSm7kT8DdEo=\n-----END CERTIFICATE-----\n', 'key': '/etc/pki/tls/private/quadlet_demo.key', 'key_content': '-----BEGIN PRIVATE KEY-----\nMIIEvQIBADANBgkqhkiG9w0BAQEFAASCBKcwggSjAgEAAoIBAQDOgGPVcoBdxhg9\nouvSd2quQ6pYNt5CWkc5CggS8sMkzBNnZEGOwZZGefwZ454Z3VjGj5GQOlHKfZqP\nDrqDwTCKTFQDQHnfpscu/rQwUBcJJYxg02y6dXzdgW16d1Z74DQZfikMWlTQswtZ\nFJZ663iiVkm/crR+vHs5Oz/NHCrrsOaXI7NCKo99eOjzOnrG0Flm2yM+LV06Waqk\nSlcWa5tTiBz7AqOylWE/ACwrPmCkCT1N8wayeF35XTLKZMx+QR3Iddn6LmgkeG0g\nTRwdhEsM1IxDua+QzVHO12q8Gr1XwWz2cQf6aupelP7uTxIw1lge6gGTkNc8udw1\n6sT3kIsTAgMBAAECggEAIYVg72C3asGB5NuCGXeuFxy/1bMOcEn0XtN8jQrgV6RT\nEAkLxU8yhD1ACbnTy9ING6ppkDaPWWX7RXa4GzQsm36Zcq+wOPh7Q+svt9oAa0kv\nmPYpI7l3rANbELK3lMNH4c2Df1BYjUpLMMoxRUHdIskMDqphl2MkQ52VFUgZxYl2\nrwDFTsnikpMaMgUqJ3DEwCSR45wrxWiBhOmrDcrgZD2iHslrVGsyaXoecM4mPR+E\nvWcsnXcXz1kjmlaAqG1BWzEL1hjFZZAL8Xa2CxwV5S5VfBfF7QZyVmupI1N9aVoh\n9Xg0tlHoMnQP6eeK18OSOx2PjZwv8UNDtkwKV4+VTQKBgQDfLQVBR1yUGqIoGDab\nTv9qepnRG+T3z6CAGrVQMXL6/aBmqeZcvLEhBSFylzM1+qowTTz0/vAohUfHmZN5\n24vOTmo/N2R5b1wbHVpdqQ/rzlIOEbGChOxL4sF+PIM9I1oO/AA+s1JsmBHrrFzf\nzro1qmpMzDIAd2lQDgiM1dUZlQKBgQDs34zUSjWs1OVXZOP8Cnw8r7QdfZcos+HA\nRlgWOU2159yoh8UUg+PSYkKK3266frA7gekSRoTq4cdPQ/g+5KT5/gYwh2jP6P40\nZq0FCZ8Sm39u0hKF2IGjeCvts6+w45jrIe/ZcqLCkH7CTwN+F4DBpJhIc+BPw3Ha\nHsUAtgN4BwKBgQDb/souVkUpbbOpaGzz3gSjluVnn2Ku8cnaXfmbapNYXYmvig2e\nHxdQYHlfNzLxI/5+DLVpIKHmn7NpS7d3qPcA0IWleVoe5qpZxC6fIxehsElsuv5L\nlfJEv+tTIB2St/V/zeQD6sr2Z1VLByaqLGOcoTeWKrejhBKZKqw9AFDHnQKBgHsr\nazFgPTTXp0GgZqRCV4O+4WuMSLOThcnTSi6UMYJrRGQl5SbbxfU5vtwzfoJfKghn\nbuhf3vG74UWoVEdv4RFVM6a8YpXp03UwBrvd2ud0D/lyJsuOXyq/bCg4DR1v2644\ngV2SrGt7c05ag7hnRD5s9fqGtpcDTApdgZbzDi61AoGAM88iJOrW6db06Qnanuos\nQxeVYpIXM7v7TPZbqZn/fsDVZjNXAndACTLCwezEe1u3u5oHsesYdeG5gyWNfBt+\nJjoolJT5p7UlB7h+L/fib3OTbs/zQnuxAA4SNoZNqnTTCVTDARDj/wJywqljAaEy\noNkaHbK282EVxBXnl/N69gw=\n-----END PRIVATE KEY-----\n', 'ca': '/etc/pki/tls/certs/quadlet_demo.crt', 'ca_content': '-----BEGIN CERTIFICATE-----\nMIIDgjCCAmqgAwIBAgIQOK2Dzgb1TkW+IA1sAmpKXDANBgkqhkiG9w0BAQsFADBQ\nMSAwHgYDVQQDDBdMb2NhbCBTaWduaW5nIEF1dGhvcml0eTEsMCoGA1UEAwwjMzhh\nZDgzY2UtMDZmNTRlNDUtYmUyMDBkNmMtMDI2YTRhNWIwHhcNMjQwMjA2MTYxNzI2\nWhcNMjUwMjA2MTYxNzI1WjAUMRIwEAYDVQQDEwlsb2NhbGhvc3QwggEiMA0GCSqG\nSIb3DQEBAQUAA4IBDwAwggEKAoIBAQDOgGPVcoBdxhg9ouvSd2quQ6pYNt5CWkc5\nCggS8sMkzBNnZEGOwZZGefwZ454Z3VjGj5GQOlHKfZqPDrqDwTCKTFQDQHnfpscu\n/rQwUBcJJYxg02y6dXzdgW16d1Z74DQZfikMWlTQswtZFJZ663iiVkm/crR+vHs5\nOz/NHCrrsOaXI7NCKo99eOjzOnrG0Flm2yM+LV06WaqkSlcWa5tTiBz7AqOylWE/\nACwrPmCkCT1N8wayeF35XTLKZMx+QR3Iddn6LmgkeG0gTRwdhEsM1IxDua+QzVHO\n12q8Gr1XwWz2cQf6aupelP7uTxIw1lge6gGTkNc8udw16sT3kIsTAgMBAAGjgZMw\ngZAwCwYDVR0PBAQDAgWgMBQGA1UdEQQNMAuCCWxvY2FsaG9zdDAdBgNVHSUEFjAU\nBggrBgEFBQcDAQYIKwYBBQUHAwIwDAYDVR0TAQH/BAIwADAdBgNVHQ4EFgQUyeLW\nnuhu2FOriupgprOKowa/EG4wHwYDVR0jBBgwFoAUv8Yqgvy0LdAzwAEvpuvk0Kk5\nf18wDQYJKoZIhvcNAQELBQADggEBAAKEkeh3sOURIl2CIw0knXgaCteuAyEo3AZ/\nZw7es4ByMGuUcHq2iPjAO85nyz8qLnOQKKWLNm9x1Q/w9nKj55p8vq+HSnybopXw\nCx0sVElP8t5R1m0mT+JFlnebvqH5ygAH2gbQhwkl1gNcIc0jsgSEsXDg+BLmzQHo\n3JriTNzY3qzE99S6Pd3IezcrkxJq+7YToJsA4i2XJivcLomCGyV8+Nn/VEUsE2WU\nZB3bSGtlmrY3De/fwo3TJSSiG+dt9KKqGuucpbloZDxj5W8Oboh84tt3d/C7pbnj\nTwFLUb627OzPSLqWbc7BdAphaYd409RruapyKWcxxSm7kT8DdEo=\n-----END CERTIFICATE-----\n'}) => { "ansible_loop_var": "item", "changed": false, "cmd": [ "getcert", "stop-tracking", "-f", "/etc/pki/tls/certs/quadlet_demo.crt" ], "delta": "0:00:00.032357", "end": "2024-02-06 16:17:27.301070", "item": { "ca": "/etc/pki/tls/certs/quadlet_demo.crt", "ca_content": "-----BEGIN CERTIFICATE-----\nMIIDgjCCAmqgAwIBAgIQOK2Dzgb1TkW+IA1sAmpKXDANBgkqhkiG9w0BAQsFADBQ\nMSAwHgYDVQQDDBdMb2NhbCBTaWduaW5nIEF1dGhvcml0eTEsMCoGA1UEAwwjMzhh\nZDgzY2UtMDZmNTRlNDUtYmUyMDBkNmMtMDI2YTRhNWIwHhcNMjQwMjA2MTYxNzI2\nWhcNMjUwMjA2MTYxNzI1WjAUMRIwEAYDVQQDEwlsb2NhbGhvc3QwggEiMA0GCSqG\nSIb3DQEBAQUAA4IBDwAwggEKAoIBAQDOgGPVcoBdxhg9ouvSd2quQ6pYNt5CWkc5\nCggS8sMkzBNnZEGOwZZGefwZ454Z3VjGj5GQOlHKfZqPDrqDwTCKTFQDQHnfpscu\n/rQwUBcJJYxg02y6dXzdgW16d1Z74DQZfikMWlTQswtZFJZ663iiVkm/crR+vHs5\nOz/NHCrrsOaXI7NCKo99eOjzOnrG0Flm2yM+LV06WaqkSlcWa5tTiBz7AqOylWE/\nACwrPmCkCT1N8wayeF35XTLKZMx+QR3Iddn6LmgkeG0gTRwdhEsM1IxDua+QzVHO\n12q8Gr1XwWz2cQf6aupelP7uTxIw1lge6gGTkNc8udw16sT3kIsTAgMBAAGjgZMw\ngZAwCwYDVR0PBAQDAgWgMBQGA1UdEQQNMAuCCWxvY2FsaG9zdDAdBgNVHSUEFjAU\nBggrBgEFBQcDAQYIKwYBBQUHAwIwDAYDVR0TAQH/BAIwADAdBgNVHQ4EFgQUyeLW\nnuhu2FOriupgprOKowa/EG4wHwYDVR0jBBgwFoAUv8Yqgvy0LdAzwAEvpuvk0Kk5\nf18wDQYJKoZIhvcNAQELBQADggEBAAKEkeh3sOURIl2CIw0knXgaCteuAyEo3AZ/\nZw7es4ByMGuUcHq2iPjAO85nyz8qLnOQKKWLNm9x1Q/w9nKj55p8vq+HSnybopXw\nCx0sVElP8t5R1m0mT+JFlnebvqH5ygAH2gbQhwkl1gNcIc0jsgSEsXDg+BLmzQHo\n3JriTNzY3qzE99S6Pd3IezcrkxJq+7YToJsA4i2XJivcLomCGyV8+Nn/VEUsE2WU\nZB3bSGtlmrY3De/fwo3TJSSiG+dt9KKqGuucpbloZDxj5W8Oboh84tt3d/C7pbnj\nTwFLUb627OzPSLqWbc7BdAphaYd409RruapyKWcxxSm7kT8DdEo=\n-----END CERTIFICATE-----\n", "cert": "/etc/pki/tls/certs/quadlet_demo.crt", "cert_content": "-----BEGIN CERTIFICATE-----\nMIIDgjCCAmqgAwIBAgIQOK2Dzgb1TkW+IA1sAmpKXDANBgkqhkiG9w0BAQsFADBQ\nMSAwHgYDVQQDDBdMb2NhbCBTaWduaW5nIEF1dGhvcml0eTEsMCoGA1UEAwwjMzhh\nZDgzY2UtMDZmNTRlNDUtYmUyMDBkNmMtMDI2YTRhNWIwHhcNMjQwMjA2MTYxNzI2\nWhcNMjUwMjA2MTYxNzI1WjAUMRIwEAYDVQQDEwlsb2NhbGhvc3QwggEiMA0GCSqG\nSIb3DQEBAQUAA4IBDwAwggEKAoIBAQDOgGPVcoBdxhg9ouvSd2quQ6pYNt5CWkc5\nCggS8sMkzBNnZEGOwZZGefwZ454Z3VjGj5GQOlHKfZqPDrqDwTCKTFQDQHnfpscu\n/rQwUBcJJYxg02y6dXzdgW16d1Z74DQZfikMWlTQswtZFJZ663iiVkm/crR+vHs5\nOz/NHCrrsOaXI7NCKo99eOjzOnrG0Flm2yM+LV06WaqkSlcWa5tTiBz7AqOylWE/\nACwrPmCkCT1N8wayeF35XTLKZMx+QR3Iddn6LmgkeG0gTRwdhEsM1IxDua+QzVHO\n12q8Gr1XwWz2cQf6aupelP7uTxIw1lge6gGTkNc8udw16sT3kIsTAgMBAAGjgZMw\ngZAwCwYDVR0PBAQDAgWgMBQGA1UdEQQNMAuCCWxvY2FsaG9zdDAdBgNVHSUEFjAU\nBggrBgEFBQcDAQYIKwYBBQUHAwIwDAYDVR0TAQH/BAIwADAdBgNVHQ4EFgQUyeLW\nnuhu2FOriupgprOKowa/EG4wHwYDVR0jBBgwFoAUv8Yqgvy0LdAzwAEvpuvk0Kk5\nf18wDQYJKoZIhvcNAQELBQADggEBAAKEkeh3sOURIl2CIw0knXgaCteuAyEo3AZ/\nZw7es4ByMGuUcHq2iPjAO85nyz8qLnOQKKWLNm9x1Q/w9nKj55p8vq+HSnybopXw\nCx0sVElP8t5R1m0mT+JFlnebvqH5ygAH2gbQhwkl1gNcIc0jsgSEsXDg+BLmzQHo\n3JriTNzY3qzE99S6Pd3IezcrkxJq+7YToJsA4i2XJivcLomCGyV8+Nn/VEUsE2WU\nZB3bSGtlmrY3De/fwo3TJSSiG+dt9KKqGuucpbloZDxj5W8Oboh84tt3d/C7pbnj\nTwFLUb627OzPSLqWbc7BdAphaYd409RruapyKWcxxSm7kT8DdEo=\n-----END CERTIFICATE-----\n", "key": "/etc/pki/tls/private/quadlet_demo.key", "key_content": "-----BEGIN PRIVATE KEY-----\nMIIEvQIBADANBgkqhkiG9w0BAQEFAASCBKcwggSjAgEAAoIBAQDOgGPVcoBdxhg9\nouvSd2quQ6pYNt5CWkc5CggS8sMkzBNnZEGOwZZGefwZ454Z3VjGj5GQOlHKfZqP\nDrqDwTCKTFQDQHnfpscu/rQwUBcJJYxg02y6dXzdgW16d1Z74DQZfikMWlTQswtZ\nFJZ663iiVkm/crR+vHs5Oz/NHCrrsOaXI7NCKo99eOjzOnrG0Flm2yM+LV06Waqk\nSlcWa5tTiBz7AqOylWE/ACwrPmCkCT1N8wayeF35XTLKZMx+QR3Iddn6LmgkeG0g\nTRwdhEsM1IxDua+QzVHO12q8Gr1XwWz2cQf6aupelP7uTxIw1lge6gGTkNc8udw1\n6sT3kIsTAgMBAAECggEAIYVg72C3asGB5NuCGXeuFxy/1bMOcEn0XtN8jQrgV6RT\nEAkLxU8yhD1ACbnTy9ING6ppkDaPWWX7RXa4GzQsm36Zcq+wOPh7Q+svt9oAa0kv\nmPYpI7l3rANbELK3lMNH4c2Df1BYjUpLMMoxRUHdIskMDqphl2MkQ52VFUgZxYl2\nrwDFTsnikpMaMgUqJ3DEwCSR45wrxWiBhOmrDcrgZD2iHslrVGsyaXoecM4mPR+E\nvWcsnXcXz1kjmlaAqG1BWzEL1hjFZZAL8Xa2CxwV5S5VfBfF7QZyVmupI1N9aVoh\n9Xg0tlHoMnQP6eeK18OSOx2PjZwv8UNDtkwKV4+VTQKBgQDfLQVBR1yUGqIoGDab\nTv9qepnRG+T3z6CAGrVQMXL6/aBmqeZcvLEhBSFylzM1+qowTTz0/vAohUfHmZN5\n24vOTmo/N2R5b1wbHVpdqQ/rzlIOEbGChOxL4sF+PIM9I1oO/AA+s1JsmBHrrFzf\nzro1qmpMzDIAd2lQDgiM1dUZlQKBgQDs34zUSjWs1OVXZOP8Cnw8r7QdfZcos+HA\nRlgWOU2159yoh8UUg+PSYkKK3266frA7gekSRoTq4cdPQ/g+5KT5/gYwh2jP6P40\nZq0FCZ8Sm39u0hKF2IGjeCvts6+w45jrIe/ZcqLCkH7CTwN+F4DBpJhIc+BPw3Ha\nHsUAtgN4BwKBgQDb/souVkUpbbOpaGzz3gSjluVnn2Ku8cnaXfmbapNYXYmvig2e\nHxdQYHlfNzLxI/5+DLVpIKHmn7NpS7d3qPcA0IWleVoe5qpZxC6fIxehsElsuv5L\nlfJEv+tTIB2St/V/zeQD6sr2Z1VLByaqLGOcoTeWKrejhBKZKqw9AFDHnQKBgHsr\nazFgPTTXp0GgZqRCV4O+4WuMSLOThcnTSi6UMYJrRGQl5SbbxfU5vtwzfoJfKghn\nbuhf3vG74UWoVEdv4RFVM6a8YpXp03UwBrvd2ud0D/lyJsuOXyq/bCg4DR1v2644\ngV2SrGt7c05ag7hnRD5s9fqGtpcDTApdgZbzDi61AoGAM88iJOrW6db06Qnanuos\nQxeVYpIXM7v7TPZbqZn/fsDVZjNXAndACTLCwezEe1u3u5oHsesYdeG5gyWNfBt+\nJjoolJT5p7UlB7h+L/fib3OTbs/zQnuxAA4SNoZNqnTTCVTDARDj/wJywqljAaEy\noNkaHbK282EVxBXnl/N69gw=\n-----END PRIVATE KEY-----\n" }, "rc": 0, "start": "2024-02-06 16:17:27.268713" } STDOUT: Request "20240206161725" removed. TASK [fedora.linux_system_roles.certificate : Remove files] ******************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:181 Tuesday 06 February 2024 16:17:27 +0000 (0:00:00.294) 0:00:13.623 ****** changed: [sut] => (item=/etc/pki/tls/certs/quadlet_demo.crt) => { "ansible_loop_var": "item", "changed": true, "item": "/etc/pki/tls/certs/quadlet_demo.crt", "path": "/etc/pki/tls/certs/quadlet_demo.crt", "state": "absent" } changed: [sut] => (item=/etc/pki/tls/private/quadlet_demo.key) => { "ansible_loop_var": "item", "changed": true, "item": "/etc/pki/tls/private/quadlet_demo.key", "path": "/etc/pki/tls/private/quadlet_demo.key", "state": "absent" } ok: [sut] => (item=/etc/pki/tls/certs/quadlet_demo.crt) => { "ansible_loop_var": "item", "changed": false, "item": "/etc/pki/tls/certs/quadlet_demo.crt", "path": "/etc/pki/tls/certs/quadlet_demo.crt", "state": "absent" } TASK [Run the role] ************************************************************ task path: /WORKDIR/git-fix-linger-handlingih6nvedg/tests/tests_quadlet_demo.yml:50 Tuesday 06 February 2024 16:17:27 +0000 (0:00:00.592) 0:00:14.216 ****** TASK [fedora.linux_system_roles.podman : Set platform/version specific variables] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:3 Tuesday 06 February 2024 16:17:27 +0000 (0:00:00.032) 0:00:14.248 ****** included: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml for sut TASK [fedora.linux_system_roles.podman : Ensure ansible_facts used by role] **** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:3 Tuesday 06 February 2024 16:17:27 +0000 (0:00:00.016) 0:00:14.265 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_required_facts | difference(ansible_facts.keys() | list) | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if system is ostree] ************ task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:11 Tuesday 06 February 2024 16:17:27 +0000 (0:00:00.020) 0:00:14.285 ****** ok: [sut] => { "changed": false, "stat": { "exists": false } } TASK [fedora.linux_system_roles.podman : Set flag to indicate system is ostree] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:16 Tuesday 06 February 2024 16:17:28 +0000 (0:00:00.192) 0:00:14.478 ****** ok: [sut] => { "ansible_facts": { "__podman_is_ostree": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Set platform/version specific variables] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:20 Tuesday 06 February 2024 16:17:28 +0000 (0:00:00.016) 0:00:14.494 ****** ok: [sut] => (item=RedHat.yml) => { "ansible_facts": { "__podman_packages": [ "podman", "shadow-utils-subid" ] }, "ansible_included_var_files": [ "/WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/vars/RedHat.yml" ], "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml" } skipping: [sut] => (item=CentOS.yml) => { "ansible_loop_var": "item", "changed": false, "false_condition": "__vars_file is file", "item": "CentOS.yml", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=CentOS_9.yml) => { "ansible_loop_var": "item", "changed": false, "false_condition": "__vars_file is file", "item": "CentOS_9.yml", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=CentOS_9.yml) => { "ansible_loop_var": "item", "changed": false, "false_condition": "__vars_file is file", "item": "CentOS_9.yml", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Gather the package facts] ************* task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:6 Tuesday 06 February 2024 16:17:28 +0000 (0:00:00.025) 0:00:14.520 ****** ok: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Enable copr if requested] ************* task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:10 Tuesday 06 February 2024 16:17:29 +0000 (0:00:01.141) 0:00:15.661 ****** skipping: [sut] => { "changed": false, "false_condition": "podman_use_copr | d(false)", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Ensure required packages are installed] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:14 Tuesday 06 February 2024 16:17:29 +0000 (0:00:00.026) 0:00:15.688 ****** changed: [sut] => { "changed": true, "rc": 0, "results": [ "Installed: libslirp-4.4.0-7.el9.x86_64", "Installed: podman-2:4.8.1-1.el9.x86_64", "Installed: netavark-2:1.9.0-1.el9.x86_64", "Installed: slirp4netns-1.2.2-1.el9.x86_64", "Installed: shadow-utils-subid-2:4.9-8.el9.x86_64", "Installed: criu-3.19-1.el9.x86_64", "Installed: yajl-2.1.0-22.el9.x86_64", "Installed: criu-libs-3.19-1.el9.x86_64", "Installed: crun-1.14-1.el9.x86_64", "Installed: conmon-2:2.1.10-1.el9.x86_64", "Installed: aardvark-dns-2:1.9.0-1.el9.x86_64", "Installed: protobuf-c-1.3.3-13.el9.x86_64", "Installed: fuse-overlayfs-1.12-1.el9.x86_64", "Installed: fuse3-3.10.2-6.el9.x86_64", "Installed: fuse3-libs-3.10.2-6.el9.x86_64", "Installed: container-selinux-3:2.229.0-1.el9.noarch", "Installed: libnet-1.2-6.el9.x86_64", "Installed: containers-common-2:1-55.el9.x86_64", "Installed: fuse-common-3.10.2-6.el9.x86_64" ] } TASK [fedora.linux_system_roles.podman : Get podman version] ******************* task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:22 Tuesday 06 February 2024 16:18:20 +0000 (0:00:50.668) 0:01:06.356 ****** ok: [sut] => { "changed": false, "cmd": [ "podman", "--version" ], "delta": "0:00:00.034764", "end": "2024-02-06 16:18:20.263601", "rc": 0, "start": "2024-02-06 16:18:20.228837" } STDOUT: podman version 4.8.1 TASK [fedora.linux_system_roles.podman : Set podman version] ******************* task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:27 Tuesday 06 February 2024 16:18:20 +0000 (0:00:00.236) 0:01:06.593 ****** ok: [sut] => { "ansible_facts": { "podman_version": "4.8.1" }, "changed": false } TASK [fedora.linux_system_roles.podman : Podman package version must be 4.2 or later] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:31 Tuesday 06 February 2024 16:18:20 +0000 (0:00:00.020) 0:01:06.613 ****** skipping: [sut] => { "changed": false, "false_condition": "podman_version is version(\"4.2\", \"<\")", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Podman package version must be 4.4 or later for quadlet, secrets] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:38 Tuesday 06 February 2024 16:18:20 +0000 (0:00:00.020) 0:01:06.634 ****** skipping: [sut] => { "changed": false, "false_condition": "podman_version is version(\"4.4\", \"<\")", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Podman package version must be 4.4 or later for quadlet, secrets] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:48 Tuesday 06 February 2024 16:18:20 +0000 (0:00:00.048) 0:01:06.682 ****** META: end_host conditional evaluated to False, continuing execution for sut skipping: [sut] => { "skip_reason": "end_host conditional evaluated to False, continuing execution for sut" } MSG: end_host conditional evaluated to false, continuing execution for sut TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:55 Tuesday 06 February 2024 16:18:20 +0000 (0:00:00.047) 0:01:06.730 ****** included: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for sut TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Tuesday 06 February 2024 16:18:20 +0000 (0:00:00.039) 0:01:06.769 ****** ok: [sut] => { "ansible_facts": { "getent_passwd": { "root": [ "x", "0", "0", "root", "/root", "/bin/bash" ] } }, "changed": false } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Tuesday 06 February 2024 16:18:20 +0000 (0:00:00.318) 0:01:07.088 ****** skipping: [sut] => { "changed": false, "false_condition": "not ansible_facts[\"getent_passwd\"][__podman_user]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Tuesday 06 February 2024 16:18:20 +0000 (0:00:00.022) 0:01:07.110 ****** ok: [sut] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get group information] **************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:27 Tuesday 06 February 2024 16:18:20 +0000 (0:00:00.027) 0:01:07.138 ****** ok: [sut] => { "ansible_facts": { "getent_group": { "root": [ "x", "0", "" ] } }, "changed": false } TASK [fedora.linux_system_roles.podman : Set group name] *********************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:34 Tuesday 06 February 2024 16:18:21 +0000 (0:00:00.206) 0:01:07.344 ****** ok: [sut] => { "ansible_facts": { "__podman_group_name": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:38 Tuesday 06 February 2024 16:18:21 +0000 (0:00:00.027) 0:01:07.371 ****** ok: [sut] => { "changed": false, "stat": { "atime": 1689166618.0, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "076f89863b474504550ab1b5c11216a77ed7123c", "ctime": 1707236269.7653096, "dev": 51713, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 4471560, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-pie-executable", "mode": "0755", "mtime": 1689166618.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 15496, "uid": 0, "version": "396954581", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************ task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:49 Tuesday 06 February 2024 16:18:21 +0000 (0:00:00.205) 0:01:07.576 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check group with getsubids] *********** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:53 Tuesday 06 February 2024 16:18:21 +0000 (0:00:00.017) 0:01:07.594 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if user is in subuid file] ****** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:63 Tuesday 06 February 2024 16:18:21 +0000 (0:00:00.018) 0:01:07.612 ****** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:71 Tuesday 06 February 2024 16:18:21 +0000 (0:00:00.018) 0:01:07.631 ****** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if group is in subgid file] ***** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:78 Tuesday 06 February 2024 16:18:21 +0000 (0:00:00.019) 0:01:07.650 ****** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] ***** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:86 Tuesday 06 February 2024 16:18:21 +0000 (0:00:00.019) 0:01:07.669 ****** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set config file paths] **************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:61 Tuesday 06 February 2024 16:18:21 +0000 (0:00:00.018) 0:01:07.688 ****** ok: [sut] => { "ansible_facts": { "__podman_container_conf_file": "/etc/containers/containers.conf.d/50-systemroles.conf", "__podman_policy_json_file": "/etc/containers/policy.json", "__podman_registries_conf_file": "/etc/containers/registries.conf.d/50-systemroles.conf", "__podman_storage_conf_file": "/etc/containers/storage.conf" }, "changed": false } TASK [fedora.linux_system_roles.podman : Handle container.conf.d] ************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:70 Tuesday 06 February 2024 16:18:21 +0000 (0:00:00.038) 0:01:07.726 ****** included: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_container_conf_d.yml for sut TASK [fedora.linux_system_roles.podman : Ensure containers.d exists] *********** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_container_conf_d.yml:5 Tuesday 06 February 2024 16:18:21 +0000 (0:00:00.036) 0:01:07.762 ****** skipping: [sut] => { "changed": false, "false_condition": "podman_containers_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Update container config file] ********* task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_container_conf_d.yml:13 Tuesday 06 February 2024 16:18:21 +0000 (0:00:00.019) 0:01:07.782 ****** skipping: [sut] => { "changed": false, "false_condition": "podman_containers_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Handle registries.conf.d] ************* task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:73 Tuesday 06 February 2024 16:18:21 +0000 (0:00:00.024) 0:01:07.806 ****** included: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_registries_conf_d.yml for sut TASK [fedora.linux_system_roles.podman : Ensure registries.d exists] *********** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_registries_conf_d.yml:5 Tuesday 06 February 2024 16:18:21 +0000 (0:00:00.036) 0:01:07.842 ****** skipping: [sut] => { "changed": false, "false_condition": "podman_registries_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Update registries config file] ******** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_registries_conf_d.yml:13 Tuesday 06 February 2024 16:18:21 +0000 (0:00:00.019) 0:01:07.861 ****** skipping: [sut] => { "changed": false, "false_condition": "podman_registries_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Handle storage.conf] ****************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:76 Tuesday 06 February 2024 16:18:21 +0000 (0:00:00.018) 0:01:07.879 ****** included: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_storage_conf.yml for sut TASK [fedora.linux_system_roles.podman : Ensure storage.conf parent dir exists] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_storage_conf.yml:5 Tuesday 06 February 2024 16:18:21 +0000 (0:00:00.057) 0:01:07.937 ****** skipping: [sut] => { "changed": false, "false_condition": "podman_storage_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Update storage config file] *********** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_storage_conf.yml:13 Tuesday 06 February 2024 16:18:21 +0000 (0:00:00.019) 0:01:07.957 ****** skipping: [sut] => { "changed": false, "false_condition": "podman_storage_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Handle policy.json] ******************* task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:79 Tuesday 06 February 2024 16:18:21 +0000 (0:00:00.019) 0:01:07.976 ****** included: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml for sut TASK [fedora.linux_system_roles.podman : Ensure policy.json parent dir exists] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:6 Tuesday 06 February 2024 16:18:21 +0000 (0:00:00.038) 0:01:08.015 ****** skipping: [sut] => { "changed": false, "false_condition": "podman_policy_json | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Stat the policy.json file] ************ task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:14 Tuesday 06 February 2024 16:18:21 +0000 (0:00:00.019) 0:01:08.035 ****** skipping: [sut] => { "changed": false, "false_condition": "podman_policy_json | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get the existing policy.json] ********* task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:19 Tuesday 06 February 2024 16:18:21 +0000 (0:00:00.027) 0:01:08.062 ****** skipping: [sut] => { "changed": false, "false_condition": "podman_policy_json | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Write new policy.json file] *********** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:25 Tuesday 06 February 2024 16:18:21 +0000 (0:00:00.024) 0:01:08.087 ****** skipping: [sut] => { "changed": false, "false_condition": "podman_policy_json | length > 0", "skip_reason": "Conditional result was False" } TASK [Manage firewall for specified ports] ************************************* task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:85 Tuesday 06 February 2024 16:18:21 +0000 (0:00:00.019) 0:01:08.106 ****** TASK [fedora.linux_system_roles.firewall : Setup firewalld] ******************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:2 Tuesday 06 February 2024 16:18:21 +0000 (0:00:00.064) 0:01:08.170 ****** included: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml for sut TASK [fedora.linux_system_roles.firewall : Ensure ansible_facts used by role] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:2 Tuesday 06 February 2024 16:18:21 +0000 (0:00:00.031) 0:01:08.201 ****** skipping: [sut] => { "changed": false, "false_condition": "__firewall_required_facts | difference(ansible_facts.keys() | list) | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Check if system is ostree] ********** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:10 Tuesday 06 February 2024 16:18:21 +0000 (0:00:00.023) 0:01:08.225 ****** ok: [sut] => { "changed": false, "stat": { "exists": false } } TASK [fedora.linux_system_roles.firewall : Set flag to indicate system is ostree] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:15 Tuesday 06 February 2024 16:18:22 +0000 (0:00:00.195) 0:01:08.420 ****** ok: [sut] => { "ansible_facts": { "__firewall_is_ostree": false }, "changed": false } TASK [fedora.linux_system_roles.firewall : Install firewalld] ****************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:19 Tuesday 06 February 2024 16:18:22 +0000 (0:00:00.022) 0:01:08.443 ****** ok: [sut] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [fedora.linux_system_roles.firewall : Collect service facts] ************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:5 Tuesday 06 February 2024 16:18:23 +0000 (0:00:00.922) 0:01:09.365 ****** skipping: [sut] => { "changed": false, "false_condition": "firewall_disable_conflicting_services | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Attempt to stop and disable conflicting services] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:9 Tuesday 06 February 2024 16:18:23 +0000 (0:00:00.018) 0:01:09.384 ****** skipping: [sut] => (item=nftables) => { "ansible_loop_var": "item", "changed": false, "false_condition": "firewall_disable_conflicting_services | bool", "item": "nftables", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=iptables) => { "ansible_loop_var": "item", "changed": false, "false_condition": "firewall_disable_conflicting_services | bool", "item": "iptables", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=ufw) => { "ansible_loop_var": "item", "changed": false, "false_condition": "firewall_disable_conflicting_services | bool", "item": "ufw", "skip_reason": "Conditional result was False" } skipping: [sut] => { "changed": false } MSG: All items skipped TASK [fedora.linux_system_roles.firewall : Unmask firewalld service] *********** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:22 Tuesday 06 February 2024 16:18:23 +0000 (0:00:00.022) 0:01:09.406 ****** ok: [sut] => { "changed": false, "name": "firewalld", "status": { "AccessSELinuxContext": "system_u:object_r:firewalld_unit_file_t:s0", "ActiveEnterTimestampMonotonic": "0", "ActiveExitTimestampMonotonic": "0", "ActiveState": "inactive", "After": "dbus-broker.service basic.target dbus.socket system.slice polkit.service sysinit.target", "AllowIsolate": "no", "AssertResult": "no", "AssertTimestampMonotonic": "0", "Before": "network-pre.target shutdown.target", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "BusName": "org.fedoraproject.FirewallD1", "CPUAccounting": "yes", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "[not set]", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "yes", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore", "CleanResult": "success", "CollectMode": "inactive", "ConditionResult": "no", "ConditionTimestampMonotonic": "0", "ConfigurationDirectoryMode": "0755", "Conflicts": "ip6tables.service ipset.service shutdown.target nftables.service ebtables.service iptables.service", "ControlGroupId": "0", "ControlPID": "0", "CoredumpFilter": "0x33", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "Delegate": "no", "Description": "firewalld - dynamic firewall daemon", "DevicePolicy": "auto", "Documentation": "\"man:firewalld(1)\"", "DynamicUser": "no", "EnvironmentFiles": "/etc/sysconfig/firewalld (ignore_errors=yes)", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "0", "ExecMainStartTimestampMonotonic": "0", "ExecMainStatus": "0", "ExecReload": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecReloadEx": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStart": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStartEx": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExitType": "main", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FinalKillSignal": "9", "FragmentPath": "/usr/lib/systemd/system/firewalld.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOReadBytes": "18446744073709551615", "IOReadOperations": "18446744073709551615", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "IOWeight": "[not set]", "IOWriteBytes": "18446744073709551615", "IOWriteOperations": "18446744073709551615", "IPAccounting": "no", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "Id": "firewalld.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestampMonotonic": "0", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "mixed", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "0", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "8388608", "LimitMEMLOCKSoft": "8388608", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "524288", "LimitNOFILESoft": "1024", "LimitNPROC": "13974", "LimitNPROCSoft": "13974", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "13974", "LimitSIGPENDINGSoft": "13974", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "0", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "ManagedOOMSwap": "auto", "MemoryAccounting": "yes", "MemoryAvailable": "infinity", "MemoryCurrent": "[not set]", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemorySwapMax": "infinity", "MountAPIVFS": "no", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAPolicy": "n/a", "Names": "firewalld.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMPolicy": "stop", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "OnSuccessJobMode": "fail", "Perpetual": "no", "PrivateDevices": "no", "PrivateIPC": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProcSubset": "all", "ProtectClock": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectHostname": "no", "ProtectKernelLogs": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectProc": "default", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "ReloadResult": "success", "ReloadSignal": "1", "RemainAfterExit": "no", "RemoveIPC": "no", "Requires": "sysinit.target system.slice dbus.socket", "Restart": "no", "RestartKillSignal": "15", "RestartUSec": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "Slice": "system.slice", "StandardError": "null", "StandardInput": "null", "StandardOutput": "null", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StateChangeTimestampMonotonic": "0", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "dead", "SuccessAction": "none", "SyslogFacility": "3", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "2147483646", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "[not set]", "TasksMax": "22359", "TimeoutAbortUSec": "1min 30s", "TimeoutCleanUSec": "infinity", "TimeoutStartFailureMode": "terminate", "TimeoutStartUSec": "1min 30s", "TimeoutStopFailureMode": "terminate", "TimeoutStopUSec": "1min 30s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "dbus", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "enabled", "UnitFileState": "disabled", "UtmpMode": "init", "Wants": "network-pre.target", "WatchdogSignal": "6", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "infinity" } } TASK [fedora.linux_system_roles.firewall : Enable and start firewalld service] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:28 Tuesday 06 February 2024 16:18:23 +0000 (0:00:00.362) 0:01:09.768 ****** changed: [sut] => { "changed": true, "enabled": true, "name": "firewalld", "state": "started", "status": { "AccessSELinuxContext": "system_u:object_r:firewalld_unit_file_t:s0", "ActiveEnterTimestampMonotonic": "0", "ActiveExitTimestampMonotonic": "0", "ActiveState": "inactive", "After": "polkit.service sysinit.target system.slice dbus-broker.service basic.target dbus.socket", "AllowIsolate": "no", "AssertResult": "no", "AssertTimestampMonotonic": "0", "Before": "network-pre.target shutdown.target", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "BusName": "org.fedoraproject.FirewallD1", "CPUAccounting": "yes", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "[not set]", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "yes", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore", "CleanResult": "success", "CollectMode": "inactive", "ConditionResult": "no", "ConditionTimestampMonotonic": "0", "ConfigurationDirectoryMode": "0755", "Conflicts": "iptables.service ip6tables.service ebtables.service shutdown.target ipset.service nftables.service", "ControlGroupId": "0", "ControlPID": "0", "CoredumpFilter": "0x33", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "Delegate": "no", "Description": "firewalld - dynamic firewall daemon", "DevicePolicy": "auto", "Documentation": "\"man:firewalld(1)\"", "DynamicUser": "no", "EnvironmentFiles": "/etc/sysconfig/firewalld (ignore_errors=yes)", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "0", "ExecMainStartTimestampMonotonic": "0", "ExecMainStatus": "0", "ExecReload": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecReloadEx": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStart": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStartEx": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExitType": "main", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FinalKillSignal": "9", "FragmentPath": "/usr/lib/systemd/system/firewalld.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOReadBytes": "18446744073709551615", "IOReadOperations": "18446744073709551615", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "IOWeight": "[not set]", "IOWriteBytes": "18446744073709551615", "IOWriteOperations": "18446744073709551615", "IPAccounting": "no", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "Id": "firewalld.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestampMonotonic": "0", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "mixed", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "0", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "8388608", "LimitMEMLOCKSoft": "8388608", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "524288", "LimitNOFILESoft": "1024", "LimitNPROC": "13974", "LimitNPROCSoft": "13974", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "13974", "LimitSIGPENDINGSoft": "13974", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "0", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "ManagedOOMSwap": "auto", "MemoryAccounting": "yes", "MemoryAvailable": "infinity", "MemoryCurrent": "[not set]", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemorySwapMax": "infinity", "MountAPIVFS": "no", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAPolicy": "n/a", "Names": "firewalld.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMPolicy": "stop", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "OnSuccessJobMode": "fail", "Perpetual": "no", "PrivateDevices": "no", "PrivateIPC": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProcSubset": "all", "ProtectClock": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectHostname": "no", "ProtectKernelLogs": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectProc": "default", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "ReloadResult": "success", "ReloadSignal": "1", "RemainAfterExit": "no", "RemoveIPC": "no", "Requires": "system.slice sysinit.target dbus.socket", "Restart": "no", "RestartKillSignal": "15", "RestartUSec": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "Slice": "system.slice", "StandardError": "null", "StandardInput": "null", "StandardOutput": "null", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StateChangeTimestampMonotonic": "0", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "dead", "SuccessAction": "none", "SyslogFacility": "3", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "2147483646", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "[not set]", "TasksMax": "22359", "TimeoutAbortUSec": "1min 30s", "TimeoutCleanUSec": "infinity", "TimeoutStartFailureMode": "terminate", "TimeoutStartUSec": "1min 30s", "TimeoutStopFailureMode": "terminate", "TimeoutStopUSec": "1min 30s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "dbus", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "enabled", "UnitFileState": "disabled", "UtmpMode": "init", "Wants": "network-pre.target", "WatchdogSignal": "6", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "infinity" } } TASK [fedora.linux_system_roles.firewall : Check if previous replaced is defined] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:34 Tuesday 06 February 2024 16:18:25 +0000 (0:00:01.722) 0:01:11.491 ****** ok: [sut] => { "ansible_facts": { "__firewall_previous_replaced": false, "__firewall_python_cmd": "/usr/bin/python3", "__firewall_report_changed": true }, "changed": false } TASK [fedora.linux_system_roles.firewall : Get config files, checksums before and remove] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:43 Tuesday 06 February 2024 16:18:25 +0000 (0:00:00.027) 0:01:11.518 ****** skipping: [sut] => { "changed": false, "false_condition": "__firewall_previous_replaced | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Tell firewall module it is able to report changed] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:55 Tuesday 06 February 2024 16:18:25 +0000 (0:00:00.019) 0:01:11.537 ****** skipping: [sut] => { "changed": false, "false_condition": "__firewall_previous_replaced | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Configure firewall] ***************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:71 Tuesday 06 February 2024 16:18:25 +0000 (0:00:00.019) 0:01:11.556 ****** changed: [sut] => (item={'port': '8000/tcp', 'state': 'enabled'}) => { "__firewall_changed": true, "ansible_loop_var": "item", "changed": true, "item": { "port": "8000/tcp", "state": "enabled" } } changed: [sut] => (item={'port': '9000/tcp', 'state': 'enabled'}) => { "__firewall_changed": true, "ansible_loop_var": "item", "changed": true, "item": { "port": "9000/tcp", "state": "enabled" } } TASK [fedora.linux_system_roles.firewall : Gather firewall config information] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:120 Tuesday 06 February 2024 16:18:26 +0000 (0:00:01.724) 0:01:13.280 ****** skipping: [sut] => (item={'port': '8000/tcp', 'state': 'enabled'}) => { "ansible_loop_var": "item", "changed": false, "false_condition": "firewall | length == 1", "item": { "port": "8000/tcp", "state": "enabled" }, "skip_reason": "Conditional result was False" } skipping: [sut] => (item={'port': '9000/tcp', 'state': 'enabled'}) => { "ansible_loop_var": "item", "changed": false, "false_condition": "firewall | length == 1", "item": { "port": "9000/tcp", "state": "enabled" }, "skip_reason": "Conditional result was False" } skipping: [sut] => { "changed": false } MSG: All items skipped TASK [fedora.linux_system_roles.firewall : Update firewalld_config fact] ******* task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:130 Tuesday 06 February 2024 16:18:27 +0000 (0:00:00.030) 0:01:13.311 ****** skipping: [sut] => { "changed": false, "false_condition": "firewall | length == 1", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Gather firewall config if no arguments] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:139 Tuesday 06 February 2024 16:18:27 +0000 (0:00:00.020) 0:01:13.331 ****** skipping: [sut] => { "changed": false, "false_condition": "firewall == None or firewall | length == 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Update firewalld_config fact] ******* task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:144 Tuesday 06 February 2024 16:18:27 +0000 (0:00:00.023) 0:01:13.355 ****** skipping: [sut] => { "changed": false, "false_condition": "firewall == None or firewall | length == 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Get config files, checksums after] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:153 Tuesday 06 February 2024 16:18:27 +0000 (0:00:00.036) 0:01:13.392 ****** skipping: [sut] => { "changed": false, "false_condition": "__firewall_previous_replaced | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Calculate what has changed] ********* task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:163 Tuesday 06 February 2024 16:18:27 +0000 (0:00:00.027) 0:01:13.419 ****** skipping: [sut] => { "changed": false, "false_condition": "__firewall_previous_replaced | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Show diffs] ************************* task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:169 Tuesday 06 February 2024 16:18:27 +0000 (0:00:00.036) 0:01:13.456 ****** skipping: [sut] => { "false_condition": "__firewall_previous_replaced | bool" } TASK [Manage selinux for specified ports] ************************************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:92 Tuesday 06 February 2024 16:18:27 +0000 (0:00:00.056) 0:01:13.512 ****** skipping: [sut] => { "changed": false, "false_condition": "podman_selinux_ports | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Keep track of users that need to cancel linger] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:99 Tuesday 06 February 2024 16:18:27 +0000 (0:00:00.040) 0:01:13.553 ****** ok: [sut] => { "ansible_facts": { "__podman_cancel_user_linger": [] }, "changed": false } TASK [fedora.linux_system_roles.podman : Handle secrets] *********************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:103 Tuesday 06 February 2024 16:18:27 +0000 (0:00:00.039) 0:01:13.592 ****** included: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml for sut => (item=(censored due to no_log)) included: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml for sut => (item=(censored due to no_log)) included: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml for sut => (item=(censored due to no_log)) TASK [fedora.linux_system_roles.podman : Set variables part 0] ***************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:3 Tuesday 06 February 2024 16:18:27 +0000 (0:00:00.133) 0:01:13.725 ****** ok: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Set variables part 1] ***************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:10 Tuesday 06 February 2024 16:18:27 +0000 (0:00:00.048) 0:01:13.774 ****** ok: [sut] => { "ansible_facts": { "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set variables part 2] ***************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:14 Tuesday 06 February 2024 16:18:27 +0000 (0:00:00.043) 0:01:13.817 ****** ok: [sut] => { "ansible_facts": { "__podman_rootless": false, "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:20 Tuesday 06 February 2024 16:18:27 +0000 (0:00:00.035) 0:01:13.853 ****** included: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for sut TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Tuesday 06 February 2024 16:18:27 +0000 (0:00:00.079) 0:01:13.933 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Tuesday 06 February 2024 16:18:27 +0000 (0:00:00.031) 0:01:13.964 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Tuesday 06 February 2024 16:18:27 +0000 (0:00:00.028) 0:01:13.993 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Stat XDG_RUNTIME_DIR] ***************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:25 Tuesday 06 February 2024 16:18:27 +0000 (0:00:00.035) 0:01:14.028 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Manage each secret] ******************* task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:36 Tuesday 06 February 2024 16:18:27 +0000 (0:00:00.037) 0:01:14.066 ****** ok: [sut] => { "changed": false } TASK [fedora.linux_system_roles.podman : Debug] ******************************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:53 Tuesday 06 February 2024 16:18:28 +0000 (0:00:00.404) 0:01:14.471 ****** changed: [sut] => { "changed": true, "cmd": [ "podman", "secret", "ls" ], "delta": "0:00:00.180260", "end": "2024-02-06 16:18:28.525229", "rc": 0, "start": "2024-02-06 16:18:28.344969" } STDOUT: ID NAME DRIVER CREATED UPDATED TASK [fedora.linux_system_roles.podman : Set variables part 0] ***************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:3 Tuesday 06 February 2024 16:18:28 +0000 (0:00:00.396) 0:01:14.867 ****** ok: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Set variables part 1] ***************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:10 Tuesday 06 February 2024 16:18:28 +0000 (0:00:00.044) 0:01:14.912 ****** ok: [sut] => { "ansible_facts": { "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set variables part 2] ***************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:14 Tuesday 06 February 2024 16:18:28 +0000 (0:00:00.027) 0:01:14.940 ****** ok: [sut] => { "ansible_facts": { "__podman_rootless": false, "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:20 Tuesday 06 February 2024 16:18:28 +0000 (0:00:00.028) 0:01:14.968 ****** included: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for sut TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Tuesday 06 February 2024 16:18:28 +0000 (0:00:00.030) 0:01:14.998 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Tuesday 06 February 2024 16:18:28 +0000 (0:00:00.018) 0:01:15.017 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Tuesday 06 February 2024 16:18:28 +0000 (0:00:00.020) 0:01:15.037 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Stat XDG_RUNTIME_DIR] ***************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:25 Tuesday 06 February 2024 16:18:28 +0000 (0:00:00.019) 0:01:15.057 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Manage each secret] ******************* task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:36 Tuesday 06 February 2024 16:18:28 +0000 (0:00:00.020) 0:01:15.077 ****** ok: [sut] => { "changed": false } TASK [fedora.linux_system_roles.podman : Debug] ******************************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:53 Tuesday 06 February 2024 16:18:29 +0000 (0:00:00.248) 0:01:15.326 ****** changed: [sut] => { "changed": true, "cmd": [ "podman", "secret", "ls" ], "delta": "0:00:00.034067", "end": "2024-02-06 16:18:29.233368", "rc": 0, "start": "2024-02-06 16:18:29.199301" } STDOUT: ID NAME DRIVER CREATED UPDATED TASK [fedora.linux_system_roles.podman : Set variables part 0] ***************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:3 Tuesday 06 February 2024 16:18:29 +0000 (0:00:00.238) 0:01:15.564 ****** ok: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Set variables part 1] ***************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:10 Tuesday 06 February 2024 16:18:29 +0000 (0:00:00.023) 0:01:15.587 ****** ok: [sut] => { "ansible_facts": { "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set variables part 2] ***************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:14 Tuesday 06 February 2024 16:18:29 +0000 (0:00:00.023) 0:01:15.611 ****** ok: [sut] => { "ansible_facts": { "__podman_rootless": false, "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:20 Tuesday 06 February 2024 16:18:29 +0000 (0:00:00.026) 0:01:15.637 ****** included: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for sut TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Tuesday 06 February 2024 16:18:29 +0000 (0:00:00.074) 0:01:15.711 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Tuesday 06 February 2024 16:18:29 +0000 (0:00:00.019) 0:01:15.730 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Tuesday 06 February 2024 16:18:29 +0000 (0:00:00.019) 0:01:15.750 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Stat XDG_RUNTIME_DIR] ***************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:25 Tuesday 06 February 2024 16:18:29 +0000 (0:00:00.020) 0:01:15.771 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Manage each secret] ******************* task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:36 Tuesday 06 February 2024 16:18:29 +0000 (0:00:00.019) 0:01:15.790 ****** ok: [sut] => { "changed": false } TASK [fedora.linux_system_roles.podman : Debug] ******************************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:53 Tuesday 06 February 2024 16:18:29 +0000 (0:00:00.251) 0:01:16.042 ****** changed: [sut] => { "changed": true, "cmd": [ "podman", "secret", "ls" ], "delta": "0:00:00.038141", "end": "2024-02-06 16:18:29.954293", "rc": 0, "start": "2024-02-06 16:18:29.916152" } STDOUT: ID NAME DRIVER CREATED UPDATED TASK [fedora.linux_system_roles.podman : Handle Kubernetes specifications] ***** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:110 Tuesday 06 February 2024 16:18:29 +0000 (0:00:00.241) 0:01:16.283 ****** skipping: [sut] => { "changed": false, "skipped_reason": "No items in the list" } TASK [fedora.linux_system_roles.podman : Handle Quadlet specifications] ******** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:116 Tuesday 06 February 2024 16:18:30 +0000 (0:00:00.018) 0:01:16.302 ****** included: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for sut => (item={'file_src': 'quadlet-demo.network'}) included: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for sut => (item={'file_src': 'quadlet-demo-mysql.volume'}) included: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for sut => (item={'template_src': 'quadlet-demo-mysql.container.j2'}) included: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for sut => (item={'file_src': 'envoy-proxy-configmap.yml'}) included: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for sut => (item={'template_src': 'quadlet-demo.yml.j2'}) included: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for sut => (item={'file_src': 'quadlet-demo.kube'}) TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14 Tuesday 06 February 2024 16:18:30 +0000 (0:00:00.087) 0:01:16.389 ****** ok: [sut] => { "ansible_facts": { "__podman_quadlet_file_src": "quadlet-demo.network", "__podman_quadlet_spec": {}, "__podman_quadlet_str": "[Network]\nSubnet=192.168.30.0/24\nGateway=192.168.30.1\nLabel=app=wordpress", "__podman_quadlet_template_src": "" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25 Tuesday 06 February 2024 16:18:30 +0000 (0:00:00.031) 0:01:16.421 ****** ok: [sut] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_pull_image": true, "__podman_state": "created", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] ***** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35 Tuesday 06 February 2024 16:18:30 +0000 (0:00:00.025) 0:01:16.446 ****** skipping: [sut] => { "changed": false, "false_condition": "not __podman_quadlet_file_src", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48 Tuesday 06 February 2024 16:18:30 +0000 (0:00:00.021) 0:01:16.468 ****** ok: [sut] => { "ansible_facts": { "__podman_quadlet_name": "quadlet-demo", "__podman_quadlet_type": "network", "__podman_rootless": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57 Tuesday 06 February 2024 16:18:30 +0000 (0:00:00.030) 0:01:16.498 ****** included: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for sut TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Tuesday 06 February 2024 16:18:30 +0000 (0:00:00.036) 0:01:16.534 ****** skipping: [sut] => { "changed": false, "false_condition": "'getent_passwd' not in ansible_facts or __podman_user not in ansible_facts['getent_passwd']", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Tuesday 06 February 2024 16:18:30 +0000 (0:00:00.022) 0:01:16.556 ****** skipping: [sut] => { "changed": false, "false_condition": "not ansible_facts[\"getent_passwd\"][__podman_user]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Tuesday 06 February 2024 16:18:30 +0000 (0:00:00.022) 0:01:16.579 ****** ok: [sut] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get group information] **************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:27 Tuesday 06 February 2024 16:18:30 +0000 (0:00:00.026) 0:01:16.606 ****** ok: [sut] => { "ansible_facts": { "getent_group": { "root": [ "x", "0", "" ] } }, "changed": false } TASK [fedora.linux_system_roles.podman : Set group name] *********************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:34 Tuesday 06 February 2024 16:18:30 +0000 (0:00:00.205) 0:01:16.812 ****** ok: [sut] => { "ansible_facts": { "__podman_group_name": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:38 Tuesday 06 February 2024 16:18:30 +0000 (0:00:00.105) 0:01:16.917 ****** ok: [sut] => { "changed": false, "stat": { "atime": 1707236301.2420683, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "076f89863b474504550ab1b5c11216a77ed7123c", "ctime": 1707236269.7653096, "dev": 51713, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 4471560, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-pie-executable", "mode": "0755", "mtime": 1689166618.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 15496, "uid": 0, "version": "396954581", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************ task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:49 Tuesday 06 February 2024 16:18:30 +0000 (0:00:00.203) 0:01:17.120 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check group with getsubids] *********** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:53 Tuesday 06 February 2024 16:18:30 +0000 (0:00:00.019) 0:01:17.140 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if user is in subuid file] ****** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:63 Tuesday 06 February 2024 16:18:30 +0000 (0:00:00.019) 0:01:17.160 ****** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:71 Tuesday 06 February 2024 16:18:30 +0000 (0:00:00.019) 0:01:17.180 ****** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if group is in subgid file] ***** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:78 Tuesday 06 February 2024 16:18:30 +0000 (0:00:00.019) 0:01:17.199 ****** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] ***** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:86 Tuesday 06 February 2024 16:18:30 +0000 (0:00:00.018) 0:01:17.218 ****** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62 Tuesday 06 February 2024 16:18:30 +0000 (0:00:00.018) 0:01:17.237 ****** ok: [sut] => { "ansible_facts": { "__podman_activate_systemd_unit": true, "__podman_images_found": [], "__podman_kube_yamls_raw": "", "__podman_service_name": "", "__podman_systemd_scope": "system", "__podman_user_home_dir": "/root", "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73 Tuesday 06 February 2024 16:18:30 +0000 (0:00:00.038) 0:01:17.275 ****** ok: [sut] => { "ansible_facts": { "__podman_quadlet_path": "/etc/containers/systemd" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get kube yaml contents] *************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77 Tuesday 06 February 2024 16:18:30 +0000 (0:00:00.021) 0:01:17.297 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_kube_yamls_raw | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87 Tuesday 06 February 2024 16:18:31 +0000 (0:00:00.020) 0:01:17.317 ****** ok: [sut] => { "ansible_facts": { "__podman_images": [], "__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo.network", "__podman_volumes": [] }, "changed": false } TASK [fedora.linux_system_roles.podman : Cleanup quadlets] ********************* task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:103 Tuesday 06 February 2024 16:18:31 +0000 (0:00:00.048) 0:01:17.365 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_state == \"absent\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create and update quadlets] *********** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:107 Tuesday 06 February 2024 16:18:31 +0000 (0:00:00.020) 0:01:17.386 ****** included: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml for sut TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:2 Tuesday 06 February 2024 16:18:31 +0000 (0:00:00.041) 0:01:17.427 ****** included: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for sut TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Tuesday 06 February 2024 16:18:31 +0000 (0:00:00.035) 0:01:17.462 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Tuesday 06 February 2024 16:18:31 +0000 (0:00:00.021) 0:01:17.484 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Tuesday 06 February 2024 16:18:31 +0000 (0:00:00.020) 0:01:17.504 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create host directories] ************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:7 Tuesday 06 February 2024 16:18:31 +0000 (0:00:00.020) 0:01:17.525 ****** skipping: [sut] => { "changed": false, "skipped_reason": "No items in the list" } TASK [fedora.linux_system_roles.podman : Ensure container images are present] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:20 Tuesday 06 February 2024 16:18:31 +0000 (0:00:00.018) 0:01:17.544 ****** skipping: [sut] => { "changed": false, "skipped_reason": "No items in the list" } TASK [fedora.linux_system_roles.podman : Ensure the quadlet directory is present] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:37 Tuesday 06 February 2024 16:18:31 +0000 (0:00:00.055) 0:01:17.599 ****** ok: [sut] => { "changed": false, "gid": 0, "group": "root", "mode": "0755", "owner": "root", "path": "/etc/containers/systemd", "secontext": "system_u:object_r:etc_t:s0", "size": 6, "state": "directory", "uid": 0 } TASK [fedora.linux_system_roles.podman : Ensure quadlet file is copied] ******** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:46 Tuesday 06 February 2024 16:18:31 +0000 (0:00:00.212) 0:01:17.812 ****** changed: [sut] => { "changed": true, "checksum": "e57c08d49aff4bae8daab138d913aeddaa8682a0", "dest": "/etc/containers/systemd/quadlet-demo.network", "gid": 0, "group": "root", "md5sum": "061f3cf318cbd8ab5794bb1173831fb8", "mode": "0644", "owner": "root", "secontext": "system_u:object_r:etc_t:s0", "size": 74, "src": "/root/.ansible/tmp/ansible-tmp-1707236311.5468745-8080-141066871357618/source", "state": "file", "uid": 0 } TASK [fedora.linux_system_roles.podman : Ensure quadlet file content is present] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:56 Tuesday 06 February 2024 16:18:32 +0000 (0:00:00.580) 0:01:18.393 ****** skipping: [sut] => { "changed": false, "false_condition": "not __podman_quadlet_file_src", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Ensure quadlet file is present] ******* task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:68 Tuesday 06 February 2024 16:18:32 +0000 (0:00:00.020) 0:01:18.414 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_copy_file is skipped", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Reload systemctl] ********************* task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:80 Tuesday 06 February 2024 16:18:32 +0000 (0:00:00.020) 0:01:18.434 ****** ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.podman : Start service] ************************ task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:108 Tuesday 06 February 2024 16:18:32 +0000 (0:00:00.528) 0:01:18.963 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_service_name | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Restart service] ********************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:123 Tuesday 06 February 2024 16:18:32 +0000 (0:00:00.023) 0:01:18.987 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_service_name | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14 Tuesday 06 February 2024 16:18:32 +0000 (0:00:00.022) 0:01:19.009 ****** ok: [sut] => { "ansible_facts": { "__podman_quadlet_file_src": "quadlet-demo-mysql.volume", "__podman_quadlet_spec": {}, "__podman_quadlet_str": "[Volume]", "__podman_quadlet_template_src": "" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25 Tuesday 06 February 2024 16:18:32 +0000 (0:00:00.030) 0:01:19.040 ****** ok: [sut] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_pull_image": true, "__podman_state": "created", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] ***** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35 Tuesday 06 February 2024 16:18:32 +0000 (0:00:00.027) 0:01:19.067 ****** skipping: [sut] => { "changed": false, "false_condition": "not __podman_quadlet_file_src", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48 Tuesday 06 February 2024 16:18:32 +0000 (0:00:00.020) 0:01:19.088 ****** ok: [sut] => { "ansible_facts": { "__podman_quadlet_name": "quadlet-demo-mysql", "__podman_quadlet_type": "volume", "__podman_rootless": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57 Tuesday 06 February 2024 16:18:32 +0000 (0:00:00.029) 0:01:19.118 ****** included: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for sut TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Tuesday 06 February 2024 16:18:32 +0000 (0:00:00.037) 0:01:19.156 ****** skipping: [sut] => { "changed": false, "false_condition": "'getent_passwd' not in ansible_facts or __podman_user not in ansible_facts['getent_passwd']", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Tuesday 06 February 2024 16:18:32 +0000 (0:00:00.023) 0:01:19.180 ****** skipping: [sut] => { "changed": false, "false_condition": "not ansible_facts[\"getent_passwd\"][__podman_user]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Tuesday 06 February 2024 16:18:32 +0000 (0:00:00.020) 0:01:19.201 ****** ok: [sut] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get group information] **************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:27 Tuesday 06 February 2024 16:18:32 +0000 (0:00:00.027) 0:01:19.228 ****** ok: [sut] => { "ansible_facts": { "getent_group": { "root": [ "x", "0", "" ] } }, "changed": false } TASK [fedora.linux_system_roles.podman : Set group name] *********************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:34 Tuesday 06 February 2024 16:18:33 +0000 (0:00:00.209) 0:01:19.438 ****** ok: [sut] => { "ansible_facts": { "__podman_group_name": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:38 Tuesday 06 February 2024 16:18:33 +0000 (0:00:00.029) 0:01:19.467 ****** ok: [sut] => { "changed": false, "stat": { "atime": 1707236301.2420683, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "076f89863b474504550ab1b5c11216a77ed7123c", "ctime": 1707236269.7653096, "dev": 51713, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 4471560, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-pie-executable", "mode": "0755", "mtime": 1689166618.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 15496, "uid": 0, "version": "396954581", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************ task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:49 Tuesday 06 February 2024 16:18:33 +0000 (0:00:00.204) 0:01:19.671 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check group with getsubids] *********** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:53 Tuesday 06 February 2024 16:18:33 +0000 (0:00:00.019) 0:01:19.691 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if user is in subuid file] ****** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:63 Tuesday 06 February 2024 16:18:33 +0000 (0:00:00.017) 0:01:19.708 ****** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:71 Tuesday 06 February 2024 16:18:33 +0000 (0:00:00.065) 0:01:19.774 ****** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if group is in subgid file] ***** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:78 Tuesday 06 February 2024 16:18:33 +0000 (0:00:00.018) 0:01:19.792 ****** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] ***** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:86 Tuesday 06 February 2024 16:18:33 +0000 (0:00:00.018) 0:01:19.811 ****** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62 Tuesday 06 February 2024 16:18:33 +0000 (0:00:00.017) 0:01:19.829 ****** ok: [sut] => { "ansible_facts": { "__podman_activate_systemd_unit": true, "__podman_images_found": [], "__podman_kube_yamls_raw": "", "__podman_service_name": "quadlet-demo-mysql-volume.service", "__podman_systemd_scope": "system", "__podman_user_home_dir": "/root", "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73 Tuesday 06 February 2024 16:18:33 +0000 (0:00:00.037) 0:01:19.866 ****** ok: [sut] => { "ansible_facts": { "__podman_quadlet_path": "/etc/containers/systemd" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get kube yaml contents] *************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77 Tuesday 06 February 2024 16:18:33 +0000 (0:00:00.023) 0:01:19.890 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_kube_yamls_raw | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87 Tuesday 06 February 2024 16:18:33 +0000 (0:00:00.021) 0:01:19.911 ****** ok: [sut] => { "ansible_facts": { "__podman_images": [], "__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo-mysql.volume", "__podman_volumes": [] }, "changed": false } TASK [fedora.linux_system_roles.podman : Cleanup quadlets] ********************* task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:103 Tuesday 06 February 2024 16:18:33 +0000 (0:00:00.047) 0:01:19.959 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_state == \"absent\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create and update quadlets] *********** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:107 Tuesday 06 February 2024 16:18:33 +0000 (0:00:00.019) 0:01:19.978 ****** included: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml for sut TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:2 Tuesday 06 February 2024 16:18:33 +0000 (0:00:00.052) 0:01:20.031 ****** included: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for sut TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Tuesday 06 February 2024 16:18:33 +0000 (0:00:00.034) 0:01:20.065 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Tuesday 06 February 2024 16:18:33 +0000 (0:00:00.018) 0:01:20.084 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Tuesday 06 February 2024 16:18:33 +0000 (0:00:00.019) 0:01:20.103 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create host directories] ************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:7 Tuesday 06 February 2024 16:18:33 +0000 (0:00:00.026) 0:01:20.129 ****** skipping: [sut] => { "changed": false, "skipped_reason": "No items in the list" } TASK [fedora.linux_system_roles.podman : Ensure container images are present] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:20 Tuesday 06 February 2024 16:18:33 +0000 (0:00:00.018) 0:01:20.148 ****** skipping: [sut] => { "changed": false, "skipped_reason": "No items in the list" } TASK [fedora.linux_system_roles.podman : Ensure the quadlet directory is present] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:37 Tuesday 06 February 2024 16:18:33 +0000 (0:00:00.022) 0:01:20.171 ****** ok: [sut] => { "changed": false, "gid": 0, "group": "root", "mode": "0755", "owner": "root", "path": "/etc/containers/systemd", "secontext": "system_u:object_r:etc_t:s0", "size": 34, "state": "directory", "uid": 0 } TASK [fedora.linux_system_roles.podman : Ensure quadlet file is copied] ******** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:46 Tuesday 06 February 2024 16:18:34 +0000 (0:00:00.215) 0:01:20.386 ****** changed: [sut] => { "changed": true, "checksum": "585f8cbdf0ec73000f9227dcffbef71e9552ea4a", "dest": "/etc/containers/systemd/quadlet-demo-mysql.volume", "gid": 0, "group": "root", "md5sum": "5ddd03a022aeb4502d9bc8ce436b4233", "mode": "0644", "owner": "root", "secontext": "system_u:object_r:etc_t:s0", "size": 9, "src": "/root/.ansible/tmp/ansible-tmp-1707236314.121802-8128-219481856671228/source", "state": "file", "uid": 0 } TASK [fedora.linux_system_roles.podman : Ensure quadlet file content is present] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:56 Tuesday 06 February 2024 16:18:34 +0000 (0:00:00.526) 0:01:20.913 ****** skipping: [sut] => { "changed": false, "false_condition": "not __podman_quadlet_file_src", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Ensure quadlet file is present] ******* task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:68 Tuesday 06 February 2024 16:18:34 +0000 (0:00:00.020) 0:01:20.933 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_copy_file is skipped", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Reload systemctl] ********************* task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:80 Tuesday 06 February 2024 16:18:34 +0000 (0:00:00.020) 0:01:20.954 ****** ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.podman : Start service] ************************ task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:108 Tuesday 06 February 2024 16:18:35 +0000 (0:00:00.562) 0:01:21.516 ****** changed: [sut] => { "changed": true, "name": "quadlet-demo-mysql-volume.service", "state": "started", "status": { "AccessSELinuxContext": "system_u:object_r:systemd_unit_file_t:s0", "ActiveEnterTimestampMonotonic": "0", "ActiveExitTimestampMonotonic": "0", "ActiveState": "inactive", "After": "basic.target system.slice -.mount sysinit.target systemd-journald.socket", "AllowIsolate": "no", "AssertResult": "no", "AssertTimestampMonotonic": "0", "Before": "shutdown.target", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "CPUAccounting": "yes", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "[not set]", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "no", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore", "CleanResult": "success", "CollectMode": "inactive", "ConditionResult": "no", "ConditionTimestampMonotonic": "0", "ConfigurationDirectoryMode": "0755", "Conflicts": "shutdown.target", "ControlGroupId": "0", "ControlPID": "0", "CoredumpFilter": "0x33", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "Delegate": "no", "Description": "quadlet-demo-mysql-volume.service", "DevicePolicy": "auto", "DynamicUser": "no", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "0", "ExecMainStartTimestampMonotonic": "0", "ExecMainStatus": "0", "ExecStart": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman volume create --ignore systemd-quadlet-demo-mysql ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStartEx": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman volume create --ignore systemd-quadlet-demo-mysql ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExitType": "main", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FinalKillSignal": "9", "FragmentPath": "/run/systemd/generator/quadlet-demo-mysql-volume.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOReadBytes": "18446744073709551615", "IOReadOperations": "18446744073709551615", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "IOWeight": "[not set]", "IOWriteBytes": "18446744073709551615", "IOWriteOperations": "18446744073709551615", "IPAccounting": "no", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "Id": "quadlet-demo-mysql-volume.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestampMonotonic": "0", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "control-group", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "0", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "8388608", "LimitMEMLOCKSoft": "8388608", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "524288", "LimitNOFILESoft": "1024", "LimitNPROC": "13974", "LimitNPROCSoft": "13974", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "13974", "LimitSIGPENDINGSoft": "13974", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "0", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "ManagedOOMSwap": "auto", "MemoryAccounting": "yes", "MemoryAvailable": "infinity", "MemoryCurrent": "[not set]", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemorySwapMax": "infinity", "MountAPIVFS": "no", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAPolicy": "n/a", "Names": "quadlet-demo-mysql-volume.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMPolicy": "stop", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "OnSuccessJobMode": "fail", "Perpetual": "no", "PrivateDevices": "no", "PrivateIPC": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProcSubset": "all", "ProtectClock": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectHostname": "no", "ProtectKernelLogs": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectProc": "default", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "ReloadResult": "success", "ReloadSignal": "1", "RemainAfterExit": "yes", "RemoveIPC": "no", "Requires": "system.slice -.mount sysinit.target", "RequiresMountsFor": "/run/containers", "Restart": "no", "RestartKillSignal": "15", "RestartUSec": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "Slice": "system.slice", "StandardError": "inherit", "StandardInput": "null", "StandardOutput": "journal", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StateChangeTimestampMonotonic": "0", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "dead", "SuccessAction": "none", "SyslogFacility": "3", "SyslogIdentifier": "quadlet-demo-mysql-volume", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "2147483646", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "[not set]", "TasksMax": "22359", "TimeoutAbortUSec": "1min 30s", "TimeoutCleanUSec": "infinity", "TimeoutStartFailureMode": "terminate", "TimeoutStartUSec": "infinity", "TimeoutStopFailureMode": "terminate", "TimeoutStopUSec": "1min 30s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "oneshot", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "disabled", "UnitFileState": "generated", "UtmpMode": "init", "WatchdogSignal": "6", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "infinity" } } TASK [fedora.linux_system_roles.podman : Restart service] ********************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:123 Tuesday 06 February 2024 16:18:35 +0000 (0:00:00.420) 0:01:21.937 ****** skipping: [sut] => { "changed": false, "false_condition": "not __podman_service_started is changed", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14 Tuesday 06 February 2024 16:18:35 +0000 (0:00:00.021) 0:01:21.959 ****** ok: [sut] => { "ansible_facts": { "__podman_quadlet_file_src": "", "__podman_quadlet_spec": {}, "__podman_quadlet_str": "[Install]\nWantedBy=default.target\n\n[Container]\nImage=quay.io/linux-system-roles/mysql:5.6\nContainerName=quadlet-demo-mysql\nVolume=quadlet-demo-mysql.volume:/var/lib/mysql\nVolume=/tmp/quadlet_demo:/var/lib/quadlet_demo:Z\nNetwork=quadlet-demo.network\nSecret=mysql-root-password-container,type=env,target=MYSQL_ROOT_PASSWORD\nHealthCmd=/usr/bin/true\nHealthOnFailure=kill\n", "__podman_quadlet_template_src": "quadlet-demo-mysql.container.j2" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25 Tuesday 06 February 2024 16:18:35 +0000 (0:00:00.064) 0:01:22.023 ****** ok: [sut] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_pull_image": true, "__podman_state": "created", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] ***** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35 Tuesday 06 February 2024 16:18:35 +0000 (0:00:00.026) 0:01:22.050 ****** skipping: [sut] => { "changed": false, "false_condition": "not __podman_quadlet_str", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48 Tuesday 06 February 2024 16:18:35 +0000 (0:00:00.023) 0:01:22.073 ****** ok: [sut] => { "ansible_facts": { "__podman_quadlet_name": "quadlet-demo-mysql", "__podman_quadlet_type": "container", "__podman_rootless": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57 Tuesday 06 February 2024 16:18:35 +0000 (0:00:00.032) 0:01:22.105 ****** included: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for sut TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Tuesday 06 February 2024 16:18:35 +0000 (0:00:00.039) 0:01:22.145 ****** skipping: [sut] => { "changed": false, "false_condition": "'getent_passwd' not in ansible_facts or __podman_user not in ansible_facts['getent_passwd']", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Tuesday 06 February 2024 16:18:35 +0000 (0:00:00.024) 0:01:22.170 ****** skipping: [sut] => { "changed": false, "false_condition": "not ansible_facts[\"getent_passwd\"][__podman_user]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Tuesday 06 February 2024 16:18:35 +0000 (0:00:00.023) 0:01:22.193 ****** ok: [sut] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get group information] **************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:27 Tuesday 06 February 2024 16:18:35 +0000 (0:00:00.029) 0:01:22.222 ****** ok: [sut] => { "ansible_facts": { "getent_group": { "root": [ "x", "0", "" ] } }, "changed": false } TASK [fedora.linux_system_roles.podman : Set group name] *********************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:34 Tuesday 06 February 2024 16:18:36 +0000 (0:00:00.211) 0:01:22.434 ****** ok: [sut] => { "ansible_facts": { "__podman_group_name": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:38 Tuesday 06 February 2024 16:18:36 +0000 (0:00:00.036) 0:01:22.470 ****** ok: [sut] => { "changed": false, "stat": { "atime": 1707236301.2420683, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "076f89863b474504550ab1b5c11216a77ed7123c", "ctime": 1707236269.7653096, "dev": 51713, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 4471560, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-pie-executable", "mode": "0755", "mtime": 1689166618.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 15496, "uid": 0, "version": "396954581", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************ task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:49 Tuesday 06 February 2024 16:18:36 +0000 (0:00:00.202) 0:01:22.673 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check group with getsubids] *********** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:53 Tuesday 06 February 2024 16:18:36 +0000 (0:00:00.018) 0:01:22.692 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if user is in subuid file] ****** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:63 Tuesday 06 February 2024 16:18:36 +0000 (0:00:00.017) 0:01:22.710 ****** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:71 Tuesday 06 February 2024 16:18:36 +0000 (0:00:00.018) 0:01:22.728 ****** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if group is in subgid file] ***** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:78 Tuesday 06 February 2024 16:18:36 +0000 (0:00:00.019) 0:01:22.747 ****** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] ***** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:86 Tuesday 06 February 2024 16:18:36 +0000 (0:00:00.019) 0:01:22.767 ****** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62 Tuesday 06 February 2024 16:18:36 +0000 (0:00:00.018) 0:01:22.785 ****** ok: [sut] => { "ansible_facts": { "__podman_activate_systemd_unit": true, "__podman_images_found": [ "quay.io/linux-system-roles/mysql:5.6" ], "__podman_kube_yamls_raw": "", "__podman_service_name": "quadlet-demo-mysql.service", "__podman_systemd_scope": "system", "__podman_user_home_dir": "/root", "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73 Tuesday 06 February 2024 16:18:36 +0000 (0:00:00.085) 0:01:22.871 ****** ok: [sut] => { "ansible_facts": { "__podman_quadlet_path": "/etc/containers/systemd" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get kube yaml contents] *************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77 Tuesday 06 February 2024 16:18:36 +0000 (0:00:00.022) 0:01:22.893 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_kube_yamls_raw | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87 Tuesday 06 February 2024 16:18:36 +0000 (0:00:00.021) 0:01:22.915 ****** ok: [sut] => { "ansible_facts": { "__podman_images": [ "quay.io/linux-system-roles/mysql:5.6" ], "__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo-mysql.container", "__podman_volumes": [ "/tmp/quadlet_demo" ] }, "changed": false } TASK [fedora.linux_system_roles.podman : Cleanup quadlets] ********************* task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:103 Tuesday 06 February 2024 16:18:36 +0000 (0:00:00.048) 0:01:22.963 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_state == \"absent\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create and update quadlets] *********** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:107 Tuesday 06 February 2024 16:18:36 +0000 (0:00:00.019) 0:01:22.983 ****** included: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml for sut TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:2 Tuesday 06 February 2024 16:18:36 +0000 (0:00:00.040) 0:01:23.024 ****** included: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for sut TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Tuesday 06 February 2024 16:18:36 +0000 (0:00:00.033) 0:01:23.057 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Tuesday 06 February 2024 16:18:36 +0000 (0:00:00.018) 0:01:23.076 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Tuesday 06 February 2024 16:18:36 +0000 (0:00:00.018) 0:01:23.095 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create host directories] ************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:7 Tuesday 06 February 2024 16:18:36 +0000 (0:00:00.018) 0:01:23.113 ****** changed: [sut] => (item=/tmp/quadlet_demo) => { "ansible_loop_var": "item", "changed": true, "gid": 0, "group": "root", "item": "/tmp/quadlet_demo", "mode": "0777", "owner": "root", "path": "/tmp/quadlet_demo", "secontext": "unconfined_u:object_r:user_tmp_t:s0", "size": 6, "state": "directory", "uid": 0 } TASK [fedora.linux_system_roles.podman : Ensure container images are present] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:20 Tuesday 06 February 2024 16:18:37 +0000 (0:00:00.226) 0:01:23.340 ****** changed: [sut] => (item=quay.io/linux-system-roles/mysql:5.6) => { "actions": [ "Pulled image quay.io/linux-system-roles/mysql:5.6" ], "ansible_loop_var": "item", "changed": true, "failed_when_result": false, "image": [ { "Annotations": {}, "Architecture": "amd64", "Author": "", "Comment": "", "Config": { "Cmd": [ "mysqld" ], "Entrypoint": [ "docker-entrypoint.sh" ], "Env": [ "PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin", "GOSU_VERSION=1.12", "MYSQL_MAJOR=5.6", "MYSQL_VERSION=5.6.51-1debian9" ], "ExposedPorts": { "3306/tcp": {} }, "Volumes": { "/var/lib/mysql": {} } }, "Created": "2021-12-21T02:57:59.587337844Z", "Digest": "sha256:2d0419de75076a55e0d0c0680b4f86ecdbb8a88de92054698a0028bcb218a3a8", "GraphDriver": { "Data": { "LowerDir": "/var/lib/containers/storage/overlay/63c8cde32bc583ca7971591c23194c1fd072197949f28fcad7f10780ad1af022/diff:/var/lib/containers/storage/overlay/4904a0ac4db0b6060c944fc7c882cba99b79b50f90151785350c0e4ba7f94390/diff:/var/lib/containers/storage/overlay/1822e40336d4b2b31d888f669f443965d4476dae35fd0a67e58eefd9c5ecb9e5/diff:/var/lib/containers/storage/overlay/988a3979591d296974121b0a893ec621ea5e9939ab3cc5e7c194dbe27f7b5677/diff:/var/lib/containers/storage/overlay/ba0315059f870fcb66ebd6517f455f714fb4eeeb4fdee9cefb3d7f59a7b0cdb7/diff:/var/lib/containers/storage/overlay/0dbdd9024cc38ecc226b2c4772718d32ff0abc288b41669d58d506274c16c67a/diff:/var/lib/containers/storage/overlay/f8ddb99f39386eac409b19f8b347a7a10ee91771cfb838c88793ad8947a233f5/diff:/var/lib/containers/storage/overlay/2ecb8e9889676e4b9b129bab56314644612905ce6e6ce1c92d5a86ac8f08f03b/diff:/var/lib/containers/storage/overlay/07e715cdf99b679f1e067f4819d1b7e1ba9ea24229ede4e6b45b9240fe5865f8/diff:/var/lib/containers/storage/overlay/2b83e5699838047f936a3875bcce7fe1b169983bf86785ae7519c5bc488558ae/diff", "UpperDir": "/var/lib/containers/storage/overlay/bb762056c5cde2e3213fe6ccf595176708249829a607b0afc5799075d4a852ec/diff", "WorkDir": "/var/lib/containers/storage/overlay/bb762056c5cde2e3213fe6ccf595176708249829a607b0afc5799075d4a852ec/work" }, "Name": "overlay" }, "History": [ { "created": "2021-12-21T01:24:41.626621944Z", "created_by": "/bin/sh -c #(nop) ADD file:b1650c19c1dd1f4cff71553f2bb0bc949944d0bc24b54c318b2880c14538648a in / " }, { "created": "2021-12-21T01:24:42.056427889Z", "created_by": "/bin/sh -c #(nop) CMD [\"bash\"]", "empty_layer": true }, { "created": "2021-12-21T02:56:59.289452902Z", "created_by": "/bin/sh -c groupadd -r mysql && useradd -r -g mysql mysql" }, { "created": "2021-12-21T02:57:05.692613282Z", "created_by": "/bin/sh -c apt-get update && apt-get install -y --no-install-recommends gnupg dirmngr && rm -rf /var/lib/apt/lists/*" }, { "created": "2021-12-21T02:57:05.921738618Z", "created_by": "/bin/sh -c #(nop) ENV GOSU_VERSION=1.12", "empty_layer": true }, { "created": "2021-12-21T02:57:18.981216995Z", "created_by": "/bin/sh -c set -eux; \tsavedAptMark=\"$(apt-mark showmanual)\"; \tapt-get update; \tapt-get install -y --no-install-recommends ca-certificates wget; \trm -rf /var/lib/apt/lists/*; \tdpkgArch=\"$(dpkg --print-architecture | awk -F- '{ print $NF }')\"; \twget -O /usr/local/bin/gosu \"https://github.com/tianon/gosu/releases/download/$GOSU_VERSION/gosu-$dpkgArch\"; \twget -O /usr/local/bin/gosu.asc \"https://github.com/tianon/gosu/releases/download/$GOSU_VERSION/gosu-$dpkgArch.asc\"; \texport GNUPGHOME=\"$(mktemp -d)\"; \tgpg --batch --keyserver hkps://keys.openpgp.org --recv-keys B42F6819007F00F88E364FD4036A9C25BF357DD4; \tgpg --batch --verify /usr/local/bin/gosu.asc /usr/local/bin/gosu; \tgpgconf --kill all; \trm -rf \"$GNUPGHOME\" /usr/local/bin/gosu.asc; \tapt-mark auto '.*' > /dev/null; \t[ -z \"$savedAptMark\" ] || apt-mark manual $savedAptMark > /dev/null; \tapt-get purge -y --auto-remove -o APT::AutoRemove::RecommendsImportant=false; \tchmod +x /usr/local/bin/gosu; \tgosu --version; \tgosu nobody true" }, { "created": "2021-12-21T02:57:19.781116502Z", "created_by": "/bin/sh -c mkdir /docker-entrypoint-initdb.d" }, { "created": "2021-12-21T02:57:25.555452451Z", "created_by": "/bin/sh -c apt-get update && apt-get install -y --no-install-recommends \t\tpwgen \t\tperl \t\txz-utils \t&& rm -rf /var/lib/apt/lists/*" }, { "created": "2021-12-21T02:57:37.684730941Z", "created_by": "/bin/sh -c set -ex; \tkey='A4A9406876FCBD3C456770C88C718D3B5072E1F5'; \texport GNUPGHOME=\"$(mktemp -d)\"; \tgpg --batch --keyserver keyserver.ubuntu.com --recv-keys \"$key\"; \tgpg --batch --export \"$key\" > /etc/apt/trusted.gpg.d/mysql.gpg; \tgpgconf --kill all; \trm -rf \"$GNUPGHOME\"; \tapt-key list > /dev/null" }, { "created": "2021-12-21T02:57:37.903115972Z", "created_by": "/bin/sh -c #(nop) ENV MYSQL_MAJOR=5.6", "empty_layer": true }, { "created": "2021-12-21T02:57:38.155528708Z", "created_by": "/bin/sh -c #(nop) ENV MYSQL_VERSION=5.6.51-1debian9", "empty_layer": true }, { "created": "2021-12-21T02:57:39.106901365Z", "created_by": "/bin/sh -c echo 'deb http://repo.mysql.com/apt/debian/ stretch mysql-5.6' > /etc/apt/sources.list.d/mysql.list" }, { "created": "2021-12-21T02:57:56.693791892Z", "created_by": "/bin/sh -c { \t\techo mysql-community-server mysql-community-server/data-dir select ''; \t\techo mysql-community-server mysql-community-server/root-pass password ''; \t\techo mysql-community-server mysql-community-server/re-root-pass password ''; \t\techo mysql-community-server mysql-community-server/remove-test-db select false; \t} | debconf-set-selections \t&& apt-get update \t&& apt-get install -y \t\tmysql-server=\"${MYSQL_VERSION}\" \t&& find /etc/mysql/ -name '*.cnf' -print0 \t\t| xargs -0 grep -lZE '^(bind-address|log)' \t\t| xargs -rt -0 sed -Ei 's/^(bind-address|log)/#&/' \t&& echo '[mysqld]\\nskip-host-cache\\nskip-name-resolve' > /etc/mysql/conf.d/docker.cnf \t&& rm -rf /var/lib/apt/lists/* \t&& rm -rf /var/lib/mysql && mkdir -p /var/lib/mysql /var/run/mysqld \t&& chown -R mysql:mysql /var/lib/mysql /var/run/mysqld \t&& chmod 1777 /var/run/mysqld /var/lib/mysql" }, { "created": "2021-12-21T02:57:57.29721744Z", "created_by": "/bin/sh -c #(nop) VOLUME [/var/lib/mysql]", "empty_layer": true }, { "created": "2021-12-21T02:57:57.591077869Z", "created_by": "/bin/sh -c #(nop) COPY file:345a22fe55d3e6783a17075612415413487e7dba27fbf1000a67c7870364b739 in /usr/local/bin/ " }, { "created": "2021-12-21T02:57:58.761392735Z", "created_by": "/bin/sh -c ln -s usr/local/bin/docker-entrypoint.sh /entrypoint.sh # backwards compat" }, { "created": "2021-12-21T02:57:59.04809683Z", "created_by": "/bin/sh -c #(nop) ENTRYPOINT [\"docker-entrypoint.sh\"]", "empty_layer": true }, { "created": "2021-12-21T02:57:59.31392792Z", "created_by": "/bin/sh -c #(nop) EXPOSE 3306", "empty_layer": true }, { "created": "2021-12-21T02:57:59.587337844Z", "created_by": "/bin/sh -c #(nop) CMD [\"mysqld\"]", "empty_layer": true } ], "Id": "dd3b2a5dcb48ff61113592ed5ddd762581be4387c7bc552375a2159422aa6bf5", "Labels": null, "ManifestType": "application/vnd.docker.distribution.manifest.v2+json", "NamesHistory": [ "quay.io/linux-system-roles/mysql:5.6" ], "Os": "linux", "Parent": "", "RepoDigests": [ "quay.io/linux-system-roles/mysql@sha256:2d0419de75076a55e0d0c0680b4f86ecdbb8a88de92054698a0028bcb218a3a8", "quay.io/linux-system-roles/mysql@sha256:4faa0a30c0055742c72daff1482976814f340d4bfb647f0fa2c3562fd1a7abda" ], "RepoTags": [ "quay.io/linux-system-roles/mysql:5.6" ], "RootFS": { "Layers": [ "sha256:2b83e5699838047f936a3875bcce7fe1b169983bf86785ae7519c5bc488558ae", "sha256:2e10295573911964a8acc37c738aed958d0bc385bc51df6f184b196fb592c975", "sha256:d414fdead0b9c934781e5611c016d7c237062c9470db810ef28b321a492dbb9c", "sha256:4085e588967d10dafb76ad52f66f661b41b9008cc16568ce02853de7cbbf7433", "sha256:7ea96a4e341b4d139f5e076e71d19229537292226c3b24ec2d588567667a1191", "sha256:e3dce1c82d4e4142e8e1bf2fb9c6fd2555196d40088167270ff95c9af37b88af", "sha256:2612088e90f60ecb4a13e68adbfb2b1366f82fc8fbf1f0970f8769cf3cd753ee", "sha256:eba393347f896c4028340bf1a40cc5d6b8f21eeb7abf3e79799d842a6babb93f", "sha256:7c5a5c1986b16b0d8ad4eaa8066627910ec7794a657467e8f18d223f1c0034fd", "sha256:49a1ca1cd2b80f84d6616df2b5123458f4bcb065bca28c43d9c3fd5a0480b907", "sha256:7137327a7221ddb123883146beb68003ee31d652ec46ab8bea3ba12c2df8df8f" ], "Type": "layers" }, "Size": 307607194, "User": "", "Version": "20.10.7", "VirtualSize": 307607194 } ], "item": "quay.io/linux-system-roles/mysql:5.6", "podman_actions": [ "/usr/bin/podman image ls quay.io/linux-system-roles/mysql:5.6 --format json", "/usr/bin/podman image exists quay.io/linux-system-roles/mysql:5.6", "/usr/bin/podman pull quay.io/linux-system-roles/mysql:5.6 -q", "/usr/bin/podman inspect dd3b2a5dcb48ff61113592ed5ddd762581be4387c7bc552375a2159422aa6bf5 --format json" ] } TASK [fedora.linux_system_roles.podman : Ensure the quadlet directory is present] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:37 Tuesday 06 February 2024 16:18:45 +0000 (0:00:08.510) 0:01:31.850 ****** ok: [sut] => { "changed": false, "gid": 0, "group": "root", "mode": "0755", "owner": "root", "path": "/etc/containers/systemd", "secontext": "system_u:object_r:etc_t:s0", "size": 67, "state": "directory", "uid": 0 } TASK [fedora.linux_system_roles.podman : Ensure quadlet file is copied] ******** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:46 Tuesday 06 February 2024 16:18:45 +0000 (0:00:00.215) 0:01:32.065 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_quadlet_file_src | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Ensure quadlet file content is present] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:56 Tuesday 06 February 2024 16:18:45 +0000 (0:00:00.020) 0:01:32.085 ****** changed: [sut] => { "changed": true, "checksum": "7bf9312d57274d4dabbd222b01717e2cc0cfa0ed", "dest": "/etc/containers/systemd/quadlet-demo-mysql.container", "gid": 0, "group": "root", "md5sum": "f1cd3875deb45b806389825bb4b5ad4a", "mode": "0644", "owner": "root", "secontext": "system_u:object_r:etc_t:s0", "size": 367, "src": "/root/.ansible/tmp/ansible-tmp-1707236325.822158-8180-172502824652855/source", "state": "file", "uid": 0 } TASK [fedora.linux_system_roles.podman : Ensure quadlet file is present] ******* task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:68 Tuesday 06 February 2024 16:18:46 +0000 (0:00:00.525) 0:01:32.611 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_copy_content is skipped", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Reload systemctl] ********************* task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:80 Tuesday 06 February 2024 16:18:46 +0000 (0:00:00.021) 0:01:32.633 ****** ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.podman : Start service] ************************ task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:108 Tuesday 06 February 2024 16:18:46 +0000 (0:00:00.555) 0:01:33.188 ****** fatal: [sut]: FAILED! => { "changed": false } MSG: Unable to start service quadlet-demo-mysql.service: Job for quadlet-demo-mysql.service failed because the control process exited with error code. See "systemctl status quadlet-demo-mysql.service" and "journalctl -xeu quadlet-demo-mysql.service" for details. TASK [Dump journal] ************************************************************ task path: /WORKDIR/git-fix-linger-handlingih6nvedg/tests/tests_quadlet_demo.yml:126 Tuesday 06 February 2024 16:18:47 +0000 (0:00:00.513) 0:01:33.701 ****** fatal: [sut]: FAILED! => { "changed": false, "cmd": [ "journalctl", "-ex" ], "delta": "0:00:00.037817", "end": "2024-02-06 16:18:47.611473", "failed_when_result": true, "rc": 0, "start": "2024-02-06 16:18:47.573656" } STDOUT: Feb 06 16:09:30 localhost systemd[1]: systemd-journald.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit systemd-journald.service has successfully entered the 'dead' state. Feb 06 16:09:30 localhost systemd[1]: Starting Flush Journal to Persistent Storage... ░░ Subject: A start job for unit systemd-journal-flush.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-journal-flush.service has begun execution. ░░ ░░ The job identifier is 166. Feb 06 16:09:30 localhost systemd[1]: Finished Load/Save Random Seed. ░░ Subject: A start job for unit systemd-random-seed.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-random-seed.service has finished successfully. ░░ ░░ The job identifier is 137. Feb 06 16:09:30 localhost systemd[1]: First Boot Complete was skipped because of an unmet condition check (ConditionFirstBoot=yes). ░░ Subject: A start job for unit first-boot-complete.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit first-boot-complete.target has finished successfully. ░░ ░░ The job identifier is 138. Feb 06 16:09:30 localhost systemd-journald[483]: Runtime Journal (/run/log/journal/acfd29633386459288749fe4a6dbe762) is 8.0M, max 70.6M, 62.6M free. ░░ Subject: Disk space used by the journal ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ Runtime Journal (/run/log/journal/acfd29633386459288749fe4a6dbe762) is currently using 8.0M. ░░ Maximum allowed usage is set to 70.6M. ░░ Leaving at least 35.3M free (of currently available 690.5M of disk space). ░░ Enforced usage limit is thus 70.6M, of which 62.6M are still available. ░░ ░░ The limits controlling how much disk space is used by the journal may ░░ be configured with SystemMaxUse=, SystemKeepFree=, SystemMaxFileSize=, ░░ RuntimeMaxUse=, RuntimeKeepFree=, RuntimeMaxFileSize= settings in ░░ /etc/systemd/journald.conf. See journald.conf(5) for details. Feb 06 16:09:31 localhost systemd-journald[483]: Received client request to flush runtime journal. Feb 06 16:09:31 localhost systemd[1]: Finished Flush Journal to Persistent Storage. ░░ Subject: A start job for unit systemd-journal-flush.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-journal-flush.service has finished successfully. ░░ ░░ The job identifier is 166. Feb 06 16:09:31 localhost systemd[1]: Finished Coldplug All udev Devices. ░░ Subject: A start job for unit systemd-udev-trigger.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-udev-trigger.service has finished successfully. ░░ ░░ The job identifier is 150. Feb 06 16:09:31 localhost systemd[1]: modprobe@configfs.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit modprobe@configfs.service has successfully entered the 'dead' state. Feb 06 16:09:31 localhost systemd[1]: Finished Load Kernel Module configfs. ░░ Subject: A start job for unit modprobe@configfs.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit modprobe@configfs.service has finished successfully. ░░ ░░ The job identifier is 158. Feb 06 16:09:31 localhost kernel: fuse: init (API version 7.36) Feb 06 16:09:31 localhost systemd[1]: modprobe@fuse.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit modprobe@fuse.service has successfully entered the 'dead' state. Feb 06 16:09:31 localhost systemd[1]: Finished Load Kernel Module fuse. ░░ Subject: A start job for unit modprobe@fuse.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit modprobe@fuse.service has finished successfully. ░░ ░░ The job identifier is 164. Feb 06 16:09:31 localhost kernel: ACPI: bus type drm_connector registered Feb 06 16:09:31 localhost systemd[1]: modprobe@drm.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit modprobe@drm.service has successfully entered the 'dead' state. Feb 06 16:09:31 localhost systemd[1]: Finished Load Kernel Module drm. ░░ Subject: A start job for unit modprobe@drm.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit modprobe@drm.service has finished successfully. ░░ ░░ The job identifier is 243. Feb 06 16:09:31 localhost systemd[1]: Mounting FUSE Control File System... ░░ Subject: A start job for unit sys-fs-fuse-connections.mount has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sys-fs-fuse-connections.mount has begun execution. ░░ ░░ The job identifier is 163. Feb 06 16:09:31 localhost systemd[1]: Mounting Kernel Configuration File System... ░░ Subject: A start job for unit sys-kernel-config.mount has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sys-kernel-config.mount has begun execution. ░░ ░░ The job identifier is 157. Feb 06 16:09:31 localhost systemd[1]: Mounted FUSE Control File System. ░░ Subject: A start job for unit sys-fs-fuse-connections.mount has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sys-fs-fuse-connections.mount has finished successfully. ░░ ░░ The job identifier is 163. Feb 06 16:09:31 localhost systemd[1]: Mounted Kernel Configuration File System. ░░ Subject: A start job for unit sys-kernel-config.mount has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sys-kernel-config.mount has finished successfully. ░░ ░░ The job identifier is 157. Feb 06 16:09:31 localhost systemd[1]: Finished Create Static Device Nodes in /dev. ░░ Subject: A start job for unit systemd-tmpfiles-setup-dev.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-tmpfiles-setup-dev.service has finished successfully. ░░ ░░ The job identifier is 141. Feb 06 16:09:31 localhost systemd[1]: Reached target Preparation for Local File Systems. ░░ Subject: A start job for unit local-fs-pre.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit local-fs-pre.target has finished successfully. ░░ ░░ The job identifier is 146. Feb 06 16:09:31 localhost systemd[1]: Reached target Local File Systems. ░░ Subject: A start job for unit local-fs.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit local-fs.target has finished successfully. ░░ ░░ The job identifier is 144. Feb 06 16:09:31 localhost systemd[1]: Rebuild Dynamic Linker Cache was skipped because no trigger condition checks were met. ░░ Subject: A start job for unit ldconfig.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit ldconfig.service has finished successfully. ░░ ░░ The job identifier is 170. Feb 06 16:09:31 localhost systemd[1]: Mark the need to relabel after reboot was skipped because of an unmet condition check (ConditionSecurity=!selinux). ░░ Subject: A start job for unit selinux-autorelabel-mark.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit selinux-autorelabel-mark.service has finished successfully. ░░ ░░ The job identifier is 143. Feb 06 16:09:31 localhost systemd[1]: Set Up Additional Binary Formats was skipped because no trigger condition checks were met. ░░ Subject: A start job for unit systemd-binfmt.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-binfmt.service has finished successfully. ░░ ░░ The job identifier is 172. Feb 06 16:09:31 localhost systemd[1]: Store a System Token in an EFI Variable was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/LoaderFeatures-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). ░░ Subject: A start job for unit systemd-boot-system-token.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-boot-system-token.service has finished successfully. ░░ ░░ The job identifier is 176. Feb 06 16:09:31 localhost systemd[1]: Starting Automatic Boot Loader Update... ░░ Subject: A start job for unit systemd-boot-update.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-boot-update.service has begun execution. ░░ ░░ The job identifier is 147. Feb 06 16:09:31 localhost systemd[1]: Starting Commit a transient machine-id on disk... ░░ Subject: A start job for unit systemd-machine-id-commit.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-machine-id-commit.service has begun execution. ░░ ░░ The job identifier is 124. Feb 06 16:09:31 localhost systemd[1]: Starting Create Volatile Files and Directories... ░░ Subject: A start job for unit systemd-tmpfiles-setup.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-tmpfiles-setup.service has begun execution. ░░ ░░ The job identifier is 161. Feb 06 16:09:31 localhost bootctl[496]: Couldn't find EFI system partition, skipping. Feb 06 16:09:31 localhost systemd[1]: Starting Rule-based Manager for Device Events and Files... ░░ Subject: A start job for unit systemd-udevd.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-udevd.service has begun execution. ░░ ░░ The job identifier is 151. Feb 06 16:09:31 localhost systemd[1]: etc-machine\x2did.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit etc-machine\x2did.mount has successfully entered the 'dead' state. Feb 06 16:09:31 localhost systemd[1]: Finished Automatic Boot Loader Update. ░░ Subject: A start job for unit systemd-boot-update.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-boot-update.service has finished successfully. ░░ ░░ The job identifier is 147. Feb 06 16:09:32 localhost systemd[1]: Finished Commit a transient machine-id on disk. ░░ Subject: A start job for unit systemd-machine-id-commit.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-machine-id-commit.service has finished successfully. ░░ ░░ The job identifier is 124. Feb 06 16:09:32 localhost systemd-udevd[500]: Using default interface naming scheme 'rhel-9.0'. Feb 06 16:09:32 localhost systemd[1]: Finished Create Volatile Files and Directories. ░░ Subject: A start job for unit systemd-tmpfiles-setup.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-tmpfiles-setup.service has finished successfully. ░░ ░░ The job identifier is 161. Feb 06 16:09:32 localhost systemd[1]: Mounting RPC Pipe File System... ░░ Subject: A start job for unit var-lib-nfs-rpc_pipefs.mount has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit var-lib-nfs-rpc_pipefs.mount has begun execution. ░░ ░░ The job identifier is 207. Feb 06 16:09:32 localhost systemd[1]: Starting Security Auditing Service... ░░ Subject: A start job for unit auditd.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit auditd.service has begun execution. ░░ ░░ The job identifier is 240. Feb 06 16:09:32 localhost systemd[1]: Starting RPC Bind... ░░ Subject: A start job for unit rpcbind.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit rpcbind.service has begun execution. ░░ ░░ The job identifier is 226. Feb 06 16:09:32 localhost systemd[1]: Rebuild Journal Catalog was skipped because of an unmet condition check (ConditionNeedsUpdate=/var). ░░ Subject: A start job for unit systemd-journal-catalog-update.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-journal-catalog-update.service has finished successfully. ░░ ░░ The job identifier is 123. Feb 06 16:09:32 localhost systemd[1]: Update is Completed was skipped because no trigger condition checks were met. ░░ Subject: A start job for unit systemd-update-done.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-update-done.service has finished successfully. ░░ ░░ The job identifier is 177. Feb 06 16:09:32 localhost kernel: RPC: Registered named UNIX socket transport module. Feb 06 16:09:32 localhost kernel: RPC: Registered udp transport module. Feb 06 16:09:32 localhost kernel: RPC: Registered tcp transport module. Feb 06 16:09:32 localhost kernel: RPC: Registered tcp-with-tls transport module. Feb 06 16:09:32 localhost kernel: RPC: Registered tcp NFSv4.1 backchannel transport module. Feb 06 16:09:32 localhost systemd[1]: Mounted RPC Pipe File System. ░░ Subject: A start job for unit var-lib-nfs-rpc_pipefs.mount has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit var-lib-nfs-rpc_pipefs.mount has finished successfully. ░░ ░░ The job identifier is 207. Feb 06 16:09:32 localhost systemd[1]: Reached target rpc_pipefs.target. ░░ Subject: A start job for unit rpc_pipefs.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit rpc_pipefs.target has finished successfully. ░░ ░░ The job identifier is 206. Feb 06 16:09:32 localhost systemd[1]: Started Rule-based Manager for Device Events and Files. ░░ Subject: A start job for unit systemd-udevd.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-udevd.service has finished successfully. ░░ ░░ The job identifier is 151. Feb 06 16:09:32 localhost systemd[1]: Starting Load Kernel Module configfs... ░░ Subject: A start job for unit modprobe@configfs.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit modprobe@configfs.service has begun execution. ░░ ░░ The job identifier is 258. Feb 06 16:09:32 localhost systemd[1]: modprobe@configfs.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit modprobe@configfs.service has successfully entered the 'dead' state. Feb 06 16:09:32 localhost systemd[1]: Finished Load Kernel Module configfs. ░░ Subject: A start job for unit modprobe@configfs.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit modprobe@configfs.service has finished successfully. ░░ ░░ The job identifier is 258. Feb 06 16:09:32 localhost kernel: input: PC Speaker as /devices/platform/pcspkr/input/input5 Feb 06 16:09:32 localhost kernel: piix4_smbus 0000:00:01.3: SMBus base address uninitialized - upgrade BIOS or use force_addr=0xaddr Feb 06 16:09:33 localhost kernel: RAPL PMU: API unit is 2^-32 Joules, 0 fixed counters, 655360 ms ovfl timer Feb 06 16:09:33 localhost systemd-udevd[512]: Network interface NamePolicy= disabled on kernel command line. Feb 06 16:09:33 localhost kernel: cirrus 0000:00:02.0: vgaarb: deactivate vga console Feb 06 16:09:33 localhost kernel: Console: switching to colour dummy device 80x25 Feb 06 16:09:33 localhost kernel: [drm] Initialized cirrus 2.0.0 2019 for 0000:00:02.0 on minor 0 Feb 06 16:09:33 localhost kernel: fbcon: cirrusdrmfb (fb0) is primary device Feb 06 16:09:33 localhost kernel: Console: switching to colour frame buffer device 128x48 Feb 06 16:09:33 localhost kernel: cirrus 0000:00:02.0: [drm] fb0: cirrusdrmfb frame buffer device Feb 06 16:09:33 localhost systemd[1]: Started RPC Bind. ░░ Subject: A start job for unit rpcbind.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit rpcbind.service has finished successfully. ░░ ░░ The job identifier is 226. Feb 06 16:09:33 localhost auditd[548]: No plugins found, not dispatching events Feb 06 16:09:33 localhost auditd[548]: Init complete, auditd 3.1.2 listening for events (startup state enable) Feb 06 16:09:34 localhost augenrules[551]: /sbin/augenrules: No change Feb 06 16:09:34 localhost augenrules[568]: No rules Feb 06 16:09:34 localhost augenrules[568]: enabled 1 Feb 06 16:09:34 localhost augenrules[568]: failure 1 Feb 06 16:09:34 localhost augenrules[568]: pid 548 Feb 06 16:09:34 localhost augenrules[568]: rate_limit 0 Feb 06 16:09:34 localhost augenrules[568]: backlog_limit 8192 Feb 06 16:09:34 localhost augenrules[568]: lost 0 Feb 06 16:09:34 localhost augenrules[568]: backlog 0 Feb 06 16:09:34 localhost augenrules[568]: backlog_wait_time 60000 Feb 06 16:09:34 localhost augenrules[568]: backlog_wait_time_actual 0 Feb 06 16:09:34 localhost augenrules[568]: enabled 1 Feb 06 16:09:34 localhost augenrules[568]: failure 1 Feb 06 16:09:34 localhost augenrules[568]: pid 548 Feb 06 16:09:34 localhost augenrules[568]: rate_limit 0 Feb 06 16:09:34 localhost augenrules[568]: backlog_limit 8192 Feb 06 16:09:34 localhost augenrules[568]: lost 0 Feb 06 16:09:34 localhost augenrules[568]: backlog 4 Feb 06 16:09:34 localhost augenrules[568]: backlog_wait_time 60000 Feb 06 16:09:34 localhost augenrules[568]: backlog_wait_time_actual 0 Feb 06 16:09:34 localhost augenrules[568]: enabled 1 Feb 06 16:09:34 localhost augenrules[568]: failure 1 Feb 06 16:09:34 localhost augenrules[568]: pid 548 Feb 06 16:09:34 localhost augenrules[568]: rate_limit 0 Feb 06 16:09:34 localhost augenrules[568]: backlog_limit 8192 Feb 06 16:09:34 localhost augenrules[568]: lost 0 Feb 06 16:09:34 localhost augenrules[568]: backlog 4 Feb 06 16:09:34 localhost augenrules[568]: backlog_wait_time 60000 Feb 06 16:09:34 localhost augenrules[568]: backlog_wait_time_actual 0 Feb 06 16:09:34 localhost systemd[1]: Started Security Auditing Service. ░░ Subject: A start job for unit auditd.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit auditd.service has finished successfully. ░░ ░░ The job identifier is 240. Feb 06 16:09:34 localhost systemd[1]: Starting Record System Boot/Shutdown in UTMP... ░░ Subject: A start job for unit systemd-update-utmp.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-update-utmp.service has begun execution. ░░ ░░ The job identifier is 213. Feb 06 16:09:34 localhost systemd[1]: Finished Record System Boot/Shutdown in UTMP. ░░ Subject: A start job for unit systemd-update-utmp.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-update-utmp.service has finished successfully. ░░ ░░ The job identifier is 213. Feb 06 16:09:34 localhost systemd[1]: Reached target System Initialization. ░░ Subject: A start job for unit sysinit.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sysinit.target has finished successfully. ░░ ░░ The job identifier is 118. Feb 06 16:09:34 localhost systemd[1]: Started dnf makecache --timer. ░░ Subject: A start job for unit dnf-makecache.timer has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit dnf-makecache.timer has finished successfully. ░░ ░░ The job identifier is 182. Feb 06 16:09:34 localhost systemd[1]: Started Daily rotation of log files. ░░ Subject: A start job for unit logrotate.timer has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit logrotate.timer has finished successfully. ░░ ░░ The job identifier is 189. Feb 06 16:09:34 localhost systemd[1]: Started Daily Cleanup of Temporary Directories. ░░ Subject: A start job for unit systemd-tmpfiles-clean.timer has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-tmpfiles-clean.timer has finished successfully. ░░ ░░ The job identifier is 181. Feb 06 16:09:34 localhost systemd[1]: Reached target Timer Units. ░░ Subject: A start job for unit timers.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit timers.target has finished successfully. ░░ ░░ The job identifier is 180. Feb 06 16:09:34 localhost systemd[1]: Listening on D-Bus System Message Bus Socket. ░░ Subject: A start job for unit dbus.socket has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit dbus.socket has finished successfully. ░░ ░░ The job identifier is 186. Feb 06 16:09:34 localhost systemd[1]: Listening on SSSD Kerberos Cache Manager responder socket. ░░ Subject: A start job for unit sssd-kcm.socket has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sssd-kcm.socket has finished successfully. ░░ ░░ The job identifier is 193. Feb 06 16:09:34 localhost systemd[1]: Reached target Socket Units. ░░ Subject: A start job for unit sockets.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sockets.target has finished successfully. ░░ ░░ The job identifier is 191. Feb 06 16:09:34 localhost systemd[1]: Starting D-Bus System Message Bus... ░░ Subject: A start job for unit dbus-broker.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit dbus-broker.service has begun execution. ░░ ░░ The job identifier is 187. Feb 06 16:09:34 localhost systemd[1]: TPM2 PCR Barrier (Initialization) was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/StubPcrKernelImage-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). ░░ Subject: A start job for unit systemd-pcrphase-sysinit.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-pcrphase-sysinit.service has finished successfully. ░░ ░░ The job identifier is 119. Feb 06 16:09:35 localhost systemd[1]: Started D-Bus System Message Bus. ░░ Subject: A start job for unit dbus-broker.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit dbus-broker.service has finished successfully. ░░ ░░ The job identifier is 187. Feb 06 16:09:35 localhost systemd[1]: Reached target Basic System. ░░ Subject: A start job for unit basic.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit basic.target has finished successfully. ░░ ░░ The job identifier is 117. Feb 06 16:09:35 localhost systemd[1]: Starting NTP client/server... ░░ Subject: A start job for unit chronyd.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit chronyd.service has begun execution. ░░ ░░ The job identifier is 235. Feb 06 16:09:35 localhost systemd[1]: Starting Initial cloud-init job (pre-networking)... ░░ Subject: A start job for unit cloud-init-local.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit cloud-init-local.service has begun execution. ░░ ░░ The job identifier is 216. Feb 06 16:09:35 localhost systemd[1]: Starting Restore /run/initramfs on shutdown... ░░ Subject: A start job for unit dracut-shutdown.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit dracut-shutdown.service has begun execution. ░░ ░░ The job identifier is 162. Feb 06 16:09:35 localhost systemd[1]: Started irqbalance daemon. ░░ Subject: A start job for unit irqbalance.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit irqbalance.service has finished successfully. ░░ ░░ The job identifier is 228. Feb 06 16:09:35 localhost systemd[1]: Load CPU microcode update was skipped because of an unmet condition check (ConditionPathExists=/sys/devices/system/cpu/microcode/reload). ░░ Subject: A start job for unit microcode.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit microcode.service has finished successfully. ░░ ░░ The job identifier is 197. Feb 06 16:09:35 localhost systemd[1]: Started Hardware RNG Entropy Gatherer Daemon. ░░ Subject: A start job for unit rngd.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit rngd.service has finished successfully. ░░ ░░ The job identifier is 227. Feb 06 16:09:35 localhost systemd[1]: OpenSSH ecdsa Server Key Generation was skipped because of an unmet condition check (ConditionPathExists=!/run/systemd/generator.early/multi-user.target.wants/cloud-init.target). ░░ Subject: A start job for unit sshd-keygen@ecdsa.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sshd-keygen@ecdsa.service has finished successfully. ░░ ░░ The job identifier is 222. Feb 06 16:09:35 localhost systemd[1]: OpenSSH ed25519 Server Key Generation was skipped because of an unmet condition check (ConditionPathExists=!/run/systemd/generator.early/multi-user.target.wants/cloud-init.target). ░░ Subject: A start job for unit sshd-keygen@ed25519.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sshd-keygen@ed25519.service has finished successfully. ░░ ░░ The job identifier is 221. Feb 06 16:09:35 localhost systemd[1]: OpenSSH rsa Server Key Generation was skipped because of an unmet condition check (ConditionPathExists=!/run/systemd/generator.early/multi-user.target.wants/cloud-init.target). ░░ Subject: A start job for unit sshd-keygen@rsa.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sshd-keygen@rsa.service has finished successfully. ░░ ░░ The job identifier is 219. Feb 06 16:09:35 localhost systemd[1]: Reached target sshd-keygen.target. ░░ Subject: A start job for unit sshd-keygen.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sshd-keygen.target has finished successfully. ░░ ░░ The job identifier is 218. Feb 06 16:09:35 localhost systemd[1]: System Security Services Daemon was skipped because no trigger condition checks were met. ░░ Subject: A start job for unit sssd.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sssd.service has finished successfully. ░░ ░░ The job identifier is 199. Feb 06 16:09:35 localhost systemd[1]: Reached target User and Group Name Lookups. ░░ Subject: A start job for unit nss-user-lookup.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit nss-user-lookup.target has finished successfully. ░░ ░░ The job identifier is 200. Feb 06 16:09:35 localhost systemd[1]: Starting User Login Management... ░░ Subject: A start job for unit systemd-logind.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-logind.service has begun execution. ░░ ░░ The job identifier is 242. Feb 06 16:09:35 localhost dbus-broker-lau[576]: Ready Feb 06 16:09:35 localhost systemd[1]: Starting Rotate log files... ░░ Subject: A start job for unit logrotate.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit logrotate.service has begun execution. ░░ ░░ The job identifier is 263. Feb 06 16:09:35 localhost systemd[1]: Finished Restore /run/initramfs on shutdown. ░░ Subject: A start job for unit dracut-shutdown.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit dracut-shutdown.service has finished successfully. ░░ ░░ The job identifier is 162. Feb 06 16:09:35 localhost systemd-logind[583]: New seat seat0. ░░ Subject: A new seat seat0 is now available ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ Documentation: sd-login(3) ░░ ░░ A new seat seat0 has been configured and is now available. Feb 06 16:09:35 localhost systemd-logind[583]: Watching system buttons on /dev/input/event0 (Power Button) Feb 06 16:09:35 localhost systemd-logind[583]: Watching system buttons on /dev/input/event1 (Sleep Button) Feb 06 16:09:35 localhost systemd-logind[583]: Watching system buttons on /dev/input/event2 (AT Translated Set 2 keyboard) Feb 06 16:09:35 localhost systemd[1]: Started User Login Management. ░░ Subject: A start job for unit systemd-logind.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-logind.service has finished successfully. ░░ ░░ The job identifier is 242. Feb 06 16:09:35 localhost systemd[1]: logrotate.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit logrotate.service has successfully entered the 'dead' state. Feb 06 16:09:35 localhost systemd[1]: Finished Rotate log files. ░░ Subject: A start job for unit logrotate.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit logrotate.service has finished successfully. ░░ ░░ The job identifier is 263. Feb 06 16:09:36 localhost rngd[582]: Disabling 7: PKCS11 Entropy generator (pkcs11) Feb 06 16:09:36 localhost rngd[582]: Disabling 5: NIST Network Entropy Beacon (nist) Feb 06 16:09:36 localhost rngd[582]: Disabling 9: Qrypt quantum entropy beacon (qrypt) Feb 06 16:09:36 localhost rngd[582]: Initializing available sources Feb 06 16:09:36 localhost rngd[582]: [hwrng ]: Initialization Failed Feb 06 16:09:36 localhost rngd[582]: [rdrand]: Enabling RDRAND rng support Feb 06 16:09:36 localhost rngd[582]: [rdrand]: Initialized Feb 06 16:09:36 localhost rngd[582]: [jitter]: JITTER timeout set to 5 sec Feb 06 16:09:36 localhost rngd[582]: [jitter]: Initializing AES buffer Feb 06 16:09:36 localhost chronyd[603]: chronyd version 4.3 starting (+CMDMON +NTP +REFCLOCK +RTC +PRIVDROP +SCFILTER +SIGND +ASYNCDNS +NTS +SECHASH +IPV6 +DEBUG) Feb 06 16:09:36 localhost chronyd[603]: Frequency 0.000 +/- 1000000.000 ppm read from /var/lib/chrony/drift Feb 06 16:09:36 localhost chronyd[603]: Using right/UTC timezone to obtain leap second data Feb 06 16:09:36 localhost chronyd[603]: Loaded seccomp filter (level 2) Feb 06 16:09:36 localhost systemd[1]: Started NTP client/server. ░░ Subject: A start job for unit chronyd.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit chronyd.service has finished successfully. ░░ ░░ The job identifier is 235. Feb 06 16:09:41 localhost rngd[582]: [jitter]: Unable to obtain AES key, disabling JITTER source Feb 06 16:09:41 localhost rngd[582]: [jitter]: Initialization Failed Feb 06 16:09:41 localhost rngd[582]: Process privileges have been dropped to 2:2 Feb 06 16:09:43 localhost cloud-init[607]: Cloud-init v. 23.4-2.el9 running 'init-local' at Tue, 06 Feb 2024 16:09:43 +0000. Up 29.90 seconds. Feb 06 16:09:43 localhost dhclient[610]: Internet Systems Consortium DHCP Client 4.4.2b1 Feb 06 16:09:43 localhost dhclient[610]: Copyright 2004-2019 Internet Systems Consortium. Feb 06 16:09:43 localhost dhclient[610]: All rights reserved. Feb 06 16:09:43 localhost dhclient[610]: For info, please visit https://www.isc.org/software/dhcp/ Feb 06 16:09:43 localhost dhclient[610]: Feb 06 16:09:43 localhost dhclient[610]: Listening on LPF/eth0/12:07:23:9c:5c:03 Feb 06 16:09:43 localhost dhclient[610]: Sending on LPF/eth0/12:07:23:9c:5c:03 Feb 06 16:09:43 localhost dhclient[610]: Sending on Socket/fallback Feb 06 16:09:43 localhost dhclient[610]: DHCPDISCOVER on eth0 to 255.255.255.255 port 67 interval 7 (xid=0xa997160f) Feb 06 16:09:43 localhost dhclient[610]: DHCPOFFER of 10.31.11.33 from 10.31.8.1 Feb 06 16:09:43 localhost dhclient[610]: DHCPREQUEST for 10.31.11.33 on eth0 to 255.255.255.255 port 67 (xid=0xa997160f) Feb 06 16:09:43 localhost dhclient[610]: DHCPACK of 10.31.11.33 from 10.31.8.1 (xid=0xa997160f) Feb 06 16:09:43 localhost dhclient[610]: bound to 10.31.11.33 -- renewal in 1470 seconds. Feb 06 16:09:44 localhost systemd[1]: Starting Hostname Service... ░░ Subject: A start job for unit systemd-hostnamed.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-hostnamed.service has begun execution. ░░ ░░ The job identifier is 326. Feb 06 16:09:44 localhost systemd[1]: Started Hostname Service. ░░ Subject: A start job for unit systemd-hostnamed.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-hostnamed.service has finished successfully. ░░ ░░ The job identifier is 326. Feb 06 16:09:44 ip-10-31-11-33.us-east-1.aws.redhat.com systemd-hostnamed[625]: Hostname set to (static) Feb 06 16:09:44 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: Finished Initial cloud-init job (pre-networking). ░░ Subject: A start job for unit cloud-init-local.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit cloud-init-local.service has finished successfully. ░░ ░░ The job identifier is 216. Feb 06 16:09:44 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: Reached target Preparation for Network. ░░ Subject: A start job for unit network-pre.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit network-pre.target has finished successfully. ░░ ░░ The job identifier is 130. Feb 06 16:09:44 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: Starting Network Manager... ░░ Subject: A start job for unit NetworkManager.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit NetworkManager.service has begun execution. ░░ ░░ The job identifier is 185. Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com /usr/sbin/irqbalance[581]: Cannot change IRQ 0 affinity: Input/output error Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com /usr/sbin/irqbalance[581]: IRQ 0 affinity is now unmanaged Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com /usr/sbin/irqbalance[581]: Cannot change IRQ 48 affinity: Input/output error Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com /usr/sbin/irqbalance[581]: IRQ 48 affinity is now unmanaged Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com /usr/sbin/irqbalance[581]: Cannot change IRQ 49 affinity: Input/output error Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com /usr/sbin/irqbalance[581]: IRQ 49 affinity is now unmanaged Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com /usr/sbin/irqbalance[581]: Cannot change IRQ 50 affinity: Input/output error Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com /usr/sbin/irqbalance[581]: IRQ 50 affinity is now unmanaged Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com /usr/sbin/irqbalance[581]: Cannot change IRQ 51 affinity: Input/output error Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com /usr/sbin/irqbalance[581]: IRQ 51 affinity is now unmanaged Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com /usr/sbin/irqbalance[581]: Cannot change IRQ 52 affinity: Input/output error Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com /usr/sbin/irqbalance[581]: IRQ 52 affinity is now unmanaged Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com /usr/sbin/irqbalance[581]: Cannot change IRQ 53 affinity: Input/output error Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com /usr/sbin/irqbalance[581]: IRQ 53 affinity is now unmanaged Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com /usr/sbin/irqbalance[581]: Cannot change IRQ 54 affinity: Input/output error Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com /usr/sbin/irqbalance[581]: IRQ 54 affinity is now unmanaged Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com /usr/sbin/irqbalance[581]: Cannot change IRQ 55 affinity: Input/output error Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com /usr/sbin/irqbalance[581]: IRQ 55 affinity is now unmanaged Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com /usr/sbin/irqbalance[581]: Cannot change IRQ 56 affinity: Input/output error Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com /usr/sbin/irqbalance[581]: IRQ 56 affinity is now unmanaged Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com /usr/sbin/irqbalance[581]: Cannot change IRQ 57 affinity: Input/output error Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com /usr/sbin/irqbalance[581]: IRQ 57 affinity is now unmanaged Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com /usr/sbin/irqbalance[581]: Cannot change IRQ 58 affinity: Input/output error Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com /usr/sbin/irqbalance[581]: IRQ 58 affinity is now unmanaged Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com /usr/sbin/irqbalance[581]: Cannot change IRQ 59 affinity: Input/output error Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com /usr/sbin/irqbalance[581]: IRQ 59 affinity is now unmanaged Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com NetworkManager[629]: [1707235785.1581] NetworkManager (version 1.45.10-1.el9) is starting... (boot:f673fda0-c26b-48bc-be20-461682a3a931) Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com NetworkManager[629]: [1707235785.1585] Read config: /etc/NetworkManager/NetworkManager.conf (run: 15-carrier-timeout.conf) Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com NetworkManager[629]: [1707235785.2637] manager[0x55624ce8f040]: monitoring kernel firmware directory '/lib/firmware'. Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com NetworkManager[629]: [1707235785.2665] hostname: hostname: using hostnamed Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com NetworkManager[629]: [1707235785.2665] hostname: static hostname changed from (none) to "ip-10-31-11-33.us-east-1.aws.redhat.com" Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com NetworkManager[629]: [1707235785.2680] dns-mgr: init: dns=default,systemd-resolved rc-manager=symlink (auto) Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com NetworkManager[629]: [1707235785.3356] manager[0x55624ce8f040]: rfkill: Wi-Fi hardware radio set enabled Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com NetworkManager[629]: [1707235785.3357] manager[0x55624ce8f040]: rfkill: WWAN hardware radio set enabled Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com NetworkManager[629]: [1707235785.3507] Loaded device plugin: NMTeamFactory (/usr/lib64/NetworkManager/1.45.10-1.el9/libnm-device-plugin-team.so) Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com NetworkManager[629]: [1707235785.3508] manager: rfkill: Wi-Fi enabled by radio killswitch; enabled by state file Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com NetworkManager[629]: [1707235785.3516] manager: rfkill: WWAN enabled by radio killswitch; enabled by state file Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com NetworkManager[629]: [1707235785.3517] manager: Networking is enabled by state file Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com NetworkManager[629]: [1707235785.3531] settings: Loaded settings plugin: keyfile (internal) Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: Listening on Load/Save RF Kill Switch Status /dev/rfkill Watch. ░░ Subject: A start job for unit systemd-rfkill.socket has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-rfkill.socket has finished successfully. ░░ ░░ The job identifier is 390. Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com NetworkManager[629]: [1707235785.3570] settings: Loaded settings plugin: ifcfg-rh ("/usr/lib64/NetworkManager/1.45.10-1.el9/libnm-settings-plugin-ifcfg-rh.so") Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com NetworkManager[629]: [1707235785.3604] Warning: the ifcfg-rh plugin is deprecated, please migrate connections to the keyfile format using "nmcli connection migrate" Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com NetworkManager[629]: [1707235785.3618] dhcp: init: Using DHCP client 'internal' Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com NetworkManager[629]: [1707235785.3621] manager: (lo): new Loopback device (/org/freedesktop/NetworkManager/Devices/1) Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com NetworkManager[629]: [1707235785.3635] device (lo): state change: unmanaged -> unavailable (reason 'connection-assumed', sys-iface-state: 'external') Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com NetworkManager[629]: [1707235785.3642] device (lo): state change: unavailable -> disconnected (reason 'connection-assumed', sys-iface-state: 'external') Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com NetworkManager[629]: [1707235785.3650] device (lo): Activation: starting connection 'lo' (6c6cb1a8-6b76-463d-a17b-0f9144112b39) Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: Starting Network Manager Script Dispatcher Service... ░░ Subject: A start job for unit NetworkManager-dispatcher.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit NetworkManager-dispatcher.service has begun execution. ░░ ░░ The job identifier is 397. Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com NetworkManager[629]: [1707235785.3664] manager: (eth0): new Ethernet device (/org/freedesktop/NetworkManager/Devices/2) Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com NetworkManager[629]: [1707235785.3669] device (eth0): state change: unmanaged -> unavailable (reason 'managed', sys-iface-state: 'external') Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: Started Network Manager. ░░ Subject: A start job for unit NetworkManager.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit NetworkManager.service has finished successfully. ░░ ░░ The job identifier is 185. Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com NetworkManager[629]: [1707235785.3691] bus-manager: acquired D-Bus service "org.freedesktop.NetworkManager" Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: Reached target Network. ░░ Subject: A start job for unit network.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit network.target has finished successfully. ░░ ░░ The job identifier is 188. Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com NetworkManager[629]: [1707235785.3701] device (lo): state change: disconnected -> prepare (reason 'none', sys-iface-state: 'external') Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com NetworkManager[629]: [1707235785.3714] device (lo): state change: prepare -> config (reason 'none', sys-iface-state: 'external') Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com NetworkManager[629]: [1707235785.3720] device (lo): state change: config -> ip-config (reason 'none', sys-iface-state: 'external') Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: Starting Network Manager Wait Online... ░░ Subject: A start job for unit NetworkManager-wait-online.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit NetworkManager-wait-online.service has begun execution. ░░ ░░ The job identifier is 184. Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com NetworkManager[629]: [1707235785.3730] device (eth0): carrier: link connected Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com NetworkManager[629]: [1707235785.3742] device (lo): state change: ip-config -> ip-check (reason 'none', sys-iface-state: 'external') Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: Starting GSSAPI Proxy Daemon... ░░ Subject: A start job for unit gssproxy.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit gssproxy.service has begun execution. ░░ ░░ The job identifier is 209. Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com NetworkManager[629]: [1707235785.3768] device (eth0): state change: unavailable -> disconnected (reason 'carrier-changed', sys-iface-state: 'managed') Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com NetworkManager[629]: [1707235785.3779] policy: auto-activating connection 'System eth0' (5fb06bd0-0bb0-7ffb-45f1-d6edd65f3e03) Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com NetworkManager[629]: [1707235785.3785] device (eth0): Activation: starting connection 'System eth0' (5fb06bd0-0bb0-7ffb-45f1-d6edd65f3e03) Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com NetworkManager[629]: [1707235785.3787] device (eth0): state change: disconnected -> prepare (reason 'none', sys-iface-state: 'managed') Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com NetworkManager[629]: [1707235785.3794] manager: NetworkManager state is now CONNECTING Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com NetworkManager[629]: [1707235785.3798] device (eth0): state change: prepare -> config (reason 'none', sys-iface-state: 'managed') Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com NetworkManager[629]: [1707235785.3806] device (eth0): state change: config -> ip-config (reason 'none', sys-iface-state: 'managed') Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com NetworkManager[629]: [1707235785.3811] dhcp4 (eth0): activation: beginning transaction (timeout in 45 seconds) Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com NetworkManager[629]: [1707235785.3832] dhcp4 (eth0): state changed new lease, address=10.31.11.33 Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com NetworkManager[629]: [1707235785.3837] policy: set 'System eth0' (eth0) as default for IPv4 routing and DNS Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com NetworkManager[629]: [1707235785.3909] device (eth0): state change: ip-config -> ip-check (reason 'none', sys-iface-state: 'managed') Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: Started GSSAPI Proxy Daemon. ░░ Subject: A start job for unit gssproxy.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit gssproxy.service has finished successfully. ░░ ░░ The job identifier is 209. Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: RPC security service for NFS client and server was skipped because of an unmet condition check (ConditionPathExists=/etc/krb5.keytab). ░░ Subject: A start job for unit rpc-gssd.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit rpc-gssd.service has finished successfully. ░░ ░░ The job identifier is 205. Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: Reached target NFS client services. ░░ Subject: A start job for unit nfs-client.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit nfs-client.target has finished successfully. ░░ ░░ The job identifier is 203. Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: Reached target Preparation for Remote File Systems. ░░ Subject: A start job for unit remote-fs-pre.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit remote-fs-pre.target has finished successfully. ░░ ░░ The job identifier is 211. Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: Reached target Remote File Systems. ░░ Subject: A start job for unit remote-fs.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit remote-fs.target has finished successfully. ░░ ░░ The job identifier is 202. Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: TPM2 PCR Barrier (User) was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/StubPcrKernelImage-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). ░░ Subject: A start job for unit systemd-pcrphase.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-pcrphase.service has finished successfully. ░░ ░░ The job identifier is 140. Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: Started Network Manager Script Dispatcher Service. ░░ Subject: A start job for unit NetworkManager-dispatcher.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit NetworkManager-dispatcher.service has finished successfully. ░░ ░░ The job identifier is 397. Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com NetworkManager[629]: [1707235785.7665] device (lo): state change: ip-check -> secondaries (reason 'none', sys-iface-state: 'external') Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com NetworkManager[629]: [1707235785.7669] device (eth0): state change: ip-check -> secondaries (reason 'none', sys-iface-state: 'managed') Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com NetworkManager[629]: [1707235785.7674] device (lo): state change: secondaries -> activated (reason 'none', sys-iface-state: 'external') Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com NetworkManager[629]: [1707235785.7688] device (lo): Activation: successful, device activated. Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com NetworkManager[629]: [1707235785.7714] device (eth0): state change: secondaries -> activated (reason 'none', sys-iface-state: 'managed') Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com NetworkManager[629]: [1707235785.7725] manager: NetworkManager state is now CONNECTED_SITE Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com NetworkManager[629]: [1707235785.7737] device (eth0): Activation: successful, device activated. Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com NetworkManager[629]: [1707235785.7742] manager: NetworkManager state is now CONNECTED_GLOBAL Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com NetworkManager[629]: [1707235785.7748] manager: startup complete Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: Finished Network Manager Wait Online. ░░ Subject: A start job for unit NetworkManager-wait-online.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit NetworkManager-wait-online.service has finished successfully. ░░ ░░ The job identifier is 184. Feb 06 16:09:45 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: Starting Initial cloud-init job (metadata service crawler)... ░░ Subject: A start job for unit cloud-init.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit cloud-init.service has begun execution. ░░ ░░ The job identifier is 215. Feb 06 16:09:46 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: Cloud-init v. 23.4-2.el9 running 'init' at Tue, 06 Feb 2024 16:09:46 +0000. Up 32.95 seconds. Feb 06 16:09:46 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: ci-info: +++++++++++++++++++++++++++++++++++++++Net device info+++++++++++++++++++++++++++++++++++++++ Feb 06 16:09:46 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: ci-info: +--------+------+------------------------------+---------------+--------+-------------------+ Feb 06 16:09:46 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: ci-info: | Device | Up | Address | Mask | Scope | Hw-Address | Feb 06 16:09:46 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: ci-info: +--------+------+------------------------------+---------------+--------+-------------------+ Feb 06 16:09:46 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: ci-info: | eth0 | True | 10.31.11.33 | 255.255.252.0 | global | 12:07:23:9c:5c:03 | Feb 06 16:09:46 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: ci-info: | eth0 | True | fe80::1007:23ff:fe9c:5c03/64 | . | link | 12:07:23:9c:5c:03 | Feb 06 16:09:46 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: ci-info: | lo | True | 127.0.0.1 | 255.0.0.0 | host | . | Feb 06 16:09:46 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: ci-info: | lo | True | ::1/128 | . | host | . | Feb 06 16:09:46 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: ci-info: +--------+------+------------------------------+---------------+--------+-------------------+ Feb 06 16:09:46 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: ci-info: ++++++++++++++++++++++++++++Route IPv4 info++++++++++++++++++++++++++++ Feb 06 16:09:46 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: ci-info: +-------+-------------+-----------+---------------+-----------+-------+ Feb 06 16:09:46 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: ci-info: | Route | Destination | Gateway | Genmask | Interface | Flags | Feb 06 16:09:46 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: ci-info: +-------+-------------+-----------+---------------+-----------+-------+ Feb 06 16:09:46 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: ci-info: | 0 | 0.0.0.0 | 10.31.8.1 | 0.0.0.0 | eth0 | UG | Feb 06 16:09:46 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: ci-info: | 1 | 10.31.8.0 | 0.0.0.0 | 255.255.252.0 | eth0 | U | Feb 06 16:09:46 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: ci-info: +-------+-------------+-----------+---------------+-----------+-------+ Feb 06 16:09:46 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: ci-info: +++++++++++++++++++Route IPv6 info+++++++++++++++++++ Feb 06 16:09:46 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: ci-info: +-------+-------------+---------+-----------+-------+ Feb 06 16:09:46 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: ci-info: | Route | Destination | Gateway | Interface | Flags | Feb 06 16:09:46 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: ci-info: +-------+-------------+---------+-----------+-------+ Feb 06 16:09:46 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: ci-info: | 1 | fe80::/64 | :: | eth0 | U | Feb 06 16:09:46 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: ci-info: | 3 | multicast | :: | eth0 | U | Feb 06 16:09:46 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: ci-info: +-------+-------------+---------+-----------+-------+ Feb 06 16:09:46 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: 2024-02-06 16:09:46,202 - __init__.py[WARNING]: Unhandled non-multipart (text/x-not-multipart) userdata: 'b'~/.citool.d/post-install'...' Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: Generating public/private rsa key pair. Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: Your identification has been saved in /etc/ssh/ssh_host_rsa_key Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: Your public key has been saved in /etc/ssh/ssh_host_rsa_key.pub Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: The key fingerprint is: Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: SHA256:UhgjPOr9FAsjeOaJ87kwnxT0/7xp89rU4Lt4fcfwwiI root@ip-10-31-11-33.us-east-1.aws.redhat.com Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: The key's randomart image is: Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: +---[RSA 3072]----+ Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: | .. o | Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: | o. + | Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: | ... .. . | Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: |..=.o .. | Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: | *.+.o.oS . | Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: |o +...o. . o . | Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: |oo.. o. o... + | Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: | =o. .oo=E.o + +| Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: | +. .B*=o o o | Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: +----[SHA256]-----+ Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: Generating public/private dsa key pair. Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: Your identification has been saved in /etc/ssh/ssh_host_dsa_key Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: Your public key has been saved in /etc/ssh/ssh_host_dsa_key.pub Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: The key fingerprint is: Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: SHA256:7lf//6rHOqGQWkXLR5+JfK22+wHU4xF6eEjI2AVaNRo root@ip-10-31-11-33.us-east-1.aws.redhat.com Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: The key's randomart image is: Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: +---[DSA 1024]----+ Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: | +E== . | Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: | .++= * .| Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: | + = B X | Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: | + = O +| Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: | So . o o | Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: | .+ o + | Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: | o.. o = o | Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: | .. o . = .| Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: | .. o=o=B| Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: +----[SHA256]-----+ Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: Generating public/private ecdsa key pair. Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: Your identification has been saved in /etc/ssh/ssh_host_ecdsa_key Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: Your public key has been saved in /etc/ssh/ssh_host_ecdsa_key.pub Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: The key fingerprint is: Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: SHA256:wGxnVZMkq682E/zDRSvVnSV0xf5/i6dM23gL9OMg1Bo root@ip-10-31-11-33.us-east-1.aws.redhat.com Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: The key's randomart image is: Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: +---[ECDSA 256]---+ Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: | oo+o..+| Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: | o . o....o| Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: | = o . . +o| Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: | . + . + ..o| Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: | .S E + .| Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: | o.o * . .| Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: | +.= + o.| Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: | +.+ + Oo+| Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: | ..o . *+=+| Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: +----[SHA256]-----+ Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: Generating public/private ed25519 key pair. Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: Your identification has been saved in /etc/ssh/ssh_host_ed25519_key Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: Your public key has been saved in /etc/ssh/ssh_host_ed25519_key.pub Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: The key fingerprint is: Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: SHA256:LkqUa3R14ycCqGNQopizTQe1kmyWZSisHx0f47d/MlQ root@ip-10-31-11-33.us-east-1.aws.redhat.com Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: The key's randomart image is: Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: +--[ED25519 256]--+ Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: |o oo+ | Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: |+*.*.oo | Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: |B.O.++.o. o | Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: |.O.+..oo.o .E | Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: |..=.+ ..S.o.. | Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: | ..+ o ....o | Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: | + . .o | Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: | o . . + . | Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: | . + | Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[719]: +----[SHA256]-----+ Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: Finished Initial cloud-init job (metadata service crawler). ░░ Subject: A start job for unit cloud-init.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit cloud-init.service has finished successfully. ░░ ░░ The job identifier is 215. Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: Reached target Cloud-config availability. ░░ Subject: A start job for unit cloud-config.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit cloud-config.target has finished successfully. ░░ ░░ The job identifier is 224. Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: Reached target Network is Online. ░░ Subject: A start job for unit network-online.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit network-online.target has finished successfully. ░░ ░░ The job identifier is 183. Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: Starting Apply the settings specified in cloud-config... ░░ Subject: A start job for unit cloud-config.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit cloud-config.service has begun execution. ░░ ░░ The job identifier is 223. Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: Starting Crash recovery kernel arming... ░░ Subject: A start job for unit kdump.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit kdump.service has begun execution. ░░ ░░ The job identifier is 239. Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: Starting The restraint harness.... ░░ Subject: A start job for unit restraintd.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit restraintd.service has begun execution. ░░ ░░ The job identifier is 238. Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: Starting Notify NFS peers of a restart... ░░ Subject: A start job for unit rpc-statd-notify.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit rpc-statd-notify.service has begun execution. ░░ ░░ The job identifier is 210. Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: Starting System Logging Service... ░░ Subject: A start job for unit rsyslog.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit rsyslog.service has begun execution. ░░ ░░ The job identifier is 229. Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: Starting OpenSSH server daemon... ░░ Subject: A start job for unit sshd.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sshd.service has begun execution. ░░ ░░ The job identifier is 217. Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com sshd[798]: Server listening on 0.0.0.0 port 22. Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com sshd[798]: Server listening on :: port 22. Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: Started OpenSSH server daemon. ░░ Subject: A start job for unit sshd.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sshd.service has finished successfully. ░░ ░░ The job identifier is 217. Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com sm-notify[796]: Version 2.5.4 starting Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: Started Notify NFS peers of a restart. ░░ Subject: A start job for unit rpc-statd-notify.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit rpc-statd-notify.service has finished successfully. ░░ ░░ The job identifier is 210. Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: Started The restraint harness.. ░░ Subject: A start job for unit restraintd.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit restraintd.service has finished successfully. ░░ ░░ The job identifier is 238. Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com rsyslogd[797]: [origin software="rsyslogd" swVersion="8.2310.0-3.el9" x-pid="797" x-info="https://www.rsyslog.com"] start Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: Started System Logging Service. ░░ Subject: A start job for unit rsyslog.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit rsyslog.service has finished successfully. ░░ ░░ The job identifier is 229. Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[835]: Cloud-init v. 23.4-2.el9 running 'modules:config' at Tue, 06 Feb 2024 16:09:49 +0000. Up 36.54 seconds. Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com rsyslogd[797]: imjournal: journal files changed, reloading... [v8.2310.0-3.el9 try https://www.rsyslog.com/e/0 ] Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com sshd[798]: Received signal 15; terminating. Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: Stopping OpenSSH server daemon... ░░ Subject: A stop job for unit sshd.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A stop job for unit sshd.service has begun execution. ░░ ░░ The job identifier is 474. Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: sshd.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit sshd.service has successfully entered the 'dead' state. Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: Stopped OpenSSH server daemon. ░░ Subject: A stop job for unit sshd.service has finished ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A stop job for unit sshd.service has finished. ░░ ░░ The job identifier is 474 and the job result is done. Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: Stopped target sshd-keygen.target. ░░ Subject: A stop job for unit sshd-keygen.target has finished ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A stop job for unit sshd-keygen.target has finished. ░░ ░░ The job identifier is 541 and the job result is done. Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: Stopping sshd-keygen.target... ░░ Subject: A stop job for unit sshd-keygen.target has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A stop job for unit sshd-keygen.target has begun execution. ░░ ░░ The job identifier is 541. Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: OpenSSH ecdsa Server Key Generation was skipped because of an unmet condition check (ConditionPathExists=!/run/systemd/generator.early/multi-user.target.wants/cloud-init.target). ░░ Subject: A start job for unit sshd-keygen@ecdsa.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sshd-keygen@ecdsa.service has finished successfully. ░░ ░░ The job identifier is 540. Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: OpenSSH ed25519 Server Key Generation was skipped because of an unmet condition check (ConditionPathExists=!/run/systemd/generator.early/multi-user.target.wants/cloud-init.target). ░░ Subject: A start job for unit sshd-keygen@ed25519.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sshd-keygen@ed25519.service has finished successfully. ░░ ░░ The job identifier is 539. Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: OpenSSH rsa Server Key Generation was skipped because of an unmet condition check (ConditionPathExists=!/run/systemd/generator.early/multi-user.target.wants/cloud-init.target). ░░ Subject: A start job for unit sshd-keygen@rsa.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sshd-keygen@rsa.service has finished successfully. ░░ ░░ The job identifier is 537. Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: Reached target sshd-keygen.target. ░░ Subject: A start job for unit sshd-keygen.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sshd-keygen.target has finished successfully. ░░ ░░ The job identifier is 541. Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: Starting OpenSSH server daemon... ░░ Subject: A start job for unit sshd.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sshd.service has begun execution. ░░ ░░ The job identifier is 474. Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com sshd[855]: Server listening on 0.0.0.0 port 22. Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com sshd[855]: Server listening on :: port 22. Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: Started OpenSSH server daemon. ░░ Subject: A start job for unit sshd.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sshd.service has finished successfully. ░░ ░░ The job identifier is 474. Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: Finished Apply the settings specified in cloud-config. ░░ Subject: A start job for unit cloud-config.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit cloud-config.service has finished successfully. ░░ ░░ The job identifier is 223. Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: Starting Execute cloud user/final scripts... ░░ Subject: A start job for unit cloud-final.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit cloud-final.service has begun execution. ░░ ░░ The job identifier is 225. Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: Starting Permit User Sessions... ░░ Subject: A start job for unit systemd-user-sessions.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-user-sessions.service has begun execution. ░░ ░░ The job identifier is 198. Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: Finished Permit User Sessions. ░░ Subject: A start job for unit systemd-user-sessions.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-user-sessions.service has finished successfully. ░░ ░░ The job identifier is 198. Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: Started Command Scheduler. ░░ Subject: A start job for unit crond.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit crond.service has finished successfully. ░░ ░░ The job identifier is 201. Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: Started Getty on tty1. ░░ Subject: A start job for unit getty@tty1.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit getty@tty1.service has finished successfully. ░░ ░░ The job identifier is 231. Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: Reached target Login Prompts. ░░ Subject: A start job for unit getty.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit getty.target has finished successfully. ░░ ░░ The job identifier is 230. Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: Reached target Multi-User System. ░░ Subject: A start job for unit multi-user.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit multi-user.target has finished successfully. ░░ ░░ The job identifier is 116. Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: Starting Record Runlevel Change in UTMP... ░░ Subject: A start job for unit systemd-update-utmp-runlevel.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-update-utmp-runlevel.service has begun execution. ░░ ░░ The job identifier is 212. Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: systemd-update-utmp-runlevel.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit systemd-update-utmp-runlevel.service has successfully entered the 'dead' state. Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: Finished Record Runlevel Change in UTMP. ░░ Subject: A start job for unit systemd-update-utmp-runlevel.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-update-utmp-runlevel.service has finished successfully. ░░ ░░ The job identifier is 212. Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com crond[858]: (CRON) STARTUP (1.5.7) Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com crond[858]: (CRON) INFO (Syslog will be used instead of sendmail.) Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com crond[858]: (CRON) INFO (RANDOM_DELAY will be scaled with factor 59% if used.) Feb 06 16:09:49 ip-10-31-11-33.us-east-1.aws.redhat.com crond[858]: (CRON) INFO (running with inotify support) Feb 06 16:09:50 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[890]: Cloud-init v. 23.4-2.el9 running 'modules:final' at Tue, 06 Feb 2024 16:09:50 +0000. Up 37.04 seconds. Feb 06 16:09:50 ip-10-31-11-33.us-east-1.aws.redhat.com restraintd[814]: Listening on http://localhost:8081 Feb 06 16:09:50 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[892]: ############################################################# Feb 06 16:09:50 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[894]: -----BEGIN SSH HOST KEY FINGERPRINTS----- Feb 06 16:09:50 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[906]: 1024 SHA256:7lf//6rHOqGQWkXLR5+JfK22+wHU4xF6eEjI2AVaNRo root@ip-10-31-11-33.us-east-1.aws.redhat.com (DSA) Feb 06 16:09:50 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[909]: 256 SHA256:wGxnVZMkq682E/zDRSvVnSV0xf5/i6dM23gL9OMg1Bo root@ip-10-31-11-33.us-east-1.aws.redhat.com (ECDSA) Feb 06 16:09:50 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[918]: 256 SHA256:LkqUa3R14ycCqGNQopizTQe1kmyWZSisHx0f47d/MlQ root@ip-10-31-11-33.us-east-1.aws.redhat.com (ED25519) Feb 06 16:09:50 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[926]: 3072 SHA256:UhgjPOr9FAsjeOaJ87kwnxT0/7xp89rU4Lt4fcfwwiI root@ip-10-31-11-33.us-east-1.aws.redhat.com (RSA) Feb 06 16:09:50 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[930]: -----END SSH HOST KEY FINGERPRINTS----- Feb 06 16:09:50 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[933]: ############################################################# Feb 06 16:09:50 ip-10-31-11-33.us-east-1.aws.redhat.com kdumpctl[801]: kdump: Detected change(s) in the following file(s): /etc/fstab Feb 06 16:09:50 ip-10-31-11-33.us-east-1.aws.redhat.com cloud-init[890]: Cloud-init v. 23.4-2.el9 finished at Tue, 06 Feb 2024 16:09:50 +0000. Datasource DataSourceEc2Local. Up 37.24 seconds Feb 06 16:09:50 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: Finished Execute cloud user/final scripts. ░░ Subject: A start job for unit cloud-final.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit cloud-final.service has finished successfully. ░░ ░░ The job identifier is 225. Feb 06 16:09:50 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: Reached target Cloud-init target. ░░ Subject: A start job for unit cloud-init.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit cloud-init.target has finished successfully. ░░ ░░ The job identifier is 214. Feb 06 16:09:52 ip-10-31-11-33.us-east-1.aws.redhat.com chronyd[603]: Selected source 10.2.32.38 Feb 06 16:09:52 ip-10-31-11-33.us-east-1.aws.redhat.com chronyd[603]: System clock TAI offset set to 37 seconds Feb 06 16:09:55 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: NetworkManager-dispatcher.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit NetworkManager-dispatcher.service has successfully entered the 'dead' state. Feb 06 16:09:56 ip-10-31-11-33.us-east-1.aws.redhat.com kernel: block xvda: the capability attribute has been deprecated. Feb 06 16:09:56 ip-10-31-11-33.us-east-1.aws.redhat.com kdumpctl[801]: kdump: Rebuilding /boot/initramfs-5.14.0-410.el9.x86_64kdump.img Feb 06 16:09:58 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1417]: dracut-057-53.git20240104.el9 Feb 06 16:09:58 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: Executing: /usr/bin/dracut --add kdumpbase --quiet --hostonly --hostonly-cmdline --hostonly-i18n --hostonly-mode strict --hostonly-nics -o "plymouth resume ifcfg earlykdump" --mount "/dev/disk/by-uuid/f72128f2-bcfd-46ab-b509-3a36632be876 /sysroot xfs rw,relatime,seclabel,attr2,inode64,logbufs=8,logbsize=32k,noquota" --squash-compressor zstd --no-hostonly-default-device -f /boot/initramfs-5.14.0-410.el9.x86_64kdump.img 5.14.0-410.el9.x86_64 Feb 06 16:10:00 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: dracut module 'systemd-networkd' will not be installed, because command 'networkctl' could not be found! Feb 06 16:10:00 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: dracut module 'systemd-networkd' will not be installed, because command '/usr/lib/systemd/systemd-networkd' could not be found! Feb 06 16:10:00 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: dracut module 'systemd-networkd' will not be installed, because command '/usr/lib/systemd/systemd-networkd-wait-online' could not be found! Feb 06 16:10:00 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: dracut module 'systemd-resolved' will not be installed, because command 'resolvectl' could not be found! Feb 06 16:10:00 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: dracut module 'systemd-resolved' will not be installed, because command '/usr/lib/systemd/systemd-resolved' could not be found! Feb 06 16:10:00 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: dracut module 'systemd-timesyncd' will not be installed, because command '/usr/lib/systemd/systemd-timesyncd' could not be found! Feb 06 16:10:00 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: dracut module 'systemd-timesyncd' will not be installed, because command '/usr/lib/systemd/systemd-time-wait-sync' could not be found! Feb 06 16:10:00 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: dracut module 'busybox' will not be installed, because command 'busybox' could not be found! Feb 06 16:10:00 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: dracut module 'dbus-daemon' will not be installed, because command 'dbus-daemon' could not be found! Feb 06 16:10:00 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: dracut module 'connman' will not be installed, because command 'connmand' could not be found! Feb 06 16:10:00 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: dracut module 'connman' will not be installed, because command 'connmanctl' could not be found! Feb 06 16:10:00 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: dracut module 'connman' will not be installed, because command 'connmand-wait-online' could not be found! Feb 06 16:10:00 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: dracut module 'network-wicked' will not be installed, because command 'wicked' could not be found! Feb 06 16:10:00 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: dracut module 'ifcfg' will not be installed, because it's in the list to be omitted! Feb 06 16:10:00 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: dracut module 'plymouth' will not be installed, because it's in the list to be omitted! Feb 06 16:10:00 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: 62bluetooth: Could not find any command of '/usr/lib/bluetooth/bluetoothd /usr/libexec/bluetooth/bluetoothd'! Feb 06 16:10:00 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: dracut module 'lvmmerge' will not be installed, because command 'lvm' could not be found! Feb 06 16:10:00 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: dracut module 'lvmthinpool-monitor' will not be installed, because command 'lvm' could not be found! Feb 06 16:10:00 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: dracut module 'btrfs' will not be installed, because command 'btrfs' could not be found! Feb 06 16:10:00 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: dracut module 'dmraid' will not be installed, because command 'dmraid' could not be found! Feb 06 16:10:00 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: dracut module 'lvm' will not be installed, because command 'lvm' could not be found! Feb 06 16:10:00 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: dracut module 'mdraid' will not be installed, because command 'mdadm' could not be found! Feb 06 16:10:00 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: dracut module 'pcsc' will not be installed, because command 'pcscd' could not be found! Feb 06 16:10:00 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: dracut module 'tpm2-tss' will not be installed, because command 'tpm2' could not be found! Feb 06 16:10:00 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: dracut module 'cifs' will not be installed, because command 'mount.cifs' could not be found! Feb 06 16:10:00 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: dracut module 'iscsi' will not be installed, because command 'iscsi-iname' could not be found! Feb 06 16:10:00 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: dracut module 'iscsi' will not be installed, because command 'iscsiadm' could not be found! Feb 06 16:10:00 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: dracut module 'iscsi' will not be installed, because command 'iscsid' could not be found! Feb 06 16:10:00 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: dracut module 'nvmf' will not be installed, because command 'nvme' could not be found! Feb 06 16:10:00 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: dracut module 'resume' will not be installed, because it's in the list to be omitted! Feb 06 16:10:00 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: dracut module 'biosdevname' will not be installed, because command 'biosdevname' could not be found! Feb 06 16:10:00 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: dracut module 'earlykdump' will not be installed, because it's in the list to be omitted! Feb 06 16:10:01 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: dracut module 'memstrack' will not be installed, because command 'memstrack' could not be found! Feb 06 16:10:01 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: memstrack is not available Feb 06 16:10:01 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: If you need to use rd.memdebug>=4, please install memstrack and procps-ng Feb 06 16:10:01 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: dracut module 'systemd-resolved' will not be installed, because command 'resolvectl' could not be found! Feb 06 16:10:01 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: dracut module 'systemd-resolved' will not be installed, because command '/usr/lib/systemd/systemd-resolved' could not be found! Feb 06 16:10:01 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: dracut module 'systemd-timesyncd' will not be installed, because command '/usr/lib/systemd/systemd-timesyncd' could not be found! Feb 06 16:10:01 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: dracut module 'systemd-timesyncd' will not be installed, because command '/usr/lib/systemd/systemd-time-wait-sync' could not be found! Feb 06 16:10:01 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: dracut module 'busybox' will not be installed, because command 'busybox' could not be found! Feb 06 16:10:01 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: dracut module 'dbus-daemon' will not be installed, because command 'dbus-daemon' could not be found! Feb 06 16:10:01 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: dracut module 'connman' will not be installed, because command 'connmand' could not be found! Feb 06 16:10:01 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: dracut module 'connman' will not be installed, because command 'connmanctl' could not be found! Feb 06 16:10:01 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: dracut module 'connman' will not be installed, because command 'connmand-wait-online' could not be found! Feb 06 16:10:01 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: dracut module 'network-wicked' will not be installed, because command 'wicked' could not be found! Feb 06 16:10:01 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: 62bluetooth: Could not find any command of '/usr/lib/bluetooth/bluetoothd /usr/libexec/bluetooth/bluetoothd'! Feb 06 16:10:01 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: dracut module 'lvmmerge' will not be installed, because command 'lvm' could not be found! Feb 06 16:10:01 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: dracut module 'lvmthinpool-monitor' will not be installed, because command 'lvm' could not be found! Feb 06 16:10:01 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: dracut module 'btrfs' will not be installed, because command 'btrfs' could not be found! Feb 06 16:10:01 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: dracut module 'dmraid' will not be installed, because command 'dmraid' could not be found! Feb 06 16:10:01 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: dracut module 'lvm' will not be installed, because command 'lvm' could not be found! Feb 06 16:10:01 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: dracut module 'mdraid' will not be installed, because command 'mdadm' could not be found! Feb 06 16:10:01 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: dracut module 'pcsc' will not be installed, because command 'pcscd' could not be found! Feb 06 16:10:01 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: dracut module 'tpm2-tss' will not be installed, because command 'tpm2' could not be found! Feb 06 16:10:01 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: dracut module 'cifs' will not be installed, because command 'mount.cifs' could not be found! Feb 06 16:10:01 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: dracut module 'iscsi' will not be installed, because command 'iscsi-iname' could not be found! Feb 06 16:10:01 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: dracut module 'iscsi' will not be installed, because command 'iscsiadm' could not be found! Feb 06 16:10:01 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: dracut module 'iscsi' will not be installed, because command 'iscsid' could not be found! Feb 06 16:10:01 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: dracut module 'nvmf' will not be installed, because command 'nvme' could not be found! Feb 06 16:10:01 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: dracut module 'memstrack' will not be installed, because command 'memstrack' could not be found! Feb 06 16:10:01 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: memstrack is not available Feb 06 16:10:01 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: If you need to use rd.memdebug>=4, please install memstrack and procps-ng Feb 06 16:10:01 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: *** Including module: systemd *** Feb 06 16:10:01 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: *** Including module: systemd-initrd *** Feb 06 16:10:01 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: *** Including module: nss-softokn *** Feb 06 16:10:01 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: *** Including module: rngd *** Feb 06 16:10:01 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: *** Including module: i18n *** Feb 06 16:10:02 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: *** Including module: drm *** Feb 06 16:10:02 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: *** Including module: prefixdevname *** Feb 06 16:10:02 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: *** Including module: kernel-modules *** Feb 06 16:10:03 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: *** Including module: kernel-modules-extra *** Feb 06 16:10:03 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: kernel-modules-extra: configuration source "/run/depmod.d" does not exist Feb 06 16:10:03 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: kernel-modules-extra: configuration source "/lib/depmod.d" does not exist Feb 06 16:10:03 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: kernel-modules-extra: parsing configuration file "/etc/depmod.d/dist.conf" Feb 06 16:10:03 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: kernel-modules-extra: /etc/depmod.d/dist.conf: added "updates extra built-in weak-updates" to the list of search directories Feb 06 16:10:03 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: *** Including module: fstab-sys *** Feb 06 16:10:03 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: *** Including module: rootfs-block *** Feb 06 16:10:03 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: *** Including module: terminfo *** Feb 06 16:10:03 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: *** Including module: udev-rules *** Feb 06 16:10:04 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: Skipping udev rule: 91-permissions.rules Feb 06 16:10:04 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: Skipping udev rule: 80-drivers-modprobe.rules Feb 06 16:10:04 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: *** Including module: dracut-systemd *** Feb 06 16:10:04 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: *** Including module: usrmount *** Feb 06 16:10:04 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: *** Including module: base *** Feb 06 16:10:04 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: *** Including module: fs-lib *** Feb 06 16:10:04 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: *** Including module: kdumpbase *** Feb 06 16:10:04 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: *** Including module: microcode_ctl-fw_dir_override *** Feb 06 16:10:04 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: microcode_ctl module: mangling fw_dir Feb 06 16:10:04 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: microcode_ctl: reset fw_dir to "/lib/firmware/updates /lib/firmware" Feb 06 16:10:04 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: microcode_ctl: processing data directory "/usr/share/microcode_ctl/ucode_with_caveats/intel"... Feb 06 16:10:04 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: microcode_ctl: intel: caveats check for kernel version "5.14.0-410.el9.x86_64" passed, adding "/usr/share/microcode_ctl/ucode_with_caveats/intel" to fw_dir variable Feb 06 16:10:04 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: microcode_ctl: processing data directory "/usr/share/microcode_ctl/ucode_with_caveats/intel-06-2d-07"... Feb 06 16:10:04 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: microcode_ctl: configuration "intel-06-2d-07" is ignored Feb 06 16:10:04 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: microcode_ctl: processing data directory "/usr/share/microcode_ctl/ucode_with_caveats/intel-06-4e-03"... Feb 06 16:10:04 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: microcode_ctl: configuration "intel-06-4e-03" is ignored Feb 06 16:10:04 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: microcode_ctl: processing data directory "/usr/share/microcode_ctl/ucode_with_caveats/intel-06-4f-01"... Feb 06 16:10:05 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: microcode_ctl: configuration "intel-06-4f-01" is ignored Feb 06 16:10:05 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: microcode_ctl: processing data directory "/usr/share/microcode_ctl/ucode_with_caveats/intel-06-55-04"... Feb 06 16:10:05 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: microcode_ctl: configuration "intel-06-55-04" is ignored Feb 06 16:10:05 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: microcode_ctl: processing data directory "/usr/share/microcode_ctl/ucode_with_caveats/intel-06-5e-03"... Feb 06 16:10:05 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: microcode_ctl: configuration "intel-06-5e-03" is ignored Feb 06 16:10:05 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: microcode_ctl: processing data directory "/usr/share/microcode_ctl/ucode_with_caveats/intel-06-8c-01"... Feb 06 16:10:05 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: microcode_ctl: configuration "intel-06-8c-01" is ignored Feb 06 16:10:05 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: microcode_ctl: processing data directory "/usr/share/microcode_ctl/ucode_with_caveats/intel-06-8e-9e-0x-0xca"... Feb 06 16:10:05 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: microcode_ctl: configuration "intel-06-8e-9e-0x-0xca" is ignored Feb 06 16:10:05 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: microcode_ctl: processing data directory "/usr/share/microcode_ctl/ucode_with_caveats/intel-06-8e-9e-0x-dell"... Feb 06 16:10:05 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: microcode_ctl: configuration "intel-06-8e-9e-0x-dell" is ignored Feb 06 16:10:05 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: microcode_ctl: final fw_dir: "/usr/share/microcode_ctl/ucode_with_caveats/intel /lib/firmware/updates /lib/firmware" Feb 06 16:10:05 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: *** Including module: shutdown *** Feb 06 16:10:05 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: *** Including module: squash *** Feb 06 16:10:05 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: *** Including modules done *** Feb 06 16:10:05 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: *** Installing kernel module dependencies *** Feb 06 16:10:06 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: *** Installing kernel module dependencies done *** Feb 06 16:10:06 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: *** Resolving executable dependencies *** Feb 06 16:10:08 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: *** Resolving executable dependencies done *** Feb 06 16:10:08 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: *** Hardlinking files *** Feb 06 16:10:08 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: Mode: real Feb 06 16:10:08 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: Files: 436 Feb 06 16:10:08 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: Linked: 1 files Feb 06 16:10:08 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: Compared: 0 xattrs Feb 06 16:10:08 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: Compared: 6 files Feb 06 16:10:08 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: Saved: 55.98 KiB Feb 06 16:10:08 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: Duration: 0.007392 seconds Feb 06 16:10:08 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: *** Hardlinking files done *** Feb 06 16:10:08 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: *** Generating early-microcode cpio image *** Feb 06 16:10:08 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: *** Constructing GenuineIntel.bin *** Feb 06 16:10:08 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: *** Constructing GenuineIntel.bin *** Feb 06 16:10:08 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: *** Store current command line parameters *** Feb 06 16:10:08 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: Stored kernel commandline: Feb 06 16:10:08 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: No dracut internal kernel commandline stored in the initramfs Feb 06 16:10:08 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: *** Install squash loader *** Feb 06 16:10:09 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: *** Stripping files *** Feb 06 16:10:10 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: *** Stripping files done *** Feb 06 16:10:10 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: *** Squashing the files inside the initramfs *** Feb 06 16:10:15 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: systemd-hostnamed.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit systemd-hostnamed.service has successfully entered the 'dead' state. Feb 06 16:10:16 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: *** Squashing the files inside the initramfs done *** Feb 06 16:10:16 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: *** Creating image file '/boot/initramfs-5.14.0-410.el9.x86_64kdump.img' *** Feb 06 16:10:16 ip-10-31-11-33.us-east-1.aws.redhat.com dracut[1419]: *** Creating initramfs image file '/boot/initramfs-5.14.0-410.el9.x86_64kdump.img' done *** Feb 06 16:10:17 ip-10-31-11-33.us-east-1.aws.redhat.com kdumpctl[801]: kdump: kexec: loaded kdump kernel Feb 06 16:10:17 ip-10-31-11-33.us-east-1.aws.redhat.com kdumpctl[801]: kdump: Starting kdump: [OK] Feb 06 16:10:17 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: Finished Crash recovery kernel arming. ░░ Subject: A start job for unit kdump.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit kdump.service has finished successfully. ░░ ░░ The job identifier is 239. Feb 06 16:10:17 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: Startup finished in 915ms (kernel) + 8.108s (initrd) + 55.115s (userspace) = 1min 4.139s. ░░ Subject: System start-up is now complete ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ All system services necessary queued for starting at boot have been ░░ started. Note that this does not mean that the machine is now idle as services ░░ might still be busy with completing start-up. ░░ ░░ Kernel start-up required 915029 microseconds. ░░ ░░ Initrd start-up required 8108853 microseconds. ░░ ░░ Userspace start-up required 55115832 microseconds. Feb 06 16:10:46 ip-10-31-11-33.us-east-1.aws.redhat.com sshd[4043]: Accepted publickey for root from 10.30.32.29 port 48242 ssh2: RSA SHA256:W3cSdmPJK+d9RwU97ardijPXIZnxHswrpTHWW9oYtEU Feb 06 16:10:46 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: Created slice User Slice of UID 0. ░░ Subject: A start job for unit user-0.slice has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit user-0.slice has finished successfully. ░░ ░░ The job identifier is 605. Feb 06 16:10:46 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: Starting User Runtime Directory /run/user/0... ░░ Subject: A start job for unit user-runtime-dir@0.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit user-runtime-dir@0.service has begun execution. ░░ ░░ The job identifier is 543. Feb 06 16:10:46 ip-10-31-11-33.us-east-1.aws.redhat.com systemd-logind[583]: New session 1 of user root. ░░ Subject: A new session 1 has been created for user root ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ Documentation: sd-login(3) ░░ ░░ A new session with the ID 1 has been created for the user root. ░░ ░░ The leading process of the session is 4043. Feb 06 16:10:47 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: Finished User Runtime Directory /run/user/0. ░░ Subject: A start job for unit user-runtime-dir@0.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit user-runtime-dir@0.service has finished successfully. ░░ ░░ The job identifier is 543. Feb 06 16:10:47 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: Starting User Manager for UID 0... ░░ Subject: A start job for unit user@0.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit user@0.service has begun execution. ░░ ░░ The job identifier is 542. Feb 06 16:10:47 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[4047]: pam_unix(systemd-user:session): session opened for user root(uid=0) by (uid=0) Feb 06 16:10:47 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[4047]: Queued start job for default target Main User Target. Feb 06 16:10:47 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[4047]: Created slice User Application Slice. ░░ Subject: A start job for unit UNIT has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has finished successfully. ░░ ░░ The job identifier is 9. Feb 06 16:10:47 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[4047]: Mark boot as successful after the user session has run 2 minutes was skipped because of an unmet condition check (ConditionUser=!@system). ░░ Subject: A start job for unit UNIT has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has finished successfully. ░░ ░░ The job identifier is 4. Feb 06 16:10:47 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[4047]: Started Daily Cleanup of User's Temporary Directories. ░░ Subject: A start job for unit UNIT has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has finished successfully. ░░ ░░ The job identifier is 6. Feb 06 16:10:47 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[4047]: Reached target Paths. ░░ Subject: A start job for unit UNIT has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has finished successfully. ░░ ░░ The job identifier is 11. Feb 06 16:10:47 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[4047]: Reached target Timers. ░░ Subject: A start job for unit UNIT has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has finished successfully. ░░ ░░ The job identifier is 3. Feb 06 16:10:47 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[4047]: Starting D-Bus User Message Bus Socket... ░░ Subject: A start job for unit UNIT has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has begun execution. ░░ ░░ The job identifier is 8. Feb 06 16:10:47 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[4047]: Starting Create User's Volatile Files and Directories... ░░ Subject: A start job for unit UNIT has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has begun execution. ░░ ░░ The job identifier is 12. Feb 06 16:10:47 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[4047]: Finished Create User's Volatile Files and Directories. ░░ Subject: A start job for unit UNIT has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has finished successfully. ░░ ░░ The job identifier is 12. Feb 06 16:10:47 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[4047]: Listening on D-Bus User Message Bus Socket. ░░ Subject: A start job for unit UNIT has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has finished successfully. ░░ ░░ The job identifier is 8. Feb 06 16:10:47 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[4047]: Reached target Sockets. ░░ Subject: A start job for unit UNIT has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has finished successfully. ░░ ░░ The job identifier is 7. Feb 06 16:10:47 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[4047]: Reached target Basic System. ░░ Subject: A start job for unit UNIT has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has finished successfully. ░░ ░░ The job identifier is 2. Feb 06 16:10:47 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[4047]: Reached target Main User Target. ░░ Subject: A start job for unit UNIT has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has finished successfully. ░░ ░░ The job identifier is 1. Feb 06 16:10:47 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[4047]: Startup finished in 72ms. ░░ Subject: User manager start-up is now complete ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The user manager instance for user 0 has been started. All services queued ░░ for starting have been started. Note that other services might still be starting ░░ up or be started at any later time. ░░ ░░ Startup of the manager took 72368 microseconds. Feb 06 16:10:47 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: Started User Manager for UID 0. ░░ Subject: A start job for unit user@0.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit user@0.service has finished successfully. ░░ ░░ The job identifier is 542. Feb 06 16:10:47 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: Started Session 1 of User root. ░░ Subject: A start job for unit session-1.scope has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit session-1.scope has finished successfully. ░░ ░░ The job identifier is 607. Feb 06 16:10:47 ip-10-31-11-33.us-east-1.aws.redhat.com sshd[4043]: pam_unix(sshd:session): session opened for user root(uid=0) by (uid=0) Feb 06 16:10:47 ip-10-31-11-33.us-east-1.aws.redhat.com sshd[4056]: Received disconnect from 10.30.32.29 port 48242:11: disconnected by user Feb 06 16:10:47 ip-10-31-11-33.us-east-1.aws.redhat.com sshd[4056]: Disconnected from user root 10.30.32.29 port 48242 Feb 06 16:10:47 ip-10-31-11-33.us-east-1.aws.redhat.com sshd[4043]: pam_unix(sshd:session): session closed for user root Feb 06 16:10:47 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: session-1.scope: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit session-1.scope has successfully entered the 'dead' state. Feb 06 16:10:47 ip-10-31-11-33.us-east-1.aws.redhat.com systemd-logind[583]: Session 1 logged out. Waiting for processes to exit. Feb 06 16:10:47 ip-10-31-11-33.us-east-1.aws.redhat.com systemd-logind[583]: Removed session 1. ░░ Subject: Session 1 has been terminated ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ Documentation: sd-login(3) ░░ ░░ A session with the ID 1 has been terminated. Feb 06 16:10:47 ip-10-31-11-33.us-east-1.aws.redhat.com sshd[4085]: error: kex_exchange_identification: Connection closed by remote host Feb 06 16:10:47 ip-10-31-11-33.us-east-1.aws.redhat.com sshd[4085]: Connection closed by 10.29.163.189 port 51532 Feb 06 16:10:47 ip-10-31-11-33.us-east-1.aws.redhat.com sshd[4086]: Accepted publickey for root from 10.29.163.189 port 51536 ssh2: RSA SHA256:W3cSdmPJK+d9RwU97ardijPXIZnxHswrpTHWW9oYtEU Feb 06 16:10:47 ip-10-31-11-33.us-east-1.aws.redhat.com systemd-logind[583]: New session 3 of user root. ░░ Subject: A new session 3 has been created for user root ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ Documentation: sd-login(3) ░░ ░░ A new session with the ID 3 has been created for the user root. ░░ ░░ The leading process of the session is 4086. Feb 06 16:10:47 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: Started Session 3 of User root. ░░ Subject: A start job for unit session-3.scope has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit session-3.scope has finished successfully. ░░ ░░ The job identifier is 673. Feb 06 16:10:47 ip-10-31-11-33.us-east-1.aws.redhat.com sshd[4086]: pam_unix(sshd:session): session opened for user root(uid=0) by (uid=0) Feb 06 16:10:47 ip-10-31-11-33.us-east-1.aws.redhat.com sshd[4089]: Received disconnect from 10.29.163.189 port 51536:11: disconnected by user Feb 06 16:10:47 ip-10-31-11-33.us-east-1.aws.redhat.com sshd[4089]: Disconnected from user root 10.29.163.189 port 51536 Feb 06 16:10:47 ip-10-31-11-33.us-east-1.aws.redhat.com sshd[4086]: pam_unix(sshd:session): session closed for user root Feb 06 16:10:47 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: session-3.scope: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit session-3.scope has successfully entered the 'dead' state. Feb 06 16:10:47 ip-10-31-11-33.us-east-1.aws.redhat.com systemd-logind[583]: Session 3 logged out. Waiting for processes to exit. Feb 06 16:10:47 ip-10-31-11-33.us-east-1.aws.redhat.com systemd-logind[583]: Removed session 3. ░░ Subject: Session 3 has been terminated ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ Documentation: sd-login(3) ░░ ░░ A session with the ID 3 has been terminated. Feb 06 16:10:47 ip-10-31-11-33.us-east-1.aws.redhat.com sshd[4106]: Accepted publickey for root from 10.29.163.189 port 51552 ssh2: RSA SHA256:W3cSdmPJK+d9RwU97ardijPXIZnxHswrpTHWW9oYtEU Feb 06 16:10:47 ip-10-31-11-33.us-east-1.aws.redhat.com systemd-logind[583]: New session 4 of user root. ░░ Subject: A new session 4 has been created for user root ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ Documentation: sd-login(3) ░░ ░░ A new session with the ID 4 has been created for the user root. ░░ ░░ The leading process of the session is 4106. Feb 06 16:10:47 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: Started Session 4 of User root. ░░ Subject: A start job for unit session-4.scope has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit session-4.scope has finished successfully. ░░ ░░ The job identifier is 739. Feb 06 16:10:47 ip-10-31-11-33.us-east-1.aws.redhat.com sshd[4106]: pam_unix(sshd:session): session opened for user root(uid=0) by (uid=0) Feb 06 16:10:47 ip-10-31-11-33.us-east-1.aws.redhat.com sshd[4109]: Received disconnect from 10.29.163.189 port 51552:11: disconnected by user Feb 06 16:10:47 ip-10-31-11-33.us-east-1.aws.redhat.com sshd[4109]: Disconnected from user root 10.29.163.189 port 51552 Feb 06 16:10:47 ip-10-31-11-33.us-east-1.aws.redhat.com sshd[4106]: pam_unix(sshd:session): session closed for user root Feb 06 16:10:47 ip-10-31-11-33.us-east-1.aws.redhat.com systemd-logind[583]: Session 4 logged out. Waiting for processes to exit. Feb 06 16:10:47 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: session-4.scope: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit session-4.scope has successfully entered the 'dead' state. Feb 06 16:10:47 ip-10-31-11-33.us-east-1.aws.redhat.com systemd-logind[583]: Removed session 4. ░░ Subject: Session 4 has been terminated ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ Documentation: sd-login(3) ░░ ░░ A session with the ID 4 has been terminated. Feb 06 16:10:48 ip-10-31-11-33.us-east-1.aws.redhat.com sshd[4126]: Accepted publickey for root from 10.29.163.189 port 49246 ssh2: RSA SHA256:W3cSdmPJK+d9RwU97ardijPXIZnxHswrpTHWW9oYtEU Feb 06 16:10:48 ip-10-31-11-33.us-east-1.aws.redhat.com systemd-logind[583]: New session 5 of user root. ░░ Subject: A new session 5 has been created for user root ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ Documentation: sd-login(3) ░░ ░░ A new session with the ID 5 has been created for the user root. ░░ ░░ The leading process of the session is 4126. Feb 06 16:10:48 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: Started Session 5 of User root. ░░ Subject: A start job for unit session-5.scope has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit session-5.scope has finished successfully. ░░ ░░ The job identifier is 805. Feb 06 16:10:48 ip-10-31-11-33.us-east-1.aws.redhat.com sshd[4126]: pam_unix(sshd:session): session opened for user root(uid=0) by (uid=0) Feb 06 16:10:48 ip-10-31-11-33.us-east-1.aws.redhat.com sshd[4129]: Received disconnect from 10.29.163.189 port 49246:11: disconnected by user Feb 06 16:10:48 ip-10-31-11-33.us-east-1.aws.redhat.com sshd[4129]: Disconnected from user root 10.29.163.189 port 49246 Feb 06 16:10:48 ip-10-31-11-33.us-east-1.aws.redhat.com sshd[4126]: pam_unix(sshd:session): session closed for user root Feb 06 16:10:48 ip-10-31-11-33.us-east-1.aws.redhat.com systemd-logind[583]: Session 5 logged out. Waiting for processes to exit. Feb 06 16:10:48 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: session-5.scope: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit session-5.scope has successfully entered the 'dead' state. Feb 06 16:10:48 ip-10-31-11-33.us-east-1.aws.redhat.com systemd-logind[583]: Removed session 5. ░░ Subject: Session 5 has been terminated ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ Documentation: sd-login(3) ░░ ░░ A session with the ID 5 has been terminated. Feb 06 16:10:49 ip-10-31-11-33.us-east-1.aws.redhat.com sshd[4146]: Accepted publickey for root from 10.29.163.189 port 49250 ssh2: RSA SHA256:W3cSdmPJK+d9RwU97ardijPXIZnxHswrpTHWW9oYtEU Feb 06 16:10:49 ip-10-31-11-33.us-east-1.aws.redhat.com systemd-logind[583]: New session 6 of user root. ░░ Subject: A new session 6 has been created for user root ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ Documentation: sd-login(3) ░░ ░░ A new session with the ID 6 has been created for the user root. ░░ ░░ The leading process of the session is 4146. Feb 06 16:10:49 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: Started Session 6 of User root. ░░ Subject: A start job for unit session-6.scope has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit session-6.scope has finished successfully. ░░ ░░ The job identifier is 871. Feb 06 16:10:49 ip-10-31-11-33.us-east-1.aws.redhat.com sshd[4146]: pam_unix(sshd:session): session opened for user root(uid=0) by (uid=0) Feb 06 16:11:23 ip-10-31-11-33.us-east-1.aws.redhat.com python3[4243]: ansible-ansible.legacy.setup Invoked with gather_subset=['all'] gather_timeout=10 filter=[] fact_path=/etc/ansible/facts.d Feb 06 16:11:25 ip-10-31-11-33.us-east-1.aws.redhat.com python3[4284]: ansible-service_facts Invoked Feb 06 16:11:26 ip-10-31-11-33.us-east-1.aws.redhat.com python3[4382]: ansible-ansible.legacy.command Invoked with _raw_params=rpm -qa _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 06 16:11:27 ip-10-31-11-33.us-east-1.aws.redhat.com python3[4400]: ansible-stat Invoked with path=/etc/yum.repos.d/qa-tools.repo follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Feb 06 16:11:28 ip-10-31-11-33.us-east-1.aws.redhat.com python3[4419]: ansible-ansible.legacy.command Invoked with _raw_params=curl -sLI --connect-timeout 5 -w '%{response_code}' http://liver2.lab.eng.brq.redhat.com/repo/ | grep ^200 _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 06 16:11:29 ip-10-31-11-33.us-east-1.aws.redhat.com python3[4440]: ansible-ansible.legacy.dnf Invoked with name=['ca-certificates'] state=latest allow_downgrade=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 allowerasing=False nobest=False conf_file=None disable_excludes=None download_dir=None list=None releasever=None Feb 06 16:11:30 ip-10-31-11-33.us-east-1.aws.redhat.com python3[4458]: ansible-ansible.legacy.dnf Invoked with name=['curl'] state=latest allow_downgrade=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 allowerasing=False nobest=False conf_file=None disable_excludes=None download_dir=None list=None releasever=None Feb 06 16:11:30 ip-10-31-11-33.us-east-1.aws.redhat.com python3[4476]: ansible-ansible.legacy.command Invoked with _raw_params=curl -skL -o /etc/pki/ca-trust/source/anchors/Current-IT-Root-CAs.pem https://certs.corp.redhat.com/certs/Current-IT-Root-CAs.pem _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 06 16:11:31 ip-10-31-11-33.us-east-1.aws.redhat.com python3[4495]: ansible-ansible.legacy.command Invoked with _raw_params=bash -c 'update-ca-trust' _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 06 16:11:32 ip-10-31-11-33.us-east-1.aws.redhat.com python3[4519]: ansible-ansible.legacy.command Invoked with _raw_params=bash -c 'update-ca-trust extract' _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 06 16:11:33 ip-10-31-11-33.us-east-1.aws.redhat.com python3[4543]: ansible-ansible.legacy.command Invoked with _raw_params=curl -v https://gitlab.cee.redhat.com _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 06 16:11:34 ip-10-31-11-33.us-east-1.aws.redhat.com python3[4562]: ansible-ansible.legacy.command Invoked with _raw_params=curl -v https://beaker.engineering.redhat.com/ _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 06 16:11:35 ip-10-31-11-33.us-east-1.aws.redhat.com python3[4581]: ansible-ansible.legacy.command Invoked with _raw_params=if [ $(find /etc/yum.repos.d/ -name 'centos-addons.repo' | wc -l ) -gt 0 ]; then sed -i "/^\[nfv-source\]/,/^$/d" /etc/yum.repos.d/centos-addons.repo; fi _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 06 16:11:35 ip-10-31-11-33.us-east-1.aws.redhat.com python3[4603]: ansible-file Invoked with path=/etc/yum.repos.d/CentOS-Media.repo state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Feb 06 16:11:35 ip-10-31-11-33.us-east-1.aws.redhat.com python3[4620]: ansible-file Invoked with path=/etc/yum.repos.d/CentOS-Linux-Media.repo state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Feb 06 16:11:35 ip-10-31-11-33.us-east-1.aws.redhat.com python3[4637]: ansible-file Invoked with path=/etc/yum.repos.d/CentOS-Stream-Media.repo state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Feb 06 16:11:36 ip-10-31-11-33.us-east-1.aws.redhat.com python3[4654]: ansible-file Invoked with path=/etc/yum.repos.d/beaker-client-testing.repo state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Feb 06 16:11:36 ip-10-31-11-33.us-east-1.aws.redhat.com python3[4671]: ansible-stat Invoked with path=/etc/yum.repos.d/beaker-client.repo follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Feb 06 16:11:36 ip-10-31-11-33.us-east-1.aws.redhat.com python3[4690]: ansible-replace Invoked with path=/etc/yum.repos.d/beaker-client.repo regexp=7|\$releasever replace=8 backup=False encoding=utf-8 unsafe_writes=False after=None before=None validate=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Feb 06 16:11:36 ip-10-31-11-33.us-east-1.aws.redhat.com python3[4707]: ansible-stat Invoked with path=/etc/yum.repos.d/beaker-client.repo follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Feb 06 16:11:37 ip-10-31-11-33.us-east-1.aws.redhat.com python3[4726]: ansible-community.general.ini_file Invoked with path=/etc/yum.repos.d/beaker-client.repo section=beaker-client option=skip_if_unavailable value=1 backup=False state=present exclusive=True no_extra_spaces=False ignore_spaces=False allow_no_value=False modify_inactive_option=True create=True follow=False unsafe_writes=False values=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Feb 06 16:11:37 ip-10-31-11-33.us-east-1.aws.redhat.com python3[4743]: ansible-ansible.legacy.command Invoked with _raw_params=for repofile in $(ls -1 /etc/yum.repos.d/); do echo "# vvvvv $repofile vvvvv vvvvv"; cat "/etc/yum.repos.d/$repofile"; echo "# ^^^^^ ^^^^^ ^^^^^ ^^^^^"; echo ""; done _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 06 16:11:37 ip-10-31-11-33.us-east-1.aws.redhat.com python3[4774]: ansible-ansible.legacy.command Invoked with _raw_params=yum -q repolist _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 06 16:11:38 ip-10-31-11-33.us-east-1.aws.redhat.com python3[4792]: ansible-stat Invoked with path=/etc/yum.repos.d/baseos-ci.repo follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Feb 06 16:11:38 ip-10-31-11-33.us-east-1.aws.redhat.com python3[4809]: ansible-ansible.legacy.command Invoked with _raw_params=for repofile in $(ls -1 /etc/yum.repos.d/); do echo "# vvvvv $repofile vvvvv vvvvv"; cat "/etc/yum.repos.d/$repofile"; echo "# ^^^^^ ^^^^^ ^^^^^ ^^^^^"; echo ""; done _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 06 16:11:38 ip-10-31-11-33.us-east-1.aws.redhat.com python3[4840]: ansible-ansible.legacy.command Invoked with _raw_params=yum -q repolist _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 06 16:11:39 ip-10-31-11-33.us-east-1.aws.redhat.com python3[4858]: ansible-ansible.legacy.command Invoked with _raw_params=for repofile in $(ls -1 /etc/yum.repos.d/); do echo "# vvvvv $repofile vvvvv vvvvv"; cat "/etc/yum.repos.d/$repofile"; echo "# ^^^^^ ^^^^^ ^^^^^ ^^^^^"; echo ""; done _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 06 16:11:39 ip-10-31-11-33.us-east-1.aws.redhat.com python3[4889]: ansible-ansible.legacy.command Invoked with _raw_params=yum -q repolist _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 06 16:11:41 ip-10-31-11-33.us-east-1.aws.redhat.com python3[4907]: ansible-ansible.legacy.command Invoked with _raw_params=for repofile in $(ls -1 /etc/yum.repos.d/); do echo "# vvvvv $repofile vvvvv vvvvv"; cat "/etc/yum.repos.d/$repofile"; echo "# ^^^^^ ^^^^^ ^^^^^ ^^^^^"; echo ""; done _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 06 16:11:41 ip-10-31-11-33.us-east-1.aws.redhat.com python3[4938]: ansible-ansible.legacy.command Invoked with _raw_params=yum -q repolist _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 06 16:11:41 ip-10-31-11-33.us-east-1.aws.redhat.com python3[4956]: ansible-ansible.legacy.command Invoked with _raw_params=for repofile in $(ls -1 /etc/yum.repos.d/); do echo "# vvvvv $repofile vvvvv vvvvv"; cat "/etc/yum.repos.d/$repofile"; echo "# ^^^^^ ^^^^^ ^^^^^ ^^^^^"; echo ""; done _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 06 16:11:42 ip-10-31-11-33.us-east-1.aws.redhat.com python3[4987]: ansible-ansible.legacy.command Invoked with _raw_params=yum -q repolist _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 06 16:11:42 ip-10-31-11-33.us-east-1.aws.redhat.com python3[5005]: ansible-ansible.legacy.command Invoked with _raw_params=dnf clean all --enablerepo=* _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 06 16:11:43 ip-10-31-11-33.us-east-1.aws.redhat.com python3[5023]: ansible-ansible.legacy.command Invoked with _raw_params=dnf makecache --enablerepo=* _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 06 16:12:52 ip-10-31-11-33.us-east-1.aws.redhat.com python3[5220]: ansible-ansible.legacy.dnf Invoked with name=['createrepo'] state=latest allow_downgrade=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 allowerasing=False nobest=False conf_file=None disable_excludes=None download_dir=None list=None releasever=None Feb 06 16:13:00 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: Started /usr/bin/systemctl start man-db-cache-update. ░░ Subject: A start job for unit run-rc33e3c21cb574e25bfac7dbe6e404b91.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit run-rc33e3c21cb574e25bfac7dbe6e404b91.service has finished successfully. ░░ ░░ The job identifier is 938. Feb 06 16:13:00 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: Starting man-db-cache-update.service... ░░ Subject: A start job for unit man-db-cache-update.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit man-db-cache-update.service has begun execution. ░░ ░░ The job identifier is 1000. Feb 06 16:13:02 ip-10-31-11-33.us-east-1.aws.redhat.com python3[5250]: ansible-ansible.legacy.dnf Invoked with name=['make'] state=latest allow_downgrade=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 allowerasing=False nobest=False conf_file=None disable_excludes=None download_dir=None list=None releasever=None Feb 06 16:13:03 ip-10-31-11-33.us-east-1.aws.redhat.com python3[5269]: ansible-ansible.legacy.dnf Invoked with name=['parted'] state=latest allow_downgrade=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 allowerasing=False nobest=False conf_file=None disable_excludes=None download_dir=None list=None releasever=None Feb 06 16:13:04 ip-10-31-11-33.us-east-1.aws.redhat.com python3[5357]: ansible-ansible.legacy.dnf Invoked with name=['psmisc'] state=latest allow_downgrade=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 allowerasing=False nobest=False conf_file=None disable_excludes=None download_dir=None list=None releasever=None Feb 06 16:13:04 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: man-db-cache-update.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit man-db-cache-update.service has successfully entered the 'dead' state. Feb 06 16:13:04 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: Finished man-db-cache-update.service. ░░ Subject: A start job for unit man-db-cache-update.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit man-db-cache-update.service has finished successfully. ░░ ░░ The job identifier is 1000. Feb 06 16:13:04 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: run-rc33e3c21cb574e25bfac7dbe6e404b91.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit run-rc33e3c21cb574e25bfac7dbe6e404b91.service has successfully entered the 'dead' state. Feb 06 16:13:05 ip-10-31-11-33.us-east-1.aws.redhat.com python3[5376]: ansible-ansible.legacy.dnf Invoked with name=['chrony'] state=latest allow_downgrade=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 allowerasing=False nobest=False conf_file=None disable_excludes=None download_dir=None list=None releasever=None Feb 06 16:13:07 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: Started /usr/bin/systemctl start man-db-cache-update. ░░ Subject: A start job for unit run-r85fb7a904aa9446a9e4cb189ad4d069b.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit run-r85fb7a904aa9446a9e4cb189ad4d069b.service has finished successfully. ░░ ░░ The job identifier is 1062. Feb 06 16:13:07 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: Starting man-db-cache-update.service... ░░ Subject: A start job for unit man-db-cache-update.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit man-db-cache-update.service has begun execution. ░░ ░░ The job identifier is 1124. Feb 06 16:13:07 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: Reloading. Feb 06 16:13:07 ip-10-31-11-33.us-east-1.aws.redhat.com systemd-rc-local-generator[5414]: /etc/rc.d/rc.local is not marked executable, skipping. Feb 06 16:13:07 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: Queuing reload/restart jobs for marked units… Feb 06 16:13:07 ip-10-31-11-33.us-east-1.aws.redhat.com chronyd[603]: chronyd exiting Feb 06 16:13:07 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: Stopping NTP client/server... ░░ Subject: A stop job for unit chronyd.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A stop job for unit chronyd.service has begun execution. ░░ ░░ The job identifier is 1186. Feb 06 16:13:07 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: chronyd.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit chronyd.service has successfully entered the 'dead' state. Feb 06 16:13:07 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: Stopped NTP client/server. ░░ Subject: A stop job for unit chronyd.service has finished ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A stop job for unit chronyd.service has finished. ░░ ░░ The job identifier is 1186 and the job result is done. Feb 06 16:13:07 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: Starting NTP client/server... ░░ Subject: A start job for unit chronyd.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit chronyd.service has begun execution. ░░ ░░ The job identifier is 1186. Feb 06 16:13:07 ip-10-31-11-33.us-east-1.aws.redhat.com chronyd[5444]: chronyd version 4.5 starting (+CMDMON +NTP +REFCLOCK +RTC +PRIVDROP +SCFILTER +SIGND +ASYNCDNS +NTS +SECHASH +IPV6 +DEBUG) Feb 06 16:13:07 ip-10-31-11-33.us-east-1.aws.redhat.com chronyd[5444]: Loaded 0 symmetric keys Feb 06 16:13:07 ip-10-31-11-33.us-east-1.aws.redhat.com chronyd[5444]: Frequency 39.257 +/- 1.061 ppm read from /var/lib/chrony/drift Feb 06 16:13:07 ip-10-31-11-33.us-east-1.aws.redhat.com chronyd[5444]: Using right/UTC timezone to obtain leap second data Feb 06 16:13:07 ip-10-31-11-33.us-east-1.aws.redhat.com chronyd[5444]: Loaded seccomp filter (level 2) Feb 06 16:13:07 ip-10-31-11-33.us-east-1.aws.redhat.com chronyd[5444]: Added source 10.11.160.238 Feb 06 16:13:07 ip-10-31-11-33.us-east-1.aws.redhat.com chronyd[5444]: Added source 10.18.100.10 Feb 06 16:13:07 ip-10-31-11-33.us-east-1.aws.redhat.com chronyd[5444]: Added source 10.2.32.37 Feb 06 16:13:07 ip-10-31-11-33.us-east-1.aws.redhat.com chronyd[5444]: Added source 10.2.32.38 Feb 06 16:13:07 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: Started NTP client/server. ░░ Subject: A start job for unit chronyd.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit chronyd.service has finished successfully. ░░ ░░ The job identifier is 1186. Feb 06 16:13:07 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: Started /usr/bin/systemctl start man-db-cache-update. ░░ Subject: A start job for unit run-ra358c39a900f4bd2a5280e2f4839bc6a.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit run-ra358c39a900f4bd2a5280e2f4839bc6a.service has finished successfully. ░░ ░░ The job identifier is 1250. Feb 06 16:13:08 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: man-db-cache-update.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit man-db-cache-update.service has successfully entered the 'dead' state. Feb 06 16:13:08 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: Finished man-db-cache-update.service. ░░ Subject: A start job for unit man-db-cache-update.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit man-db-cache-update.service has finished successfully. ░░ ░░ The job identifier is 1124. Feb 06 16:13:08 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: run-r85fb7a904aa9446a9e4cb189ad4d069b.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit run-r85fb7a904aa9446a9e4cb189ad4d069b.service has successfully entered the 'dead' state. Feb 06 16:13:08 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: run-ra358c39a900f4bd2a5280e2f4839bc6a.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit run-ra358c39a900f4bd2a5280e2f4839bc6a.service has successfully entered the 'dead' state. Feb 06 16:13:08 ip-10-31-11-33.us-east-1.aws.redhat.com python3[5632]: ansible-ansible.legacy.command Invoked with _raw_params=bash -c "chronyc sources" _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 06 16:13:09 ip-10-31-11-33.us-east-1.aws.redhat.com python3[5650]: ansible-ansible.legacy.command Invoked with _raw_params=bash -c "chronyc sources | grep redhat | grep -v util.phx2.redhat || chronyc add server clock.corp.redhat.com iburst" _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 06 16:13:09 ip-10-31-11-33.us-east-1.aws.redhat.com python3[5671]: ansible-ansible.legacy.command Invoked with _raw_params=bash -c "chronyc sources" _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 06 16:13:10 ip-10-31-11-33.us-east-1.aws.redhat.com python3[5689]: ansible-ansible.legacy.command Invoked with _raw_params=chronyc waitsync 5 _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 06 16:13:13 ip-10-31-11-33.us-east-1.aws.redhat.com chronyd[5444]: Selected source 10.2.32.38 Feb 06 16:13:13 ip-10-31-11-33.us-east-1.aws.redhat.com chronyd[5444]: System clock TAI offset set to 37 seconds Feb 06 16:13:20 ip-10-31-11-33.us-east-1.aws.redhat.com python3[5707]: ansible-ansible.legacy.command Invoked with _raw_params=setenforce 0; timedatectl set-timezone UTC; setenforce 1 _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 06 16:13:20 ip-10-31-11-33.us-east-1.aws.redhat.com dbus-broker-launch[577]: avc: op=setenforce lsm=selinux enforcing=0 res=1 Feb 06 16:13:20 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: Starting Time & Date Service... ░░ Subject: A start job for unit systemd-timedated.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-timedated.service has begun execution. ░░ ░░ The job identifier is 1374. Feb 06 16:13:20 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: Started Time & Date Service. ░░ Subject: A start job for unit systemd-timedated.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-timedated.service has finished successfully. ░░ ░░ The job identifier is 1374. Feb 06 16:13:20 ip-10-31-11-33.us-east-1.aws.redhat.com systemd-timedated[5711]: Changed time zone to 'UTC' (UTC). ░░ Subject: Time zone change to UTC ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The system timezone has been changed to UTC. Feb 06 16:13:21 ip-10-31-11-33.us-east-1.aws.redhat.com python3[5728]: ansible-yum_repository Invoked with state=present name=beaker-tasks description=Beaker tasks baseurl=['http://beaker.engineering.redhat.com/rpms/'] gpgcheck=False enabled=True reposdir=/etc/yum.repos.d unsafe_writes=False bandwidth=None cost=None deltarpm_metadata_percentage=None deltarpm_percentage=None enablegroups=None exclude=None failovermethod=None file=None gpgcakey=None gpgkey=None module_hotfixes=None http_caching=None include=None includepkgs=None ip_resolve=None keepalive=None keepcache=None metadata_expire=None metadata_expire_filter=None metalink=None mirrorlist=None mirrorlist_expire=None params=None password=NOT_LOGGING_PARAMETER priority=None protect=None proxy=None proxy_password=NOT_LOGGING_PARAMETER proxy_username=None repo_gpgcheck=None retries=None s3_enabled=None skip_if_unavailable=None sslcacert=None ssl_check_cert_permissions=None sslclientcert=None sslclientkey=None sslverify=None throttle=None timeout=None ui_repoid_vars=None username=None async=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Feb 06 16:13:21 ip-10-31-11-33.us-east-1.aws.redhat.com python3[5745]: ansible-user Invoked with name=root update_password=always password=NOT_LOGGING_PARAMETER state=present non_unique=False force=False remove=False create_home=True system=False move_home=False append=False ssh_key_bits=0 ssh_key_type=rsa ssh_key_comment=ansible-generated on ip-10-31-11-33.us-east-1.aws.redhat.com uid=None group=None groups=None comment=None home=None shell=None login_class=None password_expire_max=None password_expire_min=None hidden=None seuser=None skeleton=None generate_ssh_key=None ssh_key_file=None ssh_key_passphrase=NOT_LOGGING_PARAMETER expires=None password_lock=None local=None profile=None authorization=None role=None umask=None Feb 06 16:13:21 ip-10-31-11-33.us-east-1.aws.redhat.com usermod[5748]: change user 'root' password Feb 06 16:13:22 ip-10-31-11-33.us-east-1.aws.redhat.com python3[5771]: ansible-lineinfile Invoked with dest=/etc/ssh/sshd_config regexp=#?PasswordAuthentication (?:yes|no) line=PasswordAuthentication yes state=present path=/etc/ssh/sshd_config backrefs=False create=False backup=False firstmatch=False unsafe_writes=False search_string=None insertafter=None insertbefore=None validate=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Feb 06 16:13:22 ip-10-31-11-33.us-east-1.aws.redhat.com python3[5788]: ansible-lineinfile Invoked with dest=/etc/ssh/sshd_config line=PermitRootLogin yes state=present path=/etc/ssh/sshd_config backrefs=False create=False backup=False firstmatch=False unsafe_writes=False regexp=None search_string=None insertafter=None insertbefore=None validate=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Feb 06 16:13:22 ip-10-31-11-33.us-east-1.aws.redhat.com python3[5805]: ansible-file Invoked with path=/var/lib/tft state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Feb 06 16:13:22 ip-10-31-11-33.us-east-1.aws.redhat.com python3[5822]: ansible-stat Invoked with path=/var/lib/tft/lib.sh follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Feb 06 16:13:23 ip-10-31-11-33.us-east-1.aws.redhat.com python3[5877]: ansible-ansible.legacy.stat Invoked with path=/var/lib/tft/lib.sh follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Feb 06 16:13:23 ip-10-31-11-33.us-east-1.aws.redhat.com python3[5927]: ansible-ansible.legacy.copy Invoked with src=/root/.ansible/tmp/ansible-tmp-1707236002.8788974-6324-92264849035175/source dest=/var/lib/tft/lib.sh owner=root group=root mode=u=rx,go= follow=False _original_basename=lib.sh.j2 checksum=cdcb4e3b7fb4fbc298e05439468daaeb6289f07c backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Feb 06 16:13:23 ip-10-31-11-33.us-east-1.aws.redhat.com python3[5961]: ansible-stat Invoked with path=/usr/local/bin/ci-extendtesttime.sh follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Feb 06 16:13:23 ip-10-31-11-33.us-east-1.aws.redhat.com python3[6016]: ansible-ansible.legacy.stat Invoked with path=/usr/local/bin/ci-extendtesttime.sh follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Feb 06 16:13:24 ip-10-31-11-33.us-east-1.aws.redhat.com python3[6066]: ansible-ansible.legacy.copy Invoked with src=/root/.ansible/tmp/ansible-tmp-1707236003.7361472-6337-120407010047358/source dest=/usr/local/bin/ci-extendtesttime.sh owner=root group=root mode=u=rx,go= follow=False _original_basename=ci-extendtesttime.sh.j2 checksum=2d0f6c296873c17e8b22c9490b000973b2a8a350 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Feb 06 16:13:24 ip-10-31-11-33.us-east-1.aws.redhat.com python3[6100]: ansible-stat Invoked with path=/usr/bin/extendtesttime.sh follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Feb 06 16:13:24 ip-10-31-11-33.us-east-1.aws.redhat.com python3[6117]: ansible-stat Invoked with path=/usr/local/bin/ci-return2pool.sh follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Feb 06 16:13:24 ip-10-31-11-33.us-east-1.aws.redhat.com python3[6172]: ansible-ansible.legacy.stat Invoked with path=/usr/local/bin/ci-return2pool.sh follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Feb 06 16:13:25 ip-10-31-11-33.us-east-1.aws.redhat.com python3[6222]: ansible-ansible.legacy.copy Invoked with src=/root/.ansible/tmp/ansible-tmp-1707236004.7075684-6357-208701873074156/source dest=/usr/local/bin/ci-return2pool.sh owner=root group=root mode=u=rx,go= follow=False _original_basename=ci-return2pool.sh.j2 checksum=f77cae3b2a729c60bcb0c01c82bf347f13a9b2c5 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Feb 06 16:13:25 ip-10-31-11-33.us-east-1.aws.redhat.com python3[6256]: ansible-stat Invoked with path=/etc/motd follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Feb 06 16:13:25 ip-10-31-11-33.us-east-1.aws.redhat.com python3[6313]: ansible-ansible.legacy.stat Invoked with path=/root/.guest-metadata.json follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Feb 06 16:13:25 ip-10-31-11-33.us-east-1.aws.redhat.com python3[6363]: ansible-ansible.legacy.copy Invoked with dest=/root/.guest-metadata.json owner=root group=root mode=u=rw,go= src=/root/.ansible/tmp/ansible-tmp-1707236005.4636712-6372-173050988815162/source _original_basename=tmp4la3j_j7 follow=False checksum=bf21a9e8fbc5a3846fb05b4fa0859e0917b2202f backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Feb 06 16:13:26 ip-10-31-11-33.us-east-1.aws.redhat.com python3[6397]: ansible-ansible.legacy.command Invoked with _raw_params=curl -sLI --connect-timeout 5 -w '%{response_code}' http://169.254.169.254/latest/meta-data/instance-id | grep ^200 _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 06 16:13:26 ip-10-31-11-33.us-east-1.aws.redhat.com python3[6417]: ansible-hostname Invoked with name=f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm use=None Feb 06 16:13:26 ip-10-31-11-33.us-east-1.aws.redhat.com dbus-broker-launch[577]: avc: op=setenforce lsm=selinux enforcing=1 res=1 Feb 06 16:13:26 ip-10-31-11-33.us-east-1.aws.redhat.com systemd-journald[483]: Field hash table of /run/log/journal/acfd29633386459288749fe4a6dbe762/system.journal has a fill level at 75.1 (250 of 333 items), suggesting rotation. Feb 06 16:13:26 ip-10-31-11-33.us-east-1.aws.redhat.com systemd-journald[483]: /run/log/journal/acfd29633386459288749fe4a6dbe762/system.journal: Journal header limits reached or header out-of-date, rotating. Feb 06 16:13:26 ip-10-31-11-33.us-east-1.aws.redhat.com rsyslogd[797]: imjournal: journal files changed, reloading... [v8.2310.0-3.el9 try https://www.rsyslog.com/e/0 ] Feb 06 16:13:26 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: Starting Hostname Service... ░░ Subject: A start job for unit systemd-hostnamed.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-hostnamed.service has begun execution. ░░ ░░ The job identifier is 1437. Feb 06 16:13:26 ip-10-31-11-33.us-east-1.aws.redhat.com systemd[1]: Started Hostname Service. ░░ Subject: A start job for unit systemd-hostnamed.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-hostnamed.service has finished successfully. ░░ ░░ The job identifier is 1437. Feb 06 16:13:26 ip-10-31-11-33.us-east-1.aws.redhat.com systemd-hostnamed[6422]: Changed pretty hostname to 'f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm' Feb 06 16:13:26 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm systemd-hostnamed[6422]: Hostname set to (static) Feb 06 16:13:26 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm NetworkManager[629]: [1707236006.8036] hostname: static hostname changed from "ip-10-31-11-33.us-east-1.aws.redhat.com" to "f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm" Feb 06 16:13:26 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm systemd[1]: Starting Network Manager Script Dispatcher Service... ░░ Subject: A start job for unit NetworkManager-dispatcher.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit NetworkManager-dispatcher.service has begun execution. ░░ ░░ The job identifier is 1500. Feb 06 16:13:26 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm systemd[1]: Started Network Manager Script Dispatcher Service. ░░ Subject: A start job for unit NetworkManager-dispatcher.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit NetworkManager-dispatcher.service has finished successfully. ░░ ░░ The job identifier is 1500. Feb 06 16:13:27 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[6451]: ansible-ansible.legacy.command Invoked with _raw_params=ping -4 -c1 -W10 $(hostname) _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 06 16:13:27 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[6470]: ansible-file Invoked with path=/var/log/messages state=touch recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Feb 06 16:13:28 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[6487]: ansible-ansible.legacy.command Invoked with _raw_params=setsebool nis_enabled on _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 06 16:13:28 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm setsebool[6488]: The nis_enabled policy boolean was changed to on by root Feb 06 16:13:28 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[6505]: ansible-stat Invoked with path=/usr/bin/rstrnt-package follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Feb 06 16:13:28 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[6524]: ansible-ansible.legacy.command Invoked with _raw_params=sed -e 's/rpm -q $package/rpm -q --whatprovides $package/' -i /usr/bin/rstrnt-package _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 06 16:13:28 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[6542]: ansible-stat Invoked with path=/var/lib/restraint follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Feb 06 16:13:28 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[6561]: ansible-file Invoked with path=/var/lib/restraint/avc_since state=touch recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Feb 06 16:13:29 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[6578]: ansible-stat Invoked with path=/usr/share/beakerlib/beakerlib.sh follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Feb 06 16:13:29 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[6597]: ansible-file Invoked with dest=/usr/lib/beakerlib state=directory path=/usr/lib/beakerlib recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Feb 06 16:13:29 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[6614]: ansible-file Invoked with dest=/usr/share/rhts-library state=directory path=/usr/share/rhts-library recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Feb 06 16:13:29 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[6631]: ansible-file Invoked with src=/usr/share/beakerlib/beakerlib.sh dest=/usr/lib/beakerlib/beakerlib.sh state=link path=/usr/lib/beakerlib/beakerlib.sh recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Feb 06 16:13:30 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[6648]: ansible-file Invoked with src=/usr/share/beakerlib/beakerlib.sh dest=/usr/share/rhts-library/rhtslib.sh state=link path=/usr/share/rhts-library/rhtslib.sh recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Feb 06 16:13:30 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[6665]: ansible-ansible.legacy.command Invoked with _raw_params=mv /var/log/audit/audit.log /var/log/audit/audit.log.bak _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 06 16:13:30 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[6683]: ansible-ansible.legacy.command Invoked with _raw_params=/usr/sbin/service auditd restart _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 06 16:13:30 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm auditd[548]: The audit daemon is exiting. Feb 06 16:13:30 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm kernel: audit: type=1305 audit(1707236010.592:719): op=set audit_pid=0 old=548 auid=4294967295 ses=4294967295 subj=system_u:system_r:auditd_t:s0 res=1 Feb 06 16:13:30 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm systemd[1]: auditd.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit auditd.service has successfully entered the 'dead' state. Feb 06 16:13:30 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm kernel: audit: type=1131 audit(1707236010.594:720): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:init_t:s0 msg='unit=auditd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 06 16:13:30 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm dbus-broker-launch[577]: avc: op=load_policy lsm=selinux seqno=2 res=1 Feb 06 16:13:32 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm systemd[1]: Starting Security Auditing Service... ░░ Subject: A start job for unit auditd.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit auditd.service has begun execution. ░░ ░░ The job identifier is 1563. Feb 06 16:13:32 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm auditd[6699]: No plugins found, not dispatching events Feb 06 16:13:32 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm auditd[6699]: Init complete, auditd 3.1.2 listening for events (startup state enable) Feb 06 16:13:32 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm augenrules[6702]: /sbin/augenrules: No change Feb 06 16:13:32 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm augenrules[6712]: No rules Feb 06 16:13:32 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm augenrules[6712]: enabled 1 Feb 06 16:13:32 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm augenrules[6712]: failure 1 Feb 06 16:13:32 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm augenrules[6712]: pid 6699 Feb 06 16:13:32 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm augenrules[6712]: rate_limit 0 Feb 06 16:13:32 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm augenrules[6712]: backlog_limit 8192 Feb 06 16:13:32 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm augenrules[6712]: lost 0 Feb 06 16:13:32 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm augenrules[6712]: backlog 0 Feb 06 16:13:32 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm augenrules[6712]: backlog_wait_time 60000 Feb 06 16:13:32 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm augenrules[6712]: backlog_wait_time_actual 0 Feb 06 16:13:32 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm augenrules[6712]: enabled 1 Feb 06 16:13:32 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm augenrules[6712]: failure 1 Feb 06 16:13:32 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm augenrules[6712]: pid 6699 Feb 06 16:13:32 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm augenrules[6712]: rate_limit 0 Feb 06 16:13:32 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm augenrules[6712]: backlog_limit 8192 Feb 06 16:13:32 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm augenrules[6712]: lost 0 Feb 06 16:13:32 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm augenrules[6712]: backlog 4 Feb 06 16:13:32 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm augenrules[6712]: backlog_wait_time 60000 Feb 06 16:13:32 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm augenrules[6712]: backlog_wait_time_actual 0 Feb 06 16:13:32 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm augenrules[6712]: enabled 1 Feb 06 16:13:32 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm augenrules[6712]: failure 1 Feb 06 16:13:32 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm augenrules[6712]: pid 6699 Feb 06 16:13:32 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm augenrules[6712]: rate_limit 0 Feb 06 16:13:32 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm augenrules[6712]: backlog_limit 8192 Feb 06 16:13:32 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm augenrules[6712]: lost 0 Feb 06 16:13:32 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm augenrules[6712]: backlog 3 Feb 06 16:13:32 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm augenrules[6712]: backlog_wait_time 60000 Feb 06 16:13:32 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm augenrules[6712]: backlog_wait_time_actual 0 Feb 06 16:13:32 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm systemd[1]: Started Security Auditing Service. ░░ Subject: A start job for unit auditd.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit auditd.service has finished successfully. ░░ ░░ The job identifier is 1563. Feb 06 16:13:32 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[6735]: ansible-stat Invoked with path=/etc/NetworkManager/conf.d follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Feb 06 16:13:33 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[6792]: ansible-ansible.legacy.stat Invoked with path=/etc/NetworkManager/conf.d/99-zzz-tft-workaround-dns-default.conf follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Feb 06 16:13:33 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[6842]: ansible-ansible.legacy.copy Invoked with dest=/etc/NetworkManager/conf.d/99-zzz-tft-workaround-dns-default.conf src=/root/.ansible/tmp/ansible-tmp-1707236012.9125493-6510-130119154099262/source _original_basename=tmpchlcqlvb follow=False checksum=0145f3ae57eef5aa08bbb678fedbb3edd001cd2d backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Feb 06 16:13:34 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[6876]: ansible-ansible.legacy.command Invoked with _raw_params=for repofile in $(ls -1 /etc/yum.repos.d/); do echo "# vvvvv $repofile vvvvv vvvvv"; cat "/etc/yum.repos.d/$repofile"; echo "# ^^^^^ ^^^^^ ^^^^^ ^^^^^"; echo ""; done _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 06 16:13:34 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[6907]: ansible-ansible.legacy.command Invoked with _raw_params=yum -q repolist _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 06 16:13:34 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[6925]: ansible-ansible.legacy.command Invoked with _raw_params=dnf config-manager --set-enabled highavailability _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 06 16:13:35 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[6943]: ansible-ansible.legacy.command Invoked with _raw_params=dnf clean all --enablerepo=* _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 06 16:13:36 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[6961]: ansible-ansible.legacy.command Invoked with _raw_params=dnf makecache --enablerepo=* _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 06 16:13:36 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm systemd[1]: NetworkManager-dispatcher.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit NetworkManager-dispatcher.service has successfully entered the 'dead' state. Feb 06 16:13:50 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm systemd[1]: systemd-timedated.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit systemd-timedated.service has successfully entered the 'dead' state. Feb 06 16:13:56 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm systemd[1]: systemd-hostnamed.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit systemd-hostnamed.service has successfully entered the 'dead' state. Feb 06 16:14:41 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[7180]: ansible-ansible.legacy.setup Invoked with gather_subset=['all'] gather_timeout=10 filter=[] fact_path=/etc/ansible/facts.d Feb 06 16:14:42 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[7221]: ansible-service_facts Invoked Feb 06 16:14:43 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[7320]: ansible-ansible.legacy.command Invoked with _raw_params=rpm -qa _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 06 16:14:44 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[7338]: ansible-ansible.legacy.command Invoked with _raw_params=if [ $(find /etc/yum.repos.d/ -name 'epel.repo' | wc -l ) -gt 0 ]; then dnf config-manager --set-disabled epel; fi _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 06 16:14:46 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[7360]: ansible-ansible.legacy.command Invoked with _raw_params=dnf clean all --enablerepo=* _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 06 16:14:46 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[7378]: ansible-ansible.legacy.command Invoked with _raw_params=dnf makecache --enablerepo=* _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 06 16:15:50 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[7593]: ansible-ansible.legacy.setup Invoked with gather_subset=['all'] gather_timeout=10 filter=[] fact_path=/etc/ansible/facts.d Feb 06 16:15:51 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[7634]: ansible-service_facts Invoked Feb 06 16:15:52 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[7733]: ansible-ansible.legacy.command Invoked with _raw_params=rpm -qa _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 06 16:15:55 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[7769]: ansible-ansible.legacy.setup Invoked with gather_subset=['all'] gather_timeout=10 filter=[] fact_path=/etc/ansible/facts.d Feb 06 16:15:56 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[7810]: ansible-ansible.legacy.dnf Invoked with name=['ca-certificates'] state=latest allow_downgrade=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 allowerasing=False nobest=False conf_file=None disable_excludes=None download_dir=None list=None releasever=None Feb 06 16:15:58 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[7828]: ansible-ansible.legacy.dnf Invoked with name=['curl'] state=latest allow_downgrade=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 allowerasing=False nobest=False conf_file=None disable_excludes=None download_dir=None list=None releasever=None Feb 06 16:15:59 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[7846]: ansible-ansible.legacy.command Invoked with _raw_params=curl -skL -o /etc/pki/ca-trust/source/anchors/Current-IT-Root-CAs.pem https://certs.corp.redhat.com/certs/Current-IT-Root-CAs.pem _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 06 16:15:59 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[7865]: ansible-ansible.legacy.command Invoked with _raw_params=bash -c 'update-ca-trust' _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 06 16:16:00 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[7889]: ansible-ansible.legacy.command Invoked with _raw_params=bash -c 'update-ca-trust extract' _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 06 16:16:01 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[7913]: ansible-ansible.legacy.command Invoked with _raw_params=curl -v https://gitlab.cee.redhat.com _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 06 16:16:02 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[7932]: ansible-ansible.legacy.command Invoked with _raw_params=curl -v https://beaker.engineering.redhat.com/ _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 06 16:16:03 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[7951]: ansible-service_facts Invoked Feb 06 16:16:04 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[8050]: ansible-ansible.legacy.command Invoked with _raw_params=rpm -qa _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 06 16:16:05 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[8068]: ansible-service_facts Invoked Feb 06 16:16:07 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[8167]: ansible-ansible.legacy.command Invoked with _raw_params=rpm -qa _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 06 16:16:08 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[8185]: ansible-ansible.legacy.command Invoked with _raw_params=for repofile in $(ls -1 /etc/yum.repos.d/); do echo "# vvvvv $repofile vvvvv vvvvv"; cat "/etc/yum.repos.d/$repofile"; echo "# ^^^^^ ^^^^^ ^^^^^ ^^^^^"; echo ""; done _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 06 16:16:08 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[8216]: ansible-ansible.legacy.command Invoked with _raw_params=yum -q repolist _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 06 16:16:09 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[8234]: ansible-ansible.legacy.command Invoked with _raw_params=dnf clean all --enablerepo=* _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 06 16:16:09 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[8252]: ansible-ansible.legacy.command Invoked with _raw_params=dnf makecache --enablerepo=* _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 06 16:16:10 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm systemd[4047]: Created slice User Background Tasks Slice. ░░ Subject: A start job for unit UNIT has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has finished successfully. ░░ ░░ The job identifier is 14. Feb 06 16:16:10 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm systemd[4047]: Starting Cleanup of User's Temporary Files and Directories... ░░ Subject: A start job for unit UNIT has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has begun execution. ░░ ░░ The job identifier is 13. Feb 06 16:16:10 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm systemd[4047]: Finished Cleanup of User's Temporary Files and Directories. ░░ Subject: A start job for unit UNIT has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has finished successfully. ░░ ░░ The job identifier is 13. Feb 06 16:17:14 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[8468]: ansible-ansible.legacy.setup Invoked with gather_subset=['all'] gather_timeout=10 filter=[] fact_path=/etc/ansible/facts.d Feb 06 16:17:14 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[8511]: ansible-stat Invoked with path=/run/ostree-booted follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Feb 06 16:17:15 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[8528]: ansible-ansible.legacy.dnf Invoked with name=['python3-pyasn1', 'python3-cryptography', 'python3-dbus'] state=present allow_downgrade=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 allowerasing=False nobest=False use_backend=auto conf_file=None disable_excludes=None download_dir=None list=None releasever=None Feb 06 16:17:21 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[8558]: ansible-ansible.legacy.dnf Invoked with name=['certmonger', 'python3-packaging'] state=present allow_downgrade=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 allowerasing=False nobest=False use_backend=auto conf_file=None disable_excludes=None download_dir=None list=None releasever=None Feb 06 16:17:22 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm dbus-broker-launch[576]: Noticed file-system modification, trigger reload. ░░ Subject: A configuration directory was written to ░░ Defined-By: dbus-broker ░░ Support: https://groups.google.com/forum/#!forum/bus1-devel ░░ ░░ A write was detected to one of the directories containing D-Bus configuration ░░ files, triggering a configuration reload. ░░ ░░ This functionality exists for backwards compatibility to pick up changes to ░░ D-Bus configuration without an explicit reolad request. Typically when ░░ installing or removing third-party software causes D-Bus configuration files ░░ to be added or removed. ░░ ░░ It is worth noting that this may cause partial configuration to be loaded in ░░ case dispatching this notification races with the writing of the configuration ░░ files. However, a future notification will then cause the configuration to be ░░ reladed again. Feb 06 16:17:22 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm dbus-broker-launch[576]: Noticed file-system modification, trigger reload. ░░ Subject: A configuration directory was written to ░░ Defined-By: dbus-broker ░░ Support: https://groups.google.com/forum/#!forum/bus1-devel ░░ ░░ A write was detected to one of the directories containing D-Bus configuration ░░ files, triggering a configuration reload. ░░ ░░ This functionality exists for backwards compatibility to pick up changes to ░░ D-Bus configuration without an explicit reolad request. Typically when ░░ installing or removing third-party software causes D-Bus configuration files ░░ to be added or removed. ░░ ░░ It is worth noting that this may cause partial configuration to be loaded in ░░ case dispatching this notification races with the writing of the configuration ░░ files. However, a future notification will then cause the configuration to be ░░ reladed again. Feb 06 16:17:22 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm dbus-broker-launch[576]: Noticed file-system modification, trigger reload. ░░ Subject: A configuration directory was written to ░░ Defined-By: dbus-broker ░░ Support: https://groups.google.com/forum/#!forum/bus1-devel ░░ ░░ A write was detected to one of the directories containing D-Bus configuration ░░ files, triggering a configuration reload. ░░ ░░ This functionality exists for backwards compatibility to pick up changes to ░░ D-Bus configuration without an explicit reolad request. Typically when ░░ installing or removing third-party software causes D-Bus configuration files ░░ to be added or removed. ░░ ░░ It is worth noting that this may cause partial configuration to be loaded in ░░ case dispatching this notification races with the writing of the configuration ░░ files. However, a future notification will then cause the configuration to be ░░ reladed again. Feb 06 16:17:23 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm systemd[1]: Reloading. Feb 06 16:17:23 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm systemd-rc-local-generator[8586]: /etc/rc.d/rc.local is not marked executable, skipping. Feb 06 16:17:23 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm systemd[1]: Started /usr/bin/systemctl start man-db-cache-update. ░░ Subject: A start job for unit run-rbfe36f10f7f44c63ab29443a74e7946b.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit run-rbfe36f10f7f44c63ab29443a74e7946b.service has finished successfully. ░░ ░░ The job identifier is 1571. Feb 06 16:17:23 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm systemd[1]: Starting man-db-cache-update.service... ░░ Subject: A start job for unit man-db-cache-update.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit man-db-cache-update.service has begun execution. ░░ ░░ The job identifier is 1633. Feb 06 16:17:23 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm systemd[1]: Reloading. Feb 06 16:17:23 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm systemd-rc-local-generator[8623]: /etc/rc.d/rc.local is not marked executable, skipping. Feb 06 16:17:23 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm systemd[1]: Queuing reload/restart jobs for marked units… Feb 06 16:17:24 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm systemd[1]: man-db-cache-update.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit man-db-cache-update.service has successfully entered the 'dead' state. Feb 06 16:17:24 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm systemd[1]: Finished man-db-cache-update.service. ░░ Subject: A start job for unit man-db-cache-update.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit man-db-cache-update.service has finished successfully. ░░ ░░ The job identifier is 1633. Feb 06 16:17:24 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm systemd[1]: run-rbfe36f10f7f44c63ab29443a74e7946b.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit run-rbfe36f10f7f44c63ab29443a74e7946b.service has successfully entered the 'dead' state. Feb 06 16:17:24 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[8918]: ansible-file Invoked with name=/etc/certmonger//pre-scripts owner=root group=root mode=0700 state=directory path=/etc/certmonger//pre-scripts recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Feb 06 16:17:24 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[8935]: ansible-file Invoked with name=/etc/certmonger//post-scripts owner=root group=root mode=0700 state=directory path=/etc/certmonger//post-scripts recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Feb 06 16:17:25 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[8952]: ansible-ansible.legacy.systemd Invoked with name=certmonger state=started enabled=True daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Feb 06 16:17:25 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm systemd[1]: Reloading. Feb 06 16:17:25 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm systemd-rc-local-generator[8971]: /etc/rc.d/rc.local is not marked executable, skipping. Feb 06 16:17:25 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm systemd[1]: Starting Certificate monitoring and PKI enrollment... ░░ Subject: A start job for unit certmonger.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit certmonger.service has begun execution. ░░ ░░ The job identifier is 1695. Feb 06 16:17:25 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm systemd[1]: Started Certificate monitoring and PKI enrollment. ░░ Subject: A start job for unit certmonger.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit certmonger.service has finished successfully. ░░ ░░ The job identifier is 1695. Feb 06 16:17:25 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[9025]: ansible-fedora.linux_system_roles.certificate_request Invoked with name=quadlet_demo dns=['localhost'] directory=/etc/pki/tls wait=True ca=self-sign __header=# # Ansible managed # # system_role:certificate provider_config_directory=/etc/certmonger provider=certmonger key_usage=['digitalSignature', 'keyEncipherment'] extended_key_usage=['id-kp-serverAuth', 'id-kp-clientAuth'] auto_renew=True ip=None email=None common_name=None country=None state=None locality=None organization=None organizational_unit=None contact_email=None key_size=None owner=None group=None mode=None principal=None run_before=None run_after=None Feb 06 16:17:25 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm certmonger[8984]: 2024-02-06 16:17:25 [8984] Wrote to /var/lib/certmonger/requests/20240206161725 Feb 06 16:17:25 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm certmonger[8984]: 2024-02-06 16:17:25 [8984] Wrote to /var/lib/certmonger/requests/20240206161725 Feb 06 16:17:25 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm certmonger[8984]: 2024-02-06 16:17:25 [8984] Wrote to /var/lib/certmonger/requests/20240206161725 Feb 06 16:17:25 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm certmonger[8984]: 2024-02-06 16:17:25 [8984] Wrote to /var/lib/certmonger/requests/20240206161725 Feb 06 16:17:25 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm certmonger[8984]: 2024-02-06 16:17:25 [8984] Wrote to /var/lib/certmonger/requests/20240206161725 Feb 06 16:17:25 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm certmonger[8984]: 2024-02-06 16:17:25 [8984] Wrote to /var/lib/certmonger/requests/20240206161725 Feb 06 16:17:25 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm certmonger[8984]: 2024-02-06 16:17:25 [8984] Wrote to /var/lib/certmonger/requests/20240206161725 Feb 06 16:17:25 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm certmonger[8984]: 2024-02-06 16:17:25 [8984] Wrote to /var/lib/certmonger/requests/20240206161725 Feb 06 16:17:25 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm certmonger[8984]: 2024-02-06 16:17:25 [8984] Wrote to /var/lib/certmonger/requests/20240206161725 Feb 06 16:17:26 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm certmonger[8984]: 2024-02-06 16:17:26 [8984] Wrote to /var/lib/certmonger/requests/20240206161725 Feb 06 16:17:26 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm certmonger[8984]: 2024-02-06 16:17:26 [8984] Wrote to /var/lib/certmonger/requests/20240206161725 Feb 06 16:17:26 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm certmonger[8984]: 2024-02-06 16:17:26 [8984] Wrote to /var/lib/certmonger/requests/20240206161725 Feb 06 16:17:26 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm certmonger[8984]: 2024-02-06 16:17:26 [8984] Wrote to /var/lib/certmonger/requests/20240206161725 Feb 06 16:17:26 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm certmonger[8984]: 2024-02-06 16:17:26 [8984] Wrote to /var/lib/certmonger/requests/20240206161725 Feb 06 16:17:26 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm certmonger[8984]: 2024-02-06 16:17:26 [8984] Wrote to /var/lib/certmonger/requests/20240206161725 Feb 06 16:17:26 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm certmonger[8984]: 2024-02-06 16:17:26 [8984] Wrote to /var/lib/certmonger/requests/20240206161725 Feb 06 16:17:26 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm certmonger[8984]: 2024-02-06 16:17:26 [8984] Wrote to /var/lib/certmonger/requests/20240206161725 Feb 06 16:17:26 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm certmonger[8984]: 2024-02-06 16:17:26 [8984] Wrote to /var/lib/certmonger/requests/20240206161725 Feb 06 16:17:26 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm certmonger[8984]: 2024-02-06 16:17:26 [8984] Wrote to /var/lib/certmonger/requests/20240206161725 Feb 06 16:17:26 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm certmonger[8984]: 2024-02-06 16:17:26 [8984] Wrote to /var/lib/certmonger/requests/20240206161725 Feb 06 16:17:26 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm certmonger[8984]: 2024-02-06 16:17:26 [8984] Wrote to /var/lib/certmonger/requests/20240206161725 Feb 06 16:17:26 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm certmonger[8984]: 2024-02-06 16:17:26 [8984] Wrote to /var/lib/certmonger/requests/20240206161725 Feb 06 16:17:26 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm certmonger[8984]: 2024-02-06 16:17:26 [8984] Wrote to /var/lib/certmonger/requests/20240206161725 Feb 06 16:17:26 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm certmonger[8984]: 2024-02-06 16:17:26 [8984] Wrote to /var/lib/certmonger/requests/20240206161725 Feb 06 16:17:26 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm certmonger[8984]: 2024-02-06 16:17:26 [8984] Wrote to /var/lib/certmonger/requests/20240206161725 Feb 06 16:17:26 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm certmonger[8984]: 2024-02-06 16:17:26 [8984] Wrote to /var/lib/certmonger/requests/20240206161725 Feb 06 16:17:26 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm certmonger[8984]: 2024-02-06 16:17:26 [8984] Wrote to /var/lib/certmonger/requests/20240206161725 Feb 06 16:17:26 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm certmonger[8984]: 2024-02-06 16:17:26 [8984] Wrote to /var/lib/certmonger/requests/20240206161725 Feb 06 16:17:26 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm certmonger[8984]: 2024-02-06 16:17:26 [8984] Wrote to /var/lib/certmonger/requests/20240206161725 Feb 06 16:17:26 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm certmonger[9043]: Certificate in file "/etc/pki/tls/certs/quadlet_demo.crt" issued by CA and saved. Feb 06 16:17:26 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm certmonger[8984]: 2024-02-06 16:17:26 [8984] Wrote to /var/lib/certmonger/requests/20240206161725 Feb 06 16:17:26 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[9060]: ansible-slurp Invoked with path=/etc/pki/tls/certs/quadlet_demo.crt src=/etc/pki/tls/certs/quadlet_demo.crt Feb 06 16:17:26 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[9077]: ansible-slurp Invoked with path=/etc/pki/tls/private/quadlet_demo.key src=/etc/pki/tls/private/quadlet_demo.key Feb 06 16:17:26 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[9094]: ansible-slurp Invoked with path=/etc/pki/tls/certs/quadlet_demo.crt src=/etc/pki/tls/certs/quadlet_demo.crt Feb 06 16:17:27 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[9111]: ansible-ansible.legacy.command Invoked with _raw_params=getcert stop-tracking -f /etc/pki/tls/certs/quadlet_demo.crt _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 06 16:17:27 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm certmonger[8984]: 2024-02-06 16:17:27 [8984] Wrote to /var/lib/certmonger/requests/20240206161725 Feb 06 16:17:27 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[9129]: ansible-file Invoked with path=/etc/pki/tls/certs/quadlet_demo.crt state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Feb 06 16:17:27 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[9146]: ansible-file Invoked with path=/etc/pki/tls/private/quadlet_demo.key state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Feb 06 16:17:27 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[9163]: ansible-file Invoked with path=/etc/pki/tls/certs/quadlet_demo.crt state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Feb 06 16:17:28 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[9180]: ansible-stat Invoked with path=/run/ostree-booted follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Feb 06 16:17:29 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[9214]: ansible-ansible.legacy.dnf Invoked with name=['podman', 'shadow-utils-subid'] state=present allow_downgrade=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 allowerasing=False nobest=False use_backend=auto conf_file=None disable_excludes=None download_dir=None list=None releasever=None Feb 06 16:17:38 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm setsebool[9244]: The virt_use_nfs policy boolean was changed to 1 by root Feb 06 16:17:38 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm setsebool[9244]: The virt_sandbox_use_all_caps policy boolean was changed to 1 by root Feb 06 16:17:49 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm kernel: SELinux: Converting 397 SID table entries... Feb 06 16:17:49 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm kernel: SELinux: policy capability network_peer_controls=1 Feb 06 16:17:49 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm kernel: SELinux: policy capability open_perms=1 Feb 06 16:17:49 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm kernel: SELinux: policy capability extended_socket_class=1 Feb 06 16:17:49 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm kernel: SELinux: policy capability always_check_network=0 Feb 06 16:17:49 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm kernel: SELinux: policy capability cgroup_seclabel=1 Feb 06 16:17:49 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm kernel: SELinux: policy capability nnp_nosuid_transition=1 Feb 06 16:17:49 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm kernel: SELinux: policy capability genfs_seclabel_symlinks=1 Feb 06 16:18:17 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm dbus-broker-launch[577]: avc: op=load_policy lsm=selinux seqno=4 res=1 Feb 06 16:18:17 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm systemd[1]: Started /usr/bin/systemctl start man-db-cache-update. ░░ Subject: A start job for unit run-r8c1ade51d1164e94b066cba5b25ce03c.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit run-r8c1ade51d1164e94b066cba5b25ce03c.service has finished successfully. ░░ ░░ The job identifier is 1758. Feb 06 16:18:17 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm systemd[1]: Starting man-db-cache-update.service... ░░ Subject: A start job for unit man-db-cache-update.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit man-db-cache-update.service has begun execution. ░░ ░░ The job identifier is 1820. Feb 06 16:18:17 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm systemd[1]: Reloading. Feb 06 16:18:18 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm systemd-rc-local-generator[9945]: /etc/rc.d/rc.local is not marked executable, skipping. Feb 06 16:18:18 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm systemd[1]: Queuing reload/restart jobs for marked units… Feb 06 16:18:19 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm systemd[1]: man-db-cache-update.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit man-db-cache-update.service has successfully entered the 'dead' state. Feb 06 16:18:19 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm systemd[1]: Finished man-db-cache-update.service. ░░ Subject: A start job for unit man-db-cache-update.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit man-db-cache-update.service has finished successfully. ░░ ░░ The job identifier is 1820. Feb 06 16:18:19 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm systemd[1]: run-r8c1ade51d1164e94b066cba5b25ce03c.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit run-r8c1ade51d1164e94b066cba5b25ce03c.service has successfully entered the 'dead' state. Feb 06 16:18:20 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[11098]: ansible-ansible.legacy.command Invoked with _raw_params=podman --version _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 06 16:18:20 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[11121]: ansible-getent Invoked with database=passwd key=root fail_key=False service=None split=None Feb 06 16:18:21 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[11139]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None Feb 06 16:18:21 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[11157]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Feb 06 16:18:22 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[11176]: ansible-stat Invoked with path=/run/ostree-booted follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Feb 06 16:18:22 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[11193]: ansible-ansible.legacy.dnf Invoked with name=['firewalld'] state=present allow_downgrade=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 allowerasing=False nobest=False use_backend=auto conf_file=None disable_excludes=None download_dir=None list=None releasever=None Feb 06 16:18:23 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[11211]: ansible-systemd Invoked with name=firewalld masked=False daemon_reload=False daemon_reexec=False scope=system no_block=False state=None enabled=None force=None Feb 06 16:18:23 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[11230]: ansible-ansible.legacy.systemd Invoked with name=firewalld state=started enabled=True daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Feb 06 16:18:23 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm systemd[1]: Reloading. Feb 06 16:18:23 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm systemd-rc-local-generator[11250]: /etc/rc.d/rc.local is not marked executable, skipping. Feb 06 16:18:24 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm systemd[1]: Starting firewalld - dynamic firewall daemon... ░░ Subject: A start job for unit firewalld.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit firewalld.service has begun execution. ░░ ░░ The job identifier is 1882. Feb 06 16:18:25 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm systemd[1]: Started firewalld - dynamic firewall daemon. ░░ Subject: A start job for unit firewalld.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit firewalld.service has finished successfully. ░░ ░░ The job identifier is 1882. Feb 06 16:18:25 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[11288]: ansible-fedora.linux_system_roles.firewall_lib Invoked with port=['8000/tcp'] permanent=True runtime=True state=enabled __report_changed=True service=[] source_port=[] forward_port=[] rich_rule=[] source=[] interface=[] interface_pci_id=[] icmp_block=[] timeout=0 ipset_entries=[] protocol=[] helper_module=[] destination=[] firewalld_conf=None masquerade=None icmp_block_inversion=None target=None zone=None set_default_zone=None ipset=None ipset_type=None description=None short=None Feb 06 16:18:25 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm kernel: Warning: Unmaintained driver is detected: ip_set Feb 06 16:18:25 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm kernel: Warning: Deprecated Driver is detected: ip_set will not be maintained in a future major release and may be disabled Feb 06 16:18:26 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[11346]: ansible-fedora.linux_system_roles.firewall_lib Invoked with port=['9000/tcp'] permanent=True runtime=True state=enabled __report_changed=True service=[] source_port=[] forward_port=[] rich_rule=[] source=[] interface=[] interface_pci_id=[] icmp_block=[] timeout=0 ipset_entries=[] protocol=[] helper_module=[] destination=[] firewalld_conf=None masquerade=None icmp_block_inversion=None target=None zone=None set_default_zone=None ipset=None ipset_type=None description=None short=None Feb 06 16:18:28 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[11363]: ansible-containers.podman.podman_secret Invoked with data=NOT_LOGGING_PARAMETER name=mysql-root-password-container skip_existing=True state=present executable=podman force=False debug=False driver=None driver_opts=None labels=None Feb 06 16:18:28 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[11386]: ansible-ansible.legacy.command Invoked with _raw_params=podman secret ls _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 06 16:18:28 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm systemd[1]: var-lib-containers-storage-overlay-compat2767603289-merged.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit var-lib-containers-storage-overlay-compat2767603289-merged.mount has successfully entered the 'dead' state. Feb 06 16:18:28 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm systemd[1]: var-lib-containers-storage-overlay-metacopy\x2dcheck2795922956-merged.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit var-lib-containers-storage-overlay-metacopy\x2dcheck2795922956-merged.mount has successfully entered the 'dead' state. Feb 06 16:18:28 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm podman[11387]: 2024-02-06 16:18:28.521691718 +0000 UTC m=+0.167424063 system refresh Feb 06 16:18:28 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[11413]: ansible-containers.podman.podman_secret Invoked with data=NOT_LOGGING_PARAMETER name=mysql-root-password-kube skip_existing=True state=present executable=podman force=False debug=False driver=None driver_opts=None labels=None Feb 06 16:18:29 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[11436]: ansible-ansible.legacy.command Invoked with _raw_params=podman secret ls _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 06 16:18:29 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Feb 06 16:18:29 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[11460]: ansible-containers.podman.podman_secret Invoked with data=NOT_LOGGING_PARAMETER name=envoy-certificates skip_existing=True state=present executable=podman force=False debug=False driver=None driver_opts=None labels=None Feb 06 16:18:29 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[11483]: ansible-ansible.legacy.command Invoked with _raw_params=podman secret ls _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 06 16:18:29 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Feb 06 16:18:30 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[11508]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None Feb 06 16:18:30 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[11526]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Feb 06 16:18:31 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[11545]: ansible-file Invoked with path=/etc/containers/systemd state=directory owner=root group=0 mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Feb 06 16:18:31 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[11600]: ansible-ansible.legacy.stat Invoked with path=/etc/containers/systemd/quadlet-demo.network follow=False get_checksum=True checksum_algorithm=sha1 get_mime=True get_attributes=True Feb 06 16:18:32 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[11650]: ansible-ansible.legacy.copy Invoked with src=/root/.ansible/tmp/ansible-tmp-1707236311.5468745-8080-141066871357618/source dest=/etc/containers/systemd/quadlet-demo.network owner=root group=0 mode=0644 _original_basename=quadlet-demo.network follow=False checksum=e57c08d49aff4bae8daab138d913aeddaa8682a0 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Feb 06 16:18:32 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[11684]: ansible-systemd Invoked with daemon_reload=True scope=system daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None Feb 06 16:18:32 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm systemd[1]: Reloading. Feb 06 16:18:32 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm systemd-rc-local-generator[11706]: /etc/rc.d/rc.local is not marked executable, skipping. Feb 06 16:18:33 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[11734]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None Feb 06 16:18:33 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[11752]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Feb 06 16:18:34 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[11771]: ansible-file Invoked with path=/etc/containers/systemd state=directory owner=root group=0 mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Feb 06 16:18:34 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[11826]: ansible-ansible.legacy.stat Invoked with path=/etc/containers/systemd/quadlet-demo-mysql.volume follow=False get_checksum=True checksum_algorithm=sha1 get_mime=True get_attributes=True Feb 06 16:18:34 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[11876]: ansible-ansible.legacy.copy Invoked with src=/root/.ansible/tmp/ansible-tmp-1707236314.121802-8128-219481856671228/source dest=/etc/containers/systemd/quadlet-demo-mysql.volume owner=root group=0 mode=0644 _original_basename=quadlet-demo-mysql.volume follow=False checksum=585f8cbdf0ec73000f9227dcffbef71e9552ea4a backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Feb 06 16:18:34 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[11910]: ansible-systemd Invoked with daemon_reload=True scope=system daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None Feb 06 16:18:34 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm systemd[1]: Reloading. Feb 06 16:18:35 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm systemd-rc-local-generator[11928]: /etc/rc.d/rc.local is not marked executable, skipping. Feb 06 16:18:35 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[11960]: ansible-systemd Invoked with name=quadlet-demo-mysql-volume.service scope=system state=started daemon_reload=False daemon_reexec=False no_block=False enabled=None force=None masked=None Feb 06 16:18:35 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm systemd[1]: Starting quadlet-demo-mysql-volume.service... ░░ Subject: A start job for unit quadlet-demo-mysql-volume.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit quadlet-demo-mysql-volume.service has begun execution. ░░ ░░ The job identifier is 1952. Feb 06 16:18:35 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm podman[11963]: 2024-02-06 16:18:35.593334176 +0000 UTC m=+0.035196676 volume create systemd-quadlet-demo-mysql Feb 06 16:18:35 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm quadlet-demo-mysql-volume[11963]: systemd-quadlet-demo-mysql Feb 06 16:18:35 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Feb 06 16:18:35 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm systemd[1]: Finished quadlet-demo-mysql-volume.service. ░░ Subject: A start job for unit quadlet-demo-mysql-volume.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit quadlet-demo-mysql-volume.service has finished successfully. ░░ ░░ The job identifier is 1952. Feb 06 16:18:36 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[11986]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None Feb 06 16:18:36 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[12004]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Feb 06 16:18:37 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[12023]: ansible-file Invoked with path=/tmp/quadlet_demo state=directory owner=root group=root mode=0777 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Feb 06 16:18:37 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[12040]: ansible-containers.podman.podman_image Invoked with name=quay.io/linux-system-roles/mysql:5.6 force=True tag=latest pull=True push=False state=present executable=podman build={'force_rm': False, 'format': 'oci', 'cache': True, 'rm': True, 'annotation': None, 'file': None, 'volume': None, 'extra_args': None, 'target': None} push_args={'compress': None, 'format': None, 'remove_signatures': None, 'sign_by': None, 'dest': None, 'transport': None} arch=None path=None validate_certs=None auth_file=None username=None password=NOT_LOGGING_PARAMETER ca_cert_dir=None Feb 06 16:18:37 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[12040]: ansible-containers.podman.podman_image PODMAN-IMAGE-DEBUG: /usr/bin/podman image ls quay.io/linux-system-roles/mysql:5.6 --format json Feb 06 16:18:37 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Feb 06 16:18:37 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[12040]: ansible-containers.podman.podman_image PODMAN-IMAGE-DEBUG: /usr/bin/podman image exists quay.io/linux-system-roles/mysql:5.6 Feb 06 16:18:37 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[12040]: ansible-containers.podman.podman_image PODMAN-IMAGE-DEBUG: /usr/bin/podman pull quay.io/linux-system-roles/mysql:5.6 -q Feb 06 16:18:39 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm kernel: overlayfs: idmapped layers are currently not supported Feb 06 16:18:39 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm systemd[1]: var-lib-containers-storage-overlay-compat785331686-lower\x2dmapped.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit var-lib-containers-storage-overlay-compat785331686-lower\x2dmapped.mount has successfully entered the 'dead' state. Feb 06 16:18:45 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm podman[12055]: 2024-02-06 16:18:45.463818074 +0000 UTC m=+8.009180375 image pull dd3b2a5dcb48ff61113592ed5ddd762581be4387c7bc552375a2159422aa6bf5 quay.io/linux-system-roles/mysql:5.6 Feb 06 16:18:45 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Feb 06 16:18:45 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[12040]: ansible-containers.podman.podman_image PODMAN-IMAGE-DEBUG: /usr/bin/podman inspect dd3b2a5dcb48ff61113592ed5ddd762581be4387c7bc552375a2159422aa6bf5 --format json Feb 06 16:18:45 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Feb 06 16:18:45 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[12155]: ansible-file Invoked with path=/etc/containers/systemd state=directory owner=root group=0 mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Feb 06 16:18:46 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[12210]: ansible-ansible.legacy.stat Invoked with path=/etc/containers/systemd/quadlet-demo-mysql.container follow=False get_checksum=True checksum_algorithm=sha1 get_mime=True get_attributes=True Feb 06 16:18:46 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[12260]: ansible-ansible.legacy.copy Invoked with dest=/etc/containers/systemd/quadlet-demo-mysql.container owner=root group=0 mode=0644 src=/root/.ansible/tmp/ansible-tmp-1707236325.822158-8180-172502824652855/source _original_basename=tmpa99kk2bs follow=False checksum=7bf9312d57274d4dabbd222b01717e2cc0cfa0ed backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Feb 06 16:18:46 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[12294]: ansible-systemd Invoked with daemon_reload=True scope=system daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None Feb 06 16:18:46 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm systemd[1]: Reloading. Feb 06 16:18:46 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm systemd-rc-local-generator[12312]: /etc/rc.d/rc.local is not marked executable, skipping. Feb 06 16:18:47 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[12344]: ansible-systemd Invoked with name=quadlet-demo-mysql.service scope=system state=started daemon_reload=False daemon_reexec=False no_block=False enabled=None force=None masked=None Feb 06 16:18:47 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm systemd[1]: Starting quadlet-demo-network.service... ░░ Subject: A start job for unit quadlet-demo-network.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit quadlet-demo-network.service has begun execution. ░░ ░░ The job identifier is 2015. Feb 06 16:18:47 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm quadlet-demo-network[12347]: systemd-quadlet-demo Feb 06 16:18:47 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Feb 06 16:18:47 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm systemd[1]: Finished quadlet-demo-network.service. ░░ Subject: A start job for unit quadlet-demo-network.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit quadlet-demo-network.service has finished successfully. ░░ ░░ The job identifier is 2015. Feb 06 16:18:47 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm systemd[1]: Starting quadlet-demo-mysql.service... ░░ Subject: A start job for unit quadlet-demo-mysql.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit quadlet-demo-mysql.service has begun execution. ░░ ░░ The job identifier is 2014. Feb 06 16:18:47 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm quadlet-demo-mysql[12354]: Error: running container create option: no secret with name or id "mysql-root-password-container": no such secret Feb 06 16:18:47 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm podman[12354]: 2024-02-06 16:18:47.308349117 +0000 UTC m=+0.032056296 image pull dd3b2a5dcb48ff61113592ed5ddd762581be4387c7bc552375a2159422aa6bf5 quay.io/linux-system-roles/mysql:5.6 Feb 06 16:18:47 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Feb 06 16:18:47 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm systemd[1]: quadlet-demo-mysql.service: Main process exited, code=exited, status=125/n/a ░░ Subject: Unit process exited ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ An ExecStart= process belonging to unit quadlet-demo-mysql.service has exited. ░░ ░░ The process' exit code is 'exited' and its exit status is 125. Feb 06 16:18:47 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm systemd[1]: quadlet-demo-mysql.service: Failed with result 'exit-code'. ░░ Subject: Unit failed ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit quadlet-demo-mysql.service has entered the 'failed' state with result 'exit-code'. Feb 06 16:18:47 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm systemd[1]: Failed to start quadlet-demo-mysql.service. ░░ Subject: A start job for unit quadlet-demo-mysql.service has failed ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit quadlet-demo-mysql.service has finished with a failure. ░░ ░░ The job identifier is 2014 and the job result is failed. Feb 06 16:18:47 f67b215a-c8a4-4b3e-8432-0214ae3c2f5d.testing-farm python3[12384]: ansible-ansible.legacy.command Invoked with _raw_params=journalctl -ex _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None TASK [Check] ******************************************************************* task path: /WORKDIR/git-fix-linger-handlingih6nvedg/tests/tests_quadlet_demo.yml:132 Tuesday 06 February 2024 16:18:47 +0000 (0:00:00.271) 0:01:33.973 ****** ok: [sut] => { "changed": false, "cmd": [ "podman", "ps", "-a" ], "delta": "0:00:00.038381", "end": "2024-02-06 16:18:47.878820", "rc": 0, "start": "2024-02-06 16:18:47.840439" } STDOUT: CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES TASK [Check pods] ************************************************************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/tests/tests_quadlet_demo.yml:136 Tuesday 06 February 2024 16:18:47 +0000 (0:00:00.284) 0:01:34.258 ****** ok: [sut] => { "changed": false, "cmd": [ "podman", "pod", "ps", "--ctr-ids", "--ctr-names", "--ctr-status" ], "delta": "0:00:00.034869", "end": "2024-02-06 16:18:48.161522", "failed_when_result": false, "rc": 0, "start": "2024-02-06 16:18:48.126653" } STDOUT: POD ID NAME STATUS CREATED INFRA ID IDS NAMES STATUS TASK [Check systemd] *********************************************************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/tests/tests_quadlet_demo.yml:141 Tuesday 06 February 2024 16:18:48 +0000 (0:00:00.234) 0:01:34.492 ****** ok: [sut] => { "changed": false, "cmd": "set -euo pipefail; systemctl list-units --all --with-dependencies | grep quadlet", "delta": "0:00:00.014155", "end": "2024-02-06 16:18:48.377605", "failed_when_result": false, "rc": 0, "start": "2024-02-06 16:18:48.363450" } STDOUT: quadlet-demo-mysql-volume.service loaded active exited quadlet-demo-mysql-volume.service ● quadlet-demo-mysql.service loaded failed failed quadlet-demo-mysql.service quadlet-demo-network.service loaded active exited quadlet-demo-network.service TASK [LS] ********************************************************************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/tests/tests_quadlet_demo.yml:149 Tuesday 06 February 2024 16:18:48 +0000 (0:00:00.215) 0:01:34.707 ****** ok: [sut] => { "changed": false, "cmd": [ "ls", "-alrtF", "/etc/systemd/system" ], "delta": "0:00:00.004302", "end": "2024-02-06 16:18:48.579080", "failed_when_result": false, "rc": 0, "start": "2024-02-06 16:18:48.574778" } STDOUT: total 8 drwxr-xr-x. 2 root root 32 Jan 30 09:28 getty.target.wants/ lrwxrwxrwx. 1 root root 37 Jan 30 09:28 ctrl-alt-del.target -> /usr/lib/systemd/system/reboot.target lrwxrwxrwx. 1 root root 43 Jan 30 09:28 dbus.service -> /usr/lib/systemd/system/dbus-broker.service drwxr-xr-x. 4 root root 166 Jan 30 09:29 ../ drwxr-xr-x. 2 root root 152 Jan 30 09:29 sysinit.target.wants/ lrwxrwxrwx. 1 root root 57 Jan 30 09:29 dbus-org.freedesktop.nm-dispatcher.service -> /usr/lib/systemd/system/NetworkManager-dispatcher.service drwxr-xr-x. 2 root root 48 Jan 30 09:29 network-online.target.wants/ drwxr-xr-x. 2 root root 56 Jan 30 09:29 timers.target.wants/ drwxr-xr-x. 2 root root 38 Jan 30 09:29 dev-virtio\x2dports-org.qemu.guest_agent.0.device.wants/ drwxr-xr-x. 2 root root 31 Jan 30 09:29 basic.target.wants/ lrwxrwxrwx. 1 root root 41 Jan 30 09:31 default.target -> /usr/lib/systemd/system/multi-user.target drwxr-xr-x. 2 root root 70 Jan 30 09:48 sockets.target.wants/ drwxr-xr-x. 2 root root 31 Jan 30 09:48 remote-fs.target.wants/ drwxr-xr-x. 2 root root 59 Jan 30 09:50 sshd-keygen@.service.d/ drwxr-xr-x. 2 root root 119 Jan 30 09:50 cloud-init.target.wants/ drwxr-xr-x. 2 root root 4096 Feb 6 16:18 multi-user.target.wants/ lrwxrwxrwx. 1 root root 41 Feb 6 16:18 dbus-org.fedoraproject.FirewallD1.service -> /usr/lib/systemd/system/firewalld.service drwxr-xr-x. 13 root root 4096 Feb 6 16:18 ./ TASK [Cleanup] ***************************************************************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/tests/tests_quadlet_demo.yml:154 Tuesday 06 February 2024 16:18:48 +0000 (0:00:00.199) 0:01:34.906 ****** TASK [fedora.linux_system_roles.podman : Set platform/version specific variables] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:3 Tuesday 06 February 2024 16:18:48 +0000 (0:00:00.056) 0:01:34.962 ****** included: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml for sut TASK [fedora.linux_system_roles.podman : Ensure ansible_facts used by role] **** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:3 Tuesday 06 February 2024 16:18:48 +0000 (0:00:00.033) 0:01:34.995 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_required_facts | difference(ansible_facts.keys() | list) | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if system is ostree] ************ task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:11 Tuesday 06 February 2024 16:18:48 +0000 (0:00:00.023) 0:01:35.019 ****** skipping: [sut] => { "changed": false, "false_condition": "not __podman_is_ostree is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set flag to indicate system is ostree] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:16 Tuesday 06 February 2024 16:18:48 +0000 (0:00:00.021) 0:01:35.040 ****** skipping: [sut] => { "changed": false, "false_condition": "not __podman_is_ostree is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set platform/version specific variables] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:20 Tuesday 06 February 2024 16:18:48 +0000 (0:00:00.020) 0:01:35.061 ****** ok: [sut] => (item=RedHat.yml) => { "ansible_facts": { "__podman_packages": [ "podman", "shadow-utils-subid" ] }, "ansible_included_var_files": [ "/WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/vars/RedHat.yml" ], "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml" } skipping: [sut] => (item=CentOS.yml) => { "ansible_loop_var": "item", "changed": false, "false_condition": "__vars_file is file", "item": "CentOS.yml", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=CentOS_9.yml) => { "ansible_loop_var": "item", "changed": false, "false_condition": "__vars_file is file", "item": "CentOS_9.yml", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=CentOS_9.yml) => { "ansible_loop_var": "item", "changed": false, "false_condition": "__vars_file is file", "item": "CentOS_9.yml", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Gather the package facts] ************* task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:6 Tuesday 06 February 2024 16:18:48 +0000 (0:00:00.046) 0:01:35.108 ****** ok: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Enable copr if requested] ************* task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:10 Tuesday 06 February 2024 16:18:49 +0000 (0:00:00.974) 0:01:36.083 ****** skipping: [sut] => { "changed": false, "false_condition": "podman_use_copr | d(false)", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Ensure required packages are installed] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:14 Tuesday 06 February 2024 16:18:49 +0000 (0:00:00.021) 0:01:36.105 ****** skipping: [sut] => { "changed": false, "false_condition": "(__podman_packages | difference(ansible_facts.packages))", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get podman version] ******************* task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:22 Tuesday 06 February 2024 16:18:49 +0000 (0:00:00.034) 0:01:36.139 ****** ok: [sut] => { "changed": false, "cmd": [ "podman", "--version" ], "delta": "0:00:00.031235", "end": "2024-02-06 16:18:50.040404", "rc": 0, "start": "2024-02-06 16:18:50.009169" } STDOUT: podman version 4.8.1 TASK [fedora.linux_system_roles.podman : Set podman version] ******************* task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:27 Tuesday 06 February 2024 16:18:50 +0000 (0:00:00.249) 0:01:36.389 ****** ok: [sut] => { "ansible_facts": { "podman_version": "4.8.1" }, "changed": false } TASK [fedora.linux_system_roles.podman : Podman package version must be 4.2 or later] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:31 Tuesday 06 February 2024 16:18:50 +0000 (0:00:00.042) 0:01:36.432 ****** skipping: [sut] => { "changed": false, "false_condition": "podman_version is version(\"4.2\", \"<\")", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Podman package version must be 4.4 or later for quadlet, secrets] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:38 Tuesday 06 February 2024 16:18:50 +0000 (0:00:00.152) 0:01:36.585 ****** skipping: [sut] => { "changed": false, "false_condition": "podman_version is version(\"4.4\", \"<\")", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Podman package version must be 4.4 or later for quadlet, secrets] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:48 Tuesday 06 February 2024 16:18:50 +0000 (0:00:00.022) 0:01:36.607 ****** META: end_host conditional evaluated to False, continuing execution for sut skipping: [sut] => { "skip_reason": "end_host conditional evaluated to False, continuing execution for sut" } MSG: end_host conditional evaluated to false, continuing execution for sut TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:55 Tuesday 06 February 2024 16:18:50 +0000 (0:00:00.022) 0:01:36.630 ****** included: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for sut TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Tuesday 06 February 2024 16:18:50 +0000 (0:00:00.041) 0:01:36.671 ****** skipping: [sut] => { "changed": false, "false_condition": "'getent_passwd' not in ansible_facts or __podman_user not in ansible_facts['getent_passwd']", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Tuesday 06 February 2024 16:18:50 +0000 (0:00:00.025) 0:01:36.697 ****** skipping: [sut] => { "changed": false, "false_condition": "not ansible_facts[\"getent_passwd\"][__podman_user]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Tuesday 06 February 2024 16:18:50 +0000 (0:00:00.023) 0:01:36.720 ****** ok: [sut] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get group information] **************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:27 Tuesday 06 February 2024 16:18:50 +0000 (0:00:00.029) 0:01:36.749 ****** ok: [sut] => { "ansible_facts": { "getent_group": { "root": [ "x", "0", "" ] } }, "changed": false } TASK [fedora.linux_system_roles.podman : Set group name] *********************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:34 Tuesday 06 February 2024 16:18:50 +0000 (0:00:00.218) 0:01:36.968 ****** ok: [sut] => { "ansible_facts": { "__podman_group_name": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:38 Tuesday 06 February 2024 16:18:50 +0000 (0:00:00.038) 0:01:37.007 ****** ok: [sut] => { "changed": false, "stat": { "atime": 1707236301.2420683, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "076f89863b474504550ab1b5c11216a77ed7123c", "ctime": 1707236269.7653096, "dev": 51713, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 4471560, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-pie-executable", "mode": "0755", "mtime": 1689166618.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 15496, "uid": 0, "version": "396954581", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************ task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:49 Tuesday 06 February 2024 16:18:50 +0000 (0:00:00.261) 0:01:37.269 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check group with getsubids] *********** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:53 Tuesday 06 February 2024 16:18:51 +0000 (0:00:00.043) 0:01:37.312 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if user is in subuid file] ****** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:63 Tuesday 06 February 2024 16:18:51 +0000 (0:00:00.037) 0:01:37.350 ****** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:71 Tuesday 06 February 2024 16:18:51 +0000 (0:00:00.045) 0:01:37.396 ****** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if group is in subgid file] ***** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:78 Tuesday 06 February 2024 16:18:51 +0000 (0:00:00.041) 0:01:37.438 ****** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] ***** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:86 Tuesday 06 February 2024 16:18:51 +0000 (0:00:00.038) 0:01:37.477 ****** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set config file paths] **************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:61 Tuesday 06 February 2024 16:18:51 +0000 (0:00:00.031) 0:01:37.508 ****** ok: [sut] => { "ansible_facts": { "__podman_container_conf_file": "/etc/containers/containers.conf.d/50-systemroles.conf", "__podman_policy_json_file": "/etc/containers/policy.json", "__podman_registries_conf_file": "/etc/containers/registries.conf.d/50-systemroles.conf", "__podman_storage_conf_file": "/etc/containers/storage.conf" }, "changed": false } TASK [fedora.linux_system_roles.podman : Handle container.conf.d] ************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:70 Tuesday 06 February 2024 16:18:51 +0000 (0:00:00.050) 0:01:37.559 ****** included: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_container_conf_d.yml for sut TASK [fedora.linux_system_roles.podman : Ensure containers.d exists] *********** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_container_conf_d.yml:5 Tuesday 06 February 2024 16:18:51 +0000 (0:00:00.049) 0:01:37.609 ****** skipping: [sut] => { "changed": false, "false_condition": "podman_containers_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Update container config file] ********* task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_container_conf_d.yml:13 Tuesday 06 February 2024 16:18:51 +0000 (0:00:00.033) 0:01:37.642 ****** skipping: [sut] => { "changed": false, "false_condition": "podman_containers_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Handle registries.conf.d] ************* task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:73 Tuesday 06 February 2024 16:18:51 +0000 (0:00:00.145) 0:01:37.787 ****** included: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_registries_conf_d.yml for sut TASK [fedora.linux_system_roles.podman : Ensure registries.d exists] *********** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_registries_conf_d.yml:5 Tuesday 06 February 2024 16:18:51 +0000 (0:00:00.077) 0:01:37.864 ****** skipping: [sut] => { "changed": false, "false_condition": "podman_registries_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Update registries config file] ******** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_registries_conf_d.yml:13 Tuesday 06 February 2024 16:18:51 +0000 (0:00:00.046) 0:01:37.910 ****** skipping: [sut] => { "changed": false, "false_condition": "podman_registries_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Handle storage.conf] ****************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:76 Tuesday 06 February 2024 16:18:51 +0000 (0:00:00.044) 0:01:37.954 ****** included: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_storage_conf.yml for sut TASK [fedora.linux_system_roles.podman : Ensure storage.conf parent dir exists] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_storage_conf.yml:5 Tuesday 06 February 2024 16:18:51 +0000 (0:00:00.077) 0:01:38.032 ****** skipping: [sut] => { "changed": false, "false_condition": "podman_storage_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Update storage config file] *********** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_storage_conf.yml:13 Tuesday 06 February 2024 16:18:51 +0000 (0:00:00.038) 0:01:38.070 ****** skipping: [sut] => { "changed": false, "false_condition": "podman_storage_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Handle policy.json] ******************* task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:79 Tuesday 06 February 2024 16:18:51 +0000 (0:00:00.040) 0:01:38.111 ****** included: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml for sut TASK [fedora.linux_system_roles.podman : Ensure policy.json parent dir exists] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:6 Tuesday 06 February 2024 16:18:51 +0000 (0:00:00.077) 0:01:38.188 ****** skipping: [sut] => { "changed": false, "false_condition": "podman_policy_json | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Stat the policy.json file] ************ task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:14 Tuesday 06 February 2024 16:18:51 +0000 (0:00:00.038) 0:01:38.227 ****** skipping: [sut] => { "changed": false, "false_condition": "podman_policy_json | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get the existing policy.json] ********* task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:19 Tuesday 06 February 2024 16:18:51 +0000 (0:00:00.032) 0:01:38.260 ****** skipping: [sut] => { "changed": false, "false_condition": "podman_policy_json | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Write new policy.json file] *********** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:25 Tuesday 06 February 2024 16:18:51 +0000 (0:00:00.025) 0:01:38.286 ****** skipping: [sut] => { "changed": false, "false_condition": "podman_policy_json | length > 0", "skip_reason": "Conditional result was False" } TASK [Manage firewall for specified ports] ************************************* task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:85 Tuesday 06 February 2024 16:18:52 +0000 (0:00:00.023) 0:01:38.309 ****** TASK [fedora.linux_system_roles.firewall : Setup firewalld] ******************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:2 Tuesday 06 February 2024 16:18:52 +0000 (0:00:00.064) 0:01:38.374 ****** included: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml for sut TASK [fedora.linux_system_roles.firewall : Ensure ansible_facts used by role] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:2 Tuesday 06 February 2024 16:18:52 +0000 (0:00:00.031) 0:01:38.405 ****** skipping: [sut] => { "changed": false, "false_condition": "__firewall_required_facts | difference(ansible_facts.keys() | list) | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Check if system is ostree] ********** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:10 Tuesday 06 February 2024 16:18:52 +0000 (0:00:00.024) 0:01:38.430 ****** skipping: [sut] => { "changed": false, "false_condition": "not __firewall_is_ostree is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Set flag to indicate system is ostree] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:15 Tuesday 06 February 2024 16:18:52 +0000 (0:00:00.070) 0:01:38.501 ****** skipping: [sut] => { "changed": false, "false_condition": "not __firewall_is_ostree is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Install firewalld] ****************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:19 Tuesday 06 February 2024 16:18:52 +0000 (0:00:00.020) 0:01:38.521 ****** ok: [sut] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [fedora.linux_system_roles.firewall : Collect service facts] ************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:5 Tuesday 06 February 2024 16:18:53 +0000 (0:00:00.914) 0:01:39.435 ****** skipping: [sut] => { "changed": false, "false_condition": "firewall_disable_conflicting_services | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Attempt to stop and disable conflicting services] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:9 Tuesday 06 February 2024 16:18:53 +0000 (0:00:00.020) 0:01:39.456 ****** skipping: [sut] => (item=nftables) => { "ansible_loop_var": "item", "changed": false, "false_condition": "firewall_disable_conflicting_services | bool", "item": "nftables", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=iptables) => { "ansible_loop_var": "item", "changed": false, "false_condition": "firewall_disable_conflicting_services | bool", "item": "iptables", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=ufw) => { "ansible_loop_var": "item", "changed": false, "false_condition": "firewall_disable_conflicting_services | bool", "item": "ufw", "skip_reason": "Conditional result was False" } skipping: [sut] => { "changed": false } MSG: All items skipped TASK [fedora.linux_system_roles.firewall : Unmask firewalld service] *********** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:22 Tuesday 06 February 2024 16:18:53 +0000 (0:00:00.025) 0:01:39.482 ****** ok: [sut] => { "changed": false, "name": "firewalld", "status": { "AccessSELinuxContext": "system_u:object_r:firewalld_unit_file_t:s0", "ActiveEnterTimestamp": "Tue 2024-02-06 16:18:25 UTC", "ActiveEnterTimestampMonotonic": "551971131", "ActiveExitTimestampMonotonic": "0", "ActiveState": "active", "After": "sysinit.target basic.target dbus.socket system.slice polkit.service dbus-broker.service", "AllowIsolate": "no", "AssertResult": "yes", "AssertTimestamp": "Tue 2024-02-06 16:18:24 UTC", "AssertTimestampMonotonic": "550889183", "Before": "network-pre.target multi-user.target shutdown.target", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "BusName": "org.fedoraproject.FirewallD1", "CPUAccounting": "yes", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "676535000", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "yes", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore", "CleanResult": "success", "CollectMode": "inactive", "ConditionResult": "yes", "ConditionTimestamp": "Tue 2024-02-06 16:18:24 UTC", "ConditionTimestampMonotonic": "550889179", "ConfigurationDirectoryMode": "0755", "Conflicts": "ipset.service ebtables.service shutdown.target nftables.service iptables.service ip6tables.service", "ControlGroup": "/system.slice/firewalld.service", "ControlGroupId": "4528", "ControlPID": "0", "CoredumpFilter": "0x33", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "Delegate": "no", "Description": "firewalld - dynamic firewall daemon", "DevicePolicy": "auto", "Documentation": "\"man:firewalld(1)\"", "DynamicUser": "no", "EffectiveCPUs": "0-1", "EffectiveMemoryNodes": "0", "EnvironmentFiles": "/etc/sysconfig/firewalld (ignore_errors=yes)", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "11267", "ExecMainStartTimestamp": "Tue 2024-02-06 16:18:24 UTC", "ExecMainStartTimestampMonotonic": "550896723", "ExecMainStatus": "0", "ExecReload": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecReloadEx": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStart": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStartEx": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExitType": "main", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FinalKillSignal": "9", "FragmentPath": "/usr/lib/systemd/system/firewalld.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOReadBytes": "18446744073709551615", "IOReadOperations": "18446744073709551615", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "IOWeight": "[not set]", "IOWriteBytes": "18446744073709551615", "IOWriteOperations": "18446744073709551615", "IPAccounting": "no", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "Id": "firewalld.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestamp": "Tue 2024-02-06 16:18:24 UTC", "InactiveExitTimestampMonotonic": "550897090", "InvocationID": "182467a88ebc4b268ebfb3b439ab39a2", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "mixed", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "0", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "8388608", "LimitMEMLOCKSoft": "8388608", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "524288", "LimitNOFILESoft": "1024", "LimitNPROC": "13974", "LimitNPROCSoft": "13974", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "13974", "LimitSIGPENDINGSoft": "13974", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "11267", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "ManagedOOMSwap": "auto", "MemoryAccounting": "yes", "MemoryAvailable": "infinity", "MemoryCurrent": "32755712", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemorySwapMax": "infinity", "MountAPIVFS": "no", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAPolicy": "n/a", "Names": "firewalld.service dbus-org.fedoraproject.FirewallD1.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMPolicy": "stop", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "OnSuccessJobMode": "fail", "Perpetual": "no", "PrivateDevices": "no", "PrivateIPC": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProcSubset": "all", "ProtectClock": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectHostname": "no", "ProtectKernelLogs": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectProc": "default", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "ReloadResult": "success", "ReloadSignal": "1", "RemainAfterExit": "no", "RemoveIPC": "no", "Requires": "system.slice sysinit.target dbus.socket", "Restart": "no", "RestartKillSignal": "15", "RestartUSec": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "Slice": "system.slice", "StandardError": "null", "StandardInput": "null", "StandardOutput": "null", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StateChangeTimestamp": "Tue 2024-02-06 16:18:46 UTC", "StateChangeTimestampMonotonic": "573707019", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "running", "SuccessAction": "none", "SyslogFacility": "3", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "2147483646", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "2", "TasksMax": "22359", "TimeoutAbortUSec": "1min 30s", "TimeoutCleanUSec": "infinity", "TimeoutStartFailureMode": "terminate", "TimeoutStartUSec": "1min 30s", "TimeoutStopFailureMode": "terminate", "TimeoutStopUSec": "1min 30s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "dbus", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "enabled", "UnitFileState": "enabled", "UtmpMode": "init", "WantedBy": "multi-user.target", "Wants": "network-pre.target", "WatchdogSignal": "6", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "0" } } TASK [fedora.linux_system_roles.firewall : Enable and start firewalld service] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:28 Tuesday 06 February 2024 16:18:53 +0000 (0:00:00.358) 0:01:39.840 ****** ok: [sut] => { "changed": false, "enabled": true, "name": "firewalld", "state": "started", "status": { "AccessSELinuxContext": "system_u:object_r:firewalld_unit_file_t:s0", "ActiveEnterTimestamp": "Tue 2024-02-06 16:18:25 UTC", "ActiveEnterTimestampMonotonic": "551971131", "ActiveExitTimestampMonotonic": "0", "ActiveState": "active", "After": "sysinit.target basic.target dbus.socket system.slice polkit.service dbus-broker.service", "AllowIsolate": "no", "AssertResult": "yes", "AssertTimestamp": "Tue 2024-02-06 16:18:24 UTC", "AssertTimestampMonotonic": "550889183", "Before": "network-pre.target multi-user.target shutdown.target", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "BusName": "org.fedoraproject.FirewallD1", "CPUAccounting": "yes", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "676535000", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "yes", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore", "CleanResult": "success", "CollectMode": "inactive", "ConditionResult": "yes", "ConditionTimestamp": "Tue 2024-02-06 16:18:24 UTC", "ConditionTimestampMonotonic": "550889179", "ConfigurationDirectoryMode": "0755", "Conflicts": "ipset.service ebtables.service shutdown.target nftables.service iptables.service ip6tables.service", "ControlGroup": "/system.slice/firewalld.service", "ControlGroupId": "4528", "ControlPID": "0", "CoredumpFilter": "0x33", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "Delegate": "no", "Description": "firewalld - dynamic firewall daemon", "DevicePolicy": "auto", "Documentation": "\"man:firewalld(1)\"", "DynamicUser": "no", "EffectiveCPUs": "0-1", "EffectiveMemoryNodes": "0", "EnvironmentFiles": "/etc/sysconfig/firewalld (ignore_errors=yes)", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "11267", "ExecMainStartTimestamp": "Tue 2024-02-06 16:18:24 UTC", "ExecMainStartTimestampMonotonic": "550896723", "ExecMainStatus": "0", "ExecReload": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecReloadEx": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStart": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStartEx": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExitType": "main", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FinalKillSignal": "9", "FragmentPath": "/usr/lib/systemd/system/firewalld.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOReadBytes": "18446744073709551615", "IOReadOperations": "18446744073709551615", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "IOWeight": "[not set]", "IOWriteBytes": "18446744073709551615", "IOWriteOperations": "18446744073709551615", "IPAccounting": "no", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "Id": "firewalld.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestamp": "Tue 2024-02-06 16:18:24 UTC", "InactiveExitTimestampMonotonic": "550897090", "InvocationID": "182467a88ebc4b268ebfb3b439ab39a2", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "mixed", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "0", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "8388608", "LimitMEMLOCKSoft": "8388608", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "524288", "LimitNOFILESoft": "1024", "LimitNPROC": "13974", "LimitNPROCSoft": "13974", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "13974", "LimitSIGPENDINGSoft": "13974", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "11267", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "ManagedOOMSwap": "auto", "MemoryAccounting": "yes", "MemoryAvailable": "infinity", "MemoryCurrent": "32755712", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemorySwapMax": "infinity", "MountAPIVFS": "no", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAPolicy": "n/a", "Names": "firewalld.service dbus-org.fedoraproject.FirewallD1.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMPolicy": "stop", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "OnSuccessJobMode": "fail", "Perpetual": "no", "PrivateDevices": "no", "PrivateIPC": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProcSubset": "all", "ProtectClock": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectHostname": "no", "ProtectKernelLogs": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectProc": "default", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "ReloadResult": "success", "ReloadSignal": "1", "RemainAfterExit": "no", "RemoveIPC": "no", "Requires": "system.slice sysinit.target dbus.socket", "Restart": "no", "RestartKillSignal": "15", "RestartUSec": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "Slice": "system.slice", "StandardError": "null", "StandardInput": "null", "StandardOutput": "null", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StateChangeTimestamp": "Tue 2024-02-06 16:18:46 UTC", "StateChangeTimestampMonotonic": "573707019", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "running", "SuccessAction": "none", "SyslogFacility": "3", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "2147483646", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "2", "TasksMax": "22359", "TimeoutAbortUSec": "1min 30s", "TimeoutCleanUSec": "infinity", "TimeoutStartFailureMode": "terminate", "TimeoutStartUSec": "1min 30s", "TimeoutStopFailureMode": "terminate", "TimeoutStopUSec": "1min 30s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "dbus", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "enabled", "UnitFileState": "enabled", "UtmpMode": "init", "WantedBy": "multi-user.target", "Wants": "network-pre.target", "WatchdogSignal": "6", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "0" } } TASK [fedora.linux_system_roles.firewall : Check if previous replaced is defined] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:34 Tuesday 06 February 2024 16:18:53 +0000 (0:00:00.357) 0:01:40.198 ****** ok: [sut] => { "ansible_facts": { "__firewall_previous_replaced": false, "__firewall_python_cmd": "/usr/bin/python3", "__firewall_report_changed": true }, "changed": false } TASK [fedora.linux_system_roles.firewall : Get config files, checksums before and remove] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:43 Tuesday 06 February 2024 16:18:53 +0000 (0:00:00.027) 0:01:40.226 ****** skipping: [sut] => { "changed": false, "false_condition": "__firewall_previous_replaced | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Tell firewall module it is able to report changed] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:55 Tuesday 06 February 2024 16:18:53 +0000 (0:00:00.021) 0:01:40.247 ****** skipping: [sut] => { "changed": false, "false_condition": "__firewall_previous_replaced | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Configure firewall] ***************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:71 Tuesday 06 February 2024 16:18:53 +0000 (0:00:00.020) 0:01:40.268 ****** ok: [sut] => (item={'port': '8000/tcp', 'state': 'enabled'}) => { "__firewall_changed": false, "ansible_loop_var": "item", "changed": false, "item": { "port": "8000/tcp", "state": "enabled" } } ok: [sut] => (item={'port': '9000/tcp', 'state': 'enabled'}) => { "__firewall_changed": false, "ansible_loop_var": "item", "changed": false, "item": { "port": "9000/tcp", "state": "enabled" } } TASK [fedora.linux_system_roles.firewall : Gather firewall config information] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:120 Tuesday 06 February 2024 16:18:54 +0000 (0:00:00.720) 0:01:40.988 ****** skipping: [sut] => (item={'port': '8000/tcp', 'state': 'enabled'}) => { "ansible_loop_var": "item", "changed": false, "false_condition": "firewall | length == 1", "item": { "port": "8000/tcp", "state": "enabled" }, "skip_reason": "Conditional result was False" } skipping: [sut] => (item={'port': '9000/tcp', 'state': 'enabled'}) => { "ansible_loop_var": "item", "changed": false, "false_condition": "firewall | length == 1", "item": { "port": "9000/tcp", "state": "enabled" }, "skip_reason": "Conditional result was False" } skipping: [sut] => { "changed": false } MSG: All items skipped TASK [fedora.linux_system_roles.firewall : Update firewalld_config fact] ******* task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:130 Tuesday 06 February 2024 16:18:54 +0000 (0:00:00.032) 0:01:41.021 ****** skipping: [sut] => { "changed": false, "false_condition": "firewall | length == 1", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Gather firewall config if no arguments] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:139 Tuesday 06 February 2024 16:18:54 +0000 (0:00:00.024) 0:01:41.045 ****** skipping: [sut] => { "changed": false, "false_condition": "firewall == None or firewall | length == 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Update firewalld_config fact] ******* task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:144 Tuesday 06 February 2024 16:18:54 +0000 (0:00:00.022) 0:01:41.068 ****** skipping: [sut] => { "changed": false, "false_condition": "firewall == None or firewall | length == 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Get config files, checksums after] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:153 Tuesday 06 February 2024 16:18:54 +0000 (0:00:00.024) 0:01:41.092 ****** skipping: [sut] => { "changed": false, "false_condition": "__firewall_previous_replaced | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Calculate what has changed] ********* task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:163 Tuesday 06 February 2024 16:18:54 +0000 (0:00:00.019) 0:01:41.111 ****** skipping: [sut] => { "changed": false, "false_condition": "__firewall_previous_replaced | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Show diffs] ************************* task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:169 Tuesday 06 February 2024 16:18:54 +0000 (0:00:00.021) 0:01:41.132 ****** skipping: [sut] => { "false_condition": "__firewall_previous_replaced | bool" } TASK [Manage selinux for specified ports] ************************************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:92 Tuesday 06 February 2024 16:18:54 +0000 (0:00:00.032) 0:01:41.165 ****** skipping: [sut] => { "changed": false, "false_condition": "podman_selinux_ports | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Keep track of users that need to cancel linger] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:99 Tuesday 06 February 2024 16:18:54 +0000 (0:00:00.022) 0:01:41.187 ****** ok: [sut] => { "ansible_facts": { "__podman_cancel_user_linger": [] }, "changed": false } TASK [fedora.linux_system_roles.podman : Handle secrets] *********************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:103 Tuesday 06 February 2024 16:18:54 +0000 (0:00:00.021) 0:01:41.209 ****** included: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml for sut => (item=(censored due to no_log)) included: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml for sut => (item=(censored due to no_log)) included: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml for sut => (item=(censored due to no_log)) TASK [fedora.linux_system_roles.podman : Set variables part 0] ***************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:3 Tuesday 06 February 2024 16:18:55 +0000 (0:00:00.109) 0:01:41.318 ****** ok: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Set variables part 1] ***************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:10 Tuesday 06 February 2024 16:18:55 +0000 (0:00:00.027) 0:01:41.345 ****** ok: [sut] => { "ansible_facts": { "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set variables part 2] ***************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:14 Tuesday 06 February 2024 16:18:55 +0000 (0:00:00.025) 0:01:41.371 ****** ok: [sut] => { "ansible_facts": { "__podman_rootless": false, "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:20 Tuesday 06 February 2024 16:18:55 +0000 (0:00:00.028) 0:01:41.399 ****** included: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for sut TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Tuesday 06 February 2024 16:18:55 +0000 (0:00:00.033) 0:01:41.433 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Tuesday 06 February 2024 16:18:55 +0000 (0:00:00.022) 0:01:41.455 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Tuesday 06 February 2024 16:18:55 +0000 (0:00:00.022) 0:01:41.477 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Stat XDG_RUNTIME_DIR] ***************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:25 Tuesday 06 February 2024 16:18:55 +0000 (0:00:00.022) 0:01:41.500 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Manage each secret] ******************* task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:36 Tuesday 06 February 2024 16:18:55 +0000 (0:00:00.020) 0:01:41.521 ****** ok: [sut] => { "changed": false } TASK [fedora.linux_system_roles.podman : Debug] ******************************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:53 Tuesday 06 February 2024 16:18:55 +0000 (0:00:00.259) 0:01:41.780 ****** changed: [sut] => { "changed": true, "cmd": [ "podman", "secret", "ls" ], "delta": "0:00:00.036882", "end": "2024-02-06 16:18:55.690737", "rc": 0, "start": "2024-02-06 16:18:55.653855" } STDOUT: ID NAME DRIVER CREATED UPDATED TASK [fedora.linux_system_roles.podman : Set variables part 0] ***************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:3 Tuesday 06 February 2024 16:18:55 +0000 (0:00:00.242) 0:01:42.023 ****** ok: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Set variables part 1] ***************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:10 Tuesday 06 February 2024 16:18:55 +0000 (0:00:00.026) 0:01:42.049 ****** ok: [sut] => { "ansible_facts": { "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set variables part 2] ***************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:14 Tuesday 06 February 2024 16:18:55 +0000 (0:00:00.026) 0:01:42.075 ****** ok: [sut] => { "ansible_facts": { "__podman_rootless": false, "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:20 Tuesday 06 February 2024 16:18:55 +0000 (0:00:00.027) 0:01:42.102 ****** included: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for sut TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Tuesday 06 February 2024 16:18:55 +0000 (0:00:00.034) 0:01:42.137 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Tuesday 06 February 2024 16:18:55 +0000 (0:00:00.023) 0:01:42.161 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Tuesday 06 February 2024 16:18:55 +0000 (0:00:00.020) 0:01:42.181 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Stat XDG_RUNTIME_DIR] ***************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:25 Tuesday 06 February 2024 16:18:55 +0000 (0:00:00.022) 0:01:42.203 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Manage each secret] ******************* task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:36 Tuesday 06 February 2024 16:18:55 +0000 (0:00:00.019) 0:01:42.223 ****** ok: [sut] => { "changed": false } TASK [fedora.linux_system_roles.podman : Debug] ******************************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:53 Tuesday 06 February 2024 16:18:56 +0000 (0:00:00.334) 0:01:42.557 ****** changed: [sut] => { "changed": true, "cmd": [ "podman", "secret", "ls" ], "delta": "0:00:00.040576", "end": "2024-02-06 16:18:56.471620", "rc": 0, "start": "2024-02-06 16:18:56.431044" } STDOUT: ID NAME DRIVER CREATED UPDATED TASK [fedora.linux_system_roles.podman : Set variables part 0] ***************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:3 Tuesday 06 February 2024 16:18:56 +0000 (0:00:00.243) 0:01:42.801 ****** ok: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Set variables part 1] ***************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:10 Tuesday 06 February 2024 16:18:56 +0000 (0:00:00.025) 0:01:42.827 ****** ok: [sut] => { "ansible_facts": { "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set variables part 2] ***************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:14 Tuesday 06 February 2024 16:18:56 +0000 (0:00:00.024) 0:01:42.851 ****** ok: [sut] => { "ansible_facts": { "__podman_rootless": false, "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:20 Tuesday 06 February 2024 16:18:56 +0000 (0:00:00.027) 0:01:42.878 ****** included: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for sut TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Tuesday 06 February 2024 16:18:56 +0000 (0:00:00.031) 0:01:42.909 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Tuesday 06 February 2024 16:18:56 +0000 (0:00:00.018) 0:01:42.928 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Tuesday 06 February 2024 16:18:56 +0000 (0:00:00.021) 0:01:42.949 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Stat XDG_RUNTIME_DIR] ***************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:25 Tuesday 06 February 2024 16:18:56 +0000 (0:00:00.021) 0:01:42.970 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Manage each secret] ******************* task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:36 Tuesday 06 February 2024 16:18:56 +0000 (0:00:00.022) 0:01:42.992 ****** ok: [sut] => { "changed": false } TASK [fedora.linux_system_roles.podman : Debug] ******************************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:53 Tuesday 06 February 2024 16:18:56 +0000 (0:00:00.258) 0:01:43.251 ****** changed: [sut] => { "changed": true, "cmd": [ "podman", "secret", "ls" ], "delta": "0:00:00.035902", "end": "2024-02-06 16:18:57.159939", "rc": 0, "start": "2024-02-06 16:18:57.124037" } STDOUT: ID NAME DRIVER CREATED UPDATED TASK [fedora.linux_system_roles.podman : Handle Kubernetes specifications] ***** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:110 Tuesday 06 February 2024 16:18:57 +0000 (0:00:00.239) 0:01:43.490 ****** skipping: [sut] => { "changed": false, "skipped_reason": "No items in the list" } TASK [fedora.linux_system_roles.podman : Handle Quadlet specifications] ******** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:116 Tuesday 06 February 2024 16:18:57 +0000 (0:00:00.018) 0:01:43.509 ****** included: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for sut => (item={'file_src': 'quadlet-demo-mysql.volume', 'state': 'absent'}) included: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for sut => (item={'template_src': 'quadlet-demo-mysql.container.j2', 'state': 'absent'}) included: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for sut => (item={'file_src': 'envoy-proxy-configmap.yml', 'state': 'absent'}) included: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for sut => (item={'template_src': 'quadlet-demo.yml.j2', 'state': 'absent'}) included: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for sut => (item={'file_src': 'quadlet-demo.kube', 'state': 'absent'}) included: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for sut => (item={'file_src': 'quadlet-demo.network', 'state': 'absent'}) TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14 Tuesday 06 February 2024 16:18:57 +0000 (0:00:00.093) 0:01:43.602 ****** ok: [sut] => { "ansible_facts": { "__podman_quadlet_file_src": "quadlet-demo-mysql.volume", "__podman_quadlet_spec": {}, "__podman_quadlet_str": "[Volume]", "__podman_quadlet_template_src": "" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25 Tuesday 06 February 2024 16:18:57 +0000 (0:00:00.031) 0:01:43.633 ****** ok: [sut] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_pull_image": true, "__podman_state": "absent", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] ***** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35 Tuesday 06 February 2024 16:18:57 +0000 (0:00:00.026) 0:01:43.660 ****** skipping: [sut] => { "changed": false, "false_condition": "not __podman_quadlet_file_src", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48 Tuesday 06 February 2024 16:18:57 +0000 (0:00:00.024) 0:01:43.684 ****** ok: [sut] => { "ansible_facts": { "__podman_quadlet_name": "quadlet-demo-mysql", "__podman_quadlet_type": "volume", "__podman_rootless": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57 Tuesday 06 February 2024 16:18:57 +0000 (0:00:00.032) 0:01:43.717 ****** included: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for sut TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Tuesday 06 February 2024 16:18:57 +0000 (0:00:00.098) 0:01:43.815 ****** skipping: [sut] => { "changed": false, "false_condition": "'getent_passwd' not in ansible_facts or __podman_user not in ansible_facts['getent_passwd']", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Tuesday 06 February 2024 16:18:57 +0000 (0:00:00.023) 0:01:43.839 ****** skipping: [sut] => { "changed": false, "false_condition": "not ansible_facts[\"getent_passwd\"][__podman_user]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Tuesday 06 February 2024 16:18:57 +0000 (0:00:00.024) 0:01:43.863 ****** ok: [sut] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get group information] **************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:27 Tuesday 06 February 2024 16:18:57 +0000 (0:00:00.028) 0:01:43.892 ****** ok: [sut] => { "ansible_facts": { "getent_group": { "root": [ "x", "0", "" ] } }, "changed": false } TASK [fedora.linux_system_roles.podman : Set group name] *********************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:34 Tuesday 06 February 2024 16:18:57 +0000 (0:00:00.208) 0:01:44.101 ****** ok: [sut] => { "ansible_facts": { "__podman_group_name": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:38 Tuesday 06 February 2024 16:18:57 +0000 (0:00:00.029) 0:01:44.130 ****** ok: [sut] => { "changed": false, "stat": { "atime": 1707236301.2420683, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "076f89863b474504550ab1b5c11216a77ed7123c", "ctime": 1707236269.7653096, "dev": 51713, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 4471560, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-pie-executable", "mode": "0755", "mtime": 1689166618.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 15496, "uid": 0, "version": "396954581", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************ task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:49 Tuesday 06 February 2024 16:18:58 +0000 (0:00:00.206) 0:01:44.337 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check group with getsubids] *********** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:53 Tuesday 06 February 2024 16:18:58 +0000 (0:00:00.021) 0:01:44.359 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if user is in subuid file] ****** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:63 Tuesday 06 February 2024 16:18:58 +0000 (0:00:00.018) 0:01:44.377 ****** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:71 Tuesday 06 February 2024 16:18:58 +0000 (0:00:00.020) 0:01:44.397 ****** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if group is in subgid file] ***** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:78 Tuesday 06 February 2024 16:18:58 +0000 (0:00:00.019) 0:01:44.417 ****** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] ***** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:86 Tuesday 06 February 2024 16:18:58 +0000 (0:00:00.020) 0:01:44.438 ****** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62 Tuesday 06 February 2024 16:18:58 +0000 (0:00:00.020) 0:01:44.458 ****** ok: [sut] => { "ansible_facts": { "__podman_activate_systemd_unit": true, "__podman_images_found": [], "__podman_kube_yamls_raw": "", "__podman_service_name": "quadlet-demo-mysql-volume.service", "__podman_systemd_scope": "system", "__podman_user_home_dir": "/root", "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73 Tuesday 06 February 2024 16:18:58 +0000 (0:00:00.036) 0:01:44.495 ****** ok: [sut] => { "ansible_facts": { "__podman_quadlet_path": "/etc/containers/systemd" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get kube yaml contents] *************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77 Tuesday 06 February 2024 16:18:58 +0000 (0:00:00.021) 0:01:44.516 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_state != \"absent\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87 Tuesday 06 February 2024 16:18:58 +0000 (0:00:00.020) 0:01:44.537 ****** ok: [sut] => { "ansible_facts": { "__podman_images": [], "__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo-mysql.volume", "__podman_volumes": [] }, "changed": false } TASK [fedora.linux_system_roles.podman : Cleanup quadlets] ********************* task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:103 Tuesday 06 February 2024 16:18:58 +0000 (0:00:00.048) 0:01:44.585 ****** included: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml for sut TASK [fedora.linux_system_roles.podman : Stat XDG_RUNTIME_DIR] ***************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:4 Tuesday 06 February 2024 16:18:58 +0000 (0:00:00.041) 0:01:44.627 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Stop and disable service] ************* task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:12 Tuesday 06 February 2024 16:18:58 +0000 (0:00:00.020) 0:01:44.647 ****** changed: [sut] => { "changed": true, "enabled": false, "failed_when_result": false, "name": "quadlet-demo-mysql-volume.service", "state": "stopped", "status": { "AccessSELinuxContext": "system_u:object_r:systemd_unit_file_t:s0", "ActiveEnterTimestamp": "Tue 2024-02-06 16:18:35 UTC", "ActiveEnterTimestampMonotonic": "562455789", "ActiveExitTimestampMonotonic": "0", "ActiveState": "active", "After": "sysinit.target systemd-journald.socket -.mount basic.target system.slice", "AllowIsolate": "no", "AssertResult": "yes", "AssertTimestamp": "Tue 2024-02-06 16:18:35 UTC", "AssertTimestampMonotonic": "562397368", "Before": "quadlet-demo-mysql.service shutdown.target", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "CPUAccounting": "yes", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "44770000", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "no", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore", "CleanResult": "success", "CollectMode": "inactive", "ConditionResult": "yes", "ConditionTimestamp": "Tue 2024-02-06 16:18:35 UTC", "ConditionTimestampMonotonic": "562397363", "ConfigurationDirectoryMode": "0755", "Conflicts": "shutdown.target", "ControlGroupId": "0", "ControlPID": "0", "CoredumpFilter": "0x33", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "Delegate": "no", "Description": "quadlet-demo-mysql-volume.service", "DevicePolicy": "auto", "DynamicUser": "no", "ExecMainCode": "1", "ExecMainExitTimestamp": "Tue 2024-02-06 16:18:35 UTC", "ExecMainExitTimestampMonotonic": "562455663", "ExecMainPID": "11963", "ExecMainStartTimestamp": "Tue 2024-02-06 16:18:35 UTC", "ExecMainStartTimestampMonotonic": "562405705", "ExecMainStatus": "0", "ExecStart": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman volume create --ignore systemd-quadlet-demo-mysql ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStartEx": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman volume create --ignore systemd-quadlet-demo-mysql ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExitType": "main", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FinalKillSignal": "9", "FragmentPath": "/run/systemd/generator/quadlet-demo-mysql-volume.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOReadBytes": "18446744073709551615", "IOReadOperations": "18446744073709551615", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "IOWeight": "[not set]", "IOWriteBytes": "18446744073709551615", "IOWriteOperations": "18446744073709551615", "IPAccounting": "no", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "Id": "quadlet-demo-mysql-volume.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestamp": "Tue 2024-02-06 16:18:35 UTC", "InactiveExitTimestampMonotonic": "562406010", "InvocationID": "6a3b28e69999429caa2d13e71c08f941", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "control-group", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "0", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "8388608", "LimitMEMLOCKSoft": "8388608", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "524288", "LimitNOFILESoft": "1024", "LimitNPROC": "13974", "LimitNPROCSoft": "13974", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "13974", "LimitSIGPENDINGSoft": "13974", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "0", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "ManagedOOMSwap": "auto", "MemoryAccounting": "yes", "MemoryAvailable": "infinity", "MemoryCurrent": "[not set]", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemorySwapMax": "infinity", "MountAPIVFS": "no", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAPolicy": "n/a", "Names": "quadlet-demo-mysql-volume.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMPolicy": "stop", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "OnSuccessJobMode": "fail", "Perpetual": "no", "PrivateDevices": "no", "PrivateIPC": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProcSubset": "all", "ProtectClock": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectHostname": "no", "ProtectKernelLogs": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectProc": "default", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "ReloadResult": "success", "ReloadSignal": "1", "RemainAfterExit": "yes", "RemoveIPC": "no", "RequiredBy": "quadlet-demo-mysql.service", "Requires": "sysinit.target system.slice -.mount", "RequiresMountsFor": "/run/containers", "Restart": "no", "RestartKillSignal": "15", "RestartUSec": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "Slice": "system.slice", "StandardError": "inherit", "StandardInput": "null", "StandardOutput": "journal", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StateChangeTimestamp": "Tue 2024-02-06 16:18:35 UTC", "StateChangeTimestampMonotonic": "562455789", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "exited", "SuccessAction": "none", "SyslogFacility": "3", "SyslogIdentifier": "quadlet-demo-mysql-volume", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "2147483646", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "[not set]", "TasksMax": "22359", "TimeoutAbortUSec": "1min 30s", "TimeoutCleanUSec": "infinity", "TimeoutStartFailureMode": "terminate", "TimeoutStartUSec": "infinity", "TimeoutStopFailureMode": "terminate", "TimeoutStopUSec": "1min 30s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "oneshot", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "disabled", "UnitFileState": "generated", "UtmpMode": "init", "WatchdogSignal": "6", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "0" } } TASK [fedora.linux_system_roles.podman : Remove quadlet file] ****************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:32 Tuesday 06 February 2024 16:18:58 +0000 (0:00:00.566) 0:01:45.213 ****** changed: [sut] => { "changed": true, "path": "/etc/containers/systemd/quadlet-demo-mysql.volume", "state": "absent" } TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:38 Tuesday 06 February 2024 16:18:59 +0000 (0:00:00.263) 0:01:45.477 ****** included: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for sut TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Tuesday 06 February 2024 16:18:59 +0000 (0:00:00.037) 0:01:45.514 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Tuesday 06 February 2024 16:18:59 +0000 (0:00:00.019) 0:01:45.533 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Tuesday 06 February 2024 16:18:59 +0000 (0:00:00.021) 0:01:45.555 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Reload systemctl] ********************* task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:46 Tuesday 06 February 2024 16:18:59 +0000 (0:00:00.019) 0:01:45.574 ****** ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.podman : Prune images no longer in use] ******** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:55 Tuesday 06 February 2024 16:18:59 +0000 (0:00:00.528) 0:01:46.103 ****** changed: [sut] => { "changed": true, "cmd": [ "podman", "image", "prune", "-f" ], "delta": "0:00:00.039204", "end": "2024-02-06 16:19:00.017969", "rc": 0, "start": "2024-02-06 16:18:59.978765" } TASK [fedora.linux_system_roles.podman : Remove network] *********************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:63 Tuesday 06 February 2024 16:19:00 +0000 (0:00:00.247) 0:01:46.351 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_quadlet_type == \"network\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create and update quadlets] *********** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:107 Tuesday 06 February 2024 16:19:00 +0000 (0:00:00.026) 0:01:46.377 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_state != \"absent\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14 Tuesday 06 February 2024 16:19:00 +0000 (0:00:00.020) 0:01:46.398 ****** ok: [sut] => { "ansible_facts": { "__podman_quadlet_file_src": "", "__podman_quadlet_spec": {}, "__podman_quadlet_str": "[Install]\nWantedBy=default.target\n\n[Container]\nImage=quay.io/linux-system-roles/mysql:5.6\nContainerName=quadlet-demo-mysql\nVolume=quadlet-demo-mysql.volume:/var/lib/mysql\nVolume=/tmp/quadlet_demo:/var/lib/quadlet_demo:Z\nNetwork=quadlet-demo.network\nSecret=mysql-root-password-container,type=env,target=MYSQL_ROOT_PASSWORD\nHealthCmd=/usr/bin/true\nHealthOnFailure=kill\n", "__podman_quadlet_template_src": "quadlet-demo-mysql.container.j2" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25 Tuesday 06 February 2024 16:19:00 +0000 (0:00:00.069) 0:01:46.467 ****** ok: [sut] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_pull_image": true, "__podman_state": "absent", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] ***** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35 Tuesday 06 February 2024 16:19:00 +0000 (0:00:00.029) 0:01:46.497 ****** skipping: [sut] => { "changed": false, "false_condition": "not __podman_quadlet_str", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48 Tuesday 06 February 2024 16:19:00 +0000 (0:00:00.023) 0:01:46.521 ****** ok: [sut] => { "ansible_facts": { "__podman_quadlet_name": "quadlet-demo-mysql", "__podman_quadlet_type": "container", "__podman_rootless": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57 Tuesday 06 February 2024 16:19:00 +0000 (0:00:00.032) 0:01:46.554 ****** included: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for sut TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Tuesday 06 February 2024 16:19:00 +0000 (0:00:00.040) 0:01:46.594 ****** skipping: [sut] => { "changed": false, "false_condition": "'getent_passwd' not in ansible_facts or __podman_user not in ansible_facts['getent_passwd']", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Tuesday 06 February 2024 16:19:00 +0000 (0:00:00.029) 0:01:46.624 ****** skipping: [sut] => { "changed": false, "false_condition": "not ansible_facts[\"getent_passwd\"][__podman_user]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Tuesday 06 February 2024 16:19:00 +0000 (0:00:00.035) 0:01:46.659 ****** ok: [sut] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get group information] **************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:27 Tuesday 06 February 2024 16:19:00 +0000 (0:00:00.028) 0:01:46.688 ****** ok: [sut] => { "ansible_facts": { "getent_group": { "root": [ "x", "0", "" ] } }, "changed": false } TASK [fedora.linux_system_roles.podman : Set group name] *********************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:34 Tuesday 06 February 2024 16:19:00 +0000 (0:00:00.222) 0:01:46.910 ****** ok: [sut] => { "ansible_facts": { "__podman_group_name": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:38 Tuesday 06 February 2024 16:19:00 +0000 (0:00:00.028) 0:01:46.939 ****** ok: [sut] => { "changed": false, "stat": { "atime": 1707236301.2420683, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "076f89863b474504550ab1b5c11216a77ed7123c", "ctime": 1707236269.7653096, "dev": 51713, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 4471560, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-pie-executable", "mode": "0755", "mtime": 1689166618.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 15496, "uid": 0, "version": "396954581", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************ task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:49 Tuesday 06 February 2024 16:19:00 +0000 (0:00:00.207) 0:01:47.146 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check group with getsubids] *********** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:53 Tuesday 06 February 2024 16:19:00 +0000 (0:00:00.076) 0:01:47.222 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if user is in subuid file] ****** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:63 Tuesday 06 February 2024 16:19:00 +0000 (0:00:00.021) 0:01:47.244 ****** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:71 Tuesday 06 February 2024 16:19:00 +0000 (0:00:00.020) 0:01:47.264 ****** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if group is in subgid file] ***** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:78 Tuesday 06 February 2024 16:19:00 +0000 (0:00:00.020) 0:01:47.285 ****** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] ***** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:86 Tuesday 06 February 2024 16:19:01 +0000 (0:00:00.019) 0:01:47.304 ****** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62 Tuesday 06 February 2024 16:19:01 +0000 (0:00:00.023) 0:01:47.327 ****** ok: [sut] => { "ansible_facts": { "__podman_activate_systemd_unit": true, "__podman_images_found": [ "quay.io/linux-system-roles/mysql:5.6" ], "__podman_kube_yamls_raw": "", "__podman_service_name": "quadlet-demo-mysql.service", "__podman_systemd_scope": "system", "__podman_user_home_dir": "/root", "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73 Tuesday 06 February 2024 16:19:01 +0000 (0:00:00.038) 0:01:47.365 ****** ok: [sut] => { "ansible_facts": { "__podman_quadlet_path": "/etc/containers/systemd" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get kube yaml contents] *************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77 Tuesday 06 February 2024 16:19:01 +0000 (0:00:00.024) 0:01:47.390 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_state != \"absent\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87 Tuesday 06 February 2024 16:19:01 +0000 (0:00:00.019) 0:01:47.410 ****** ok: [sut] => { "ansible_facts": { "__podman_images": [ "quay.io/linux-system-roles/mysql:5.6" ], "__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo-mysql.container", "__podman_volumes": [ "/tmp/quadlet_demo" ] }, "changed": false } TASK [fedora.linux_system_roles.podman : Cleanup quadlets] ********************* task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:103 Tuesday 06 February 2024 16:19:01 +0000 (0:00:00.052) 0:01:47.462 ****** included: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml for sut TASK [fedora.linux_system_roles.podman : Stat XDG_RUNTIME_DIR] ***************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:4 Tuesday 06 February 2024 16:19:01 +0000 (0:00:00.041) 0:01:47.504 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Stop and disable service] ************* task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:12 Tuesday 06 February 2024 16:19:01 +0000 (0:00:00.018) 0:01:47.523 ****** changed: [sut] => { "changed": true, "enabled": false, "failed_when_result": false, "name": "quadlet-demo-mysql.service", "state": "stopped", "status": { "AccessSELinuxContext": "system_u:object_r:systemd_unit_file_t:s0", "ActiveEnterTimestampMonotonic": "0", "ActiveExitTimestampMonotonic": "0", "ActiveState": "failed", "After": "sysinit.target systemd-journald.socket tmp.mount -.mount quadlet-demo-network.service basic.target system.slice quadlet-demo-mysql-volume.service", "AllowIsolate": "no", "AssertResult": "yes", "AssertTimestamp": "Tue 2024-02-06 16:18:47 UTC", "AssertTimestampMonotonic": "574122831", "Before": "multi-user.target shutdown.target", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "CPUAccounting": "yes", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "93252000", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "no", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore", "CleanResult": "success", "CollectMode": "inactive", "ConditionResult": "yes", "ConditionTimestamp": "Tue 2024-02-06 16:18:47 UTC", "ConditionTimestampMonotonic": "574122824", "ConfigurationDirectoryMode": "0755", "Conflicts": "shutdown.target", "ControlGroupId": "0", "ControlPID": "0", "CoredumpFilter": "0x33", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "Delegate": "yes", "DelegateControllers": "cpu cpuacct cpuset io blkio memory devices pids bpf-firewall bpf-devices bpf-foreign bpf-socket-bind bpf-restrict-network-interfaces", "Description": "quadlet-demo-mysql.service", "DevicePolicy": "auto", "DynamicUser": "no", "Environment": "PODMAN_SYSTEMD_UNIT=quadlet-demo-mysql.service", "ExecMainCode": "1", "ExecMainExitTimestamp": "Tue 2024-02-06 16:18:47 UTC", "ExecMainExitTimestampMonotonic": "574177524", "ExecMainPID": "12354", "ExecMainStartTimestamp": "Tue 2024-02-06 16:18:47 UTC", "ExecMainStartTimestampMonotonic": "574126851", "ExecMainStatus": "125", "ExecStart": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman run --name=quadlet-demo-mysql --cidfile=/run/quadlet-demo-mysql.cid --replace --rm --cgroups=split --network=systemd-quadlet-demo --sdnotify=conmon -d -v systemd-quadlet-demo-mysql:/var/lib/mysql -v /tmp/quadlet_demo:/var/lib/quadlet_demo:Z --secret mysql-root-password-container,type=env,target=MYSQL_ROOT_PASSWORD --health-cmd /usr/bin/true --health-on-failure kill quay.io/linux-system-roles/mysql:5.6 ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStartEx": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman run --name=quadlet-demo-mysql --cidfile=/run/quadlet-demo-mysql.cid --replace --rm --cgroups=split --network=systemd-quadlet-demo --sdnotify=conmon -d -v systemd-quadlet-demo-mysql:/var/lib/mysql -v /tmp/quadlet_demo:/var/lib/quadlet_demo:Z --secret mysql-root-password-container,type=env,target=MYSQL_ROOT_PASSWORD --health-cmd /usr/bin/true --health-on-failure kill quay.io/linux-system-roles/mysql:5.6 ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStop": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman rm -v -f -i --cidfile=/run/quadlet-demo-mysql.cid ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStopEx": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman rm -v -f -i --cidfile=/run/quadlet-demo-mysql.cid ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStopPost": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman rm -v -f -i --cidfile=/run/quadlet-demo-mysql.cid ; ignore_errors=yes ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStopPostEx": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman rm -v -f -i --cidfile=/run/quadlet-demo-mysql.cid ; flags=ignore-failure ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExitType": "main", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FinalKillSignal": "9", "FragmentPath": "/run/systemd/generator/quadlet-demo-mysql.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOReadBytes": "18446744073709551615", "IOReadOperations": "18446744073709551615", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "IOWeight": "[not set]", "IOWriteBytes": "18446744073709551615", "IOWriteOperations": "18446744073709551615", "IPAccounting": "no", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "Id": "quadlet-demo-mysql.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestamp": "Tue 2024-02-06 16:18:47 UTC", "InactiveEnterTimestampMonotonic": "574222533", "InactiveExitTimestamp": "Tue 2024-02-06 16:18:47 UTC", "InactiveExitTimestampMonotonic": "574127296", "InvocationID": "1b15feff71844513a6955a4a9fdfc719", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "mixed", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "0", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "8388608", "LimitMEMLOCKSoft": "8388608", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "524288", "LimitNOFILESoft": "1024", "LimitNPROC": "13974", "LimitNPROCSoft": "13974", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "13974", "LimitSIGPENDINGSoft": "13974", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "0", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "ManagedOOMSwap": "auto", "MemoryAccounting": "yes", "MemoryAvailable": "infinity", "MemoryCurrent": "[not set]", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemorySwapMax": "infinity", "MountAPIVFS": "no", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAPolicy": "n/a", "Names": "quadlet-demo-mysql.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "all", "OOMPolicy": "continue", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "OnSuccessJobMode": "fail", "Perpetual": "no", "PrivateDevices": "no", "PrivateIPC": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProcSubset": "all", "ProtectClock": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectHostname": "no", "ProtectKernelLogs": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectProc": "default", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "ReloadResult": "success", "ReloadSignal": "1", "RemainAfterExit": "no", "RemoveIPC": "no", "Requires": "quadlet-demo-network.service system.slice quadlet-demo-mysql-volume.service sysinit.target -.mount", "RequiresMountsFor": "/run/containers /tmp/quadlet_demo", "Restart": "no", "RestartKillSignal": "15", "RestartUSec": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "exit-code", "RootDirectoryStartOnly": "no", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "Slice": "system.slice", "SourcePath": "/etc/containers/systemd/quadlet-demo-mysql.container", "StandardError": "inherit", "StandardInput": "null", "StandardOutput": "journal", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StateChangeTimestamp": "Tue 2024-02-06 16:18:47 UTC", "StateChangeTimestampMonotonic": "574222533", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "failed", "SuccessAction": "none", "SyslogFacility": "3", "SyslogIdentifier": "quadlet-demo-mysql", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "2147483646", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "[not set]", "TasksMax": "22359", "TimeoutAbortUSec": "1min 30s", "TimeoutCleanUSec": "infinity", "TimeoutStartFailureMode": "terminate", "TimeoutStartUSec": "1min 30s", "TimeoutStopFailureMode": "terminate", "TimeoutStopUSec": "1min 30s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "notify", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "disabled", "UnitFileState": "generated", "UtmpMode": "init", "WantedBy": "multi-user.target", "WatchdogSignal": "6", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "0" } } TASK [fedora.linux_system_roles.podman : Remove quadlet file] ****************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:32 Tuesday 06 February 2024 16:19:01 +0000 (0:00:00.576) 0:01:48.099 ****** changed: [sut] => { "changed": true, "path": "/etc/containers/systemd/quadlet-demo-mysql.container", "state": "absent" } TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:38 Tuesday 06 February 2024 16:19:02 +0000 (0:00:00.214) 0:01:48.314 ****** included: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for sut TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Tuesday 06 February 2024 16:19:02 +0000 (0:00:00.033) 0:01:48.347 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Tuesday 06 February 2024 16:19:02 +0000 (0:00:00.021) 0:01:48.368 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Tuesday 06 February 2024 16:19:02 +0000 (0:00:00.027) 0:01:48.396 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Reload systemctl] ********************* task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:46 Tuesday 06 February 2024 16:19:02 +0000 (0:00:00.020) 0:01:48.416 ****** ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.podman : Prune images no longer in use] ******** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:55 Tuesday 06 February 2024 16:19:02 +0000 (0:00:00.584) 0:01:49.001 ****** changed: [sut] => { "changed": true, "cmd": [ "podman", "image", "prune", "-f" ], "delta": "0:00:00.042810", "end": "2024-02-06 16:19:02.922229", "rc": 0, "start": "2024-02-06 16:19:02.879419" } TASK [fedora.linux_system_roles.podman : Remove network] *********************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:63 Tuesday 06 February 2024 16:19:03 +0000 (0:00:00.301) 0:01:49.302 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_quadlet_type == \"network\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create and update quadlets] *********** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:107 Tuesday 06 February 2024 16:19:03 +0000 (0:00:00.022) 0:01:49.324 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_state != \"absent\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14 Tuesday 06 February 2024 16:19:03 +0000 (0:00:00.020) 0:01:49.345 ****** ok: [sut] => { "ansible_facts": { "__podman_quadlet_file_src": "envoy-proxy-configmap.yml", "__podman_quadlet_spec": {}, "__podman_quadlet_str": "---\napiVersion: v1\nkind: ConfigMap\nmetadata:\n name: envoy-proxy-config\ndata:\n envoy.yaml: |\n admin:\n address:\n socket_address:\n address: 0.0.0.0\n port_value: 9901\n\n static_resources:\n listeners:\n - name: listener_0\n address:\n socket_address:\n address: 0.0.0.0\n port_value: 8080\n filter_chains:\n - filters:\n - name: envoy.filters.network.http_connection_manager\n typed_config:\n \"@type\": type.googleapis.com/envoy.extensions.filters.network.http_connection_manager.v3.HttpConnectionManager\n stat_prefix: ingress_http\n codec_type: AUTO\n route_config:\n name: local_route\n virtual_hosts:\n - name: local_service\n domains: [\"*\"]\n routes:\n - match:\n prefix: \"/\"\n route:\n cluster: backend\n http_filters:\n - name: envoy.filters.http.router\n typed_config:\n \"@type\": type.googleapis.com/envoy.extensions.filters.http.router.v3.Router\n transport_socket:\n name: envoy.transport_sockets.tls\n typed_config:\n \"@type\": type.googleapis.com/envoy.extensions.transport_sockets.tls.v3.DownstreamTlsContext\n common_tls_context:\n tls_certificates:\n - certificate_chain:\n filename: /etc/envoy-certificates/certificate.pem\n private_key:\n filename: /etc/envoy-certificates/certificate.key\n clusters:\n - name: backend\n connect_timeout: 5s\n type: STATIC\n dns_refresh_rate: 1800s\n lb_policy: ROUND_ROBIN\n load_assignment:\n cluster_name: backend\n endpoints:\n - lb_endpoints:\n - endpoint:\n address:\n socket_address:\n address: 127.0.0.1\n port_value: 80", "__podman_quadlet_template_src": "" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25 Tuesday 06 February 2024 16:19:03 +0000 (0:00:00.030) 0:01:49.376 ****** ok: [sut] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_pull_image": true, "__podman_state": "absent", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] ***** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35 Tuesday 06 February 2024 16:19:03 +0000 (0:00:00.026) 0:01:49.403 ****** skipping: [sut] => { "changed": false, "false_condition": "not __podman_quadlet_file_src", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48 Tuesday 06 February 2024 16:19:03 +0000 (0:00:00.021) 0:01:49.424 ****** ok: [sut] => { "ansible_facts": { "__podman_quadlet_name": "envoy-proxy-configmap", "__podman_quadlet_type": "yml", "__podman_rootless": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57 Tuesday 06 February 2024 16:19:03 +0000 (0:00:00.032) 0:01:49.457 ****** included: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for sut TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Tuesday 06 February 2024 16:19:03 +0000 (0:00:00.038) 0:01:49.495 ****** skipping: [sut] => { "changed": false, "false_condition": "'getent_passwd' not in ansible_facts or __podman_user not in ansible_facts['getent_passwd']", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Tuesday 06 February 2024 16:19:03 +0000 (0:00:00.023) 0:01:49.519 ****** skipping: [sut] => { "changed": false, "false_condition": "not ansible_facts[\"getent_passwd\"][__podman_user]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Tuesday 06 February 2024 16:19:03 +0000 (0:00:00.024) 0:01:49.543 ****** ok: [sut] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get group information] **************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:27 Tuesday 06 February 2024 16:19:03 +0000 (0:00:00.028) 0:01:49.572 ****** ok: [sut] => { "ansible_facts": { "getent_group": { "root": [ "x", "0", "" ] } }, "changed": false } TASK [fedora.linux_system_roles.podman : Set group name] *********************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:34 Tuesday 06 February 2024 16:19:03 +0000 (0:00:00.210) 0:01:49.782 ****** ok: [sut] => { "ansible_facts": { "__podman_group_name": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:38 Tuesday 06 February 2024 16:19:03 +0000 (0:00:00.027) 0:01:49.810 ****** ok: [sut] => { "changed": false, "stat": { "atime": 1707236301.2420683, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "076f89863b474504550ab1b5c11216a77ed7123c", "ctime": 1707236269.7653096, "dev": 51713, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 4471560, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-pie-executable", "mode": "0755", "mtime": 1689166618.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 15496, "uid": 0, "version": "396954581", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************ task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:49 Tuesday 06 February 2024 16:19:03 +0000 (0:00:00.202) 0:01:50.013 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check group with getsubids] *********** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:53 Tuesday 06 February 2024 16:19:03 +0000 (0:00:00.020) 0:01:50.034 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if user is in subuid file] ****** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:63 Tuesday 06 February 2024 16:19:03 +0000 (0:00:00.022) 0:01:50.056 ****** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:71 Tuesday 06 February 2024 16:19:03 +0000 (0:00:00.020) 0:01:50.077 ****** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if group is in subgid file] ***** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:78 Tuesday 06 February 2024 16:19:03 +0000 (0:00:00.020) 0:01:50.097 ****** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] ***** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:86 Tuesday 06 February 2024 16:19:03 +0000 (0:00:00.019) 0:01:50.117 ****** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62 Tuesday 06 February 2024 16:19:03 +0000 (0:00:00.029) 0:01:50.147 ****** ok: [sut] => { "ansible_facts": { "__podman_activate_systemd_unit": true, "__podman_images_found": [], "__podman_kube_yamls_raw": "", "__podman_service_name": "", "__podman_systemd_scope": "system", "__podman_user_home_dir": "/root", "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73 Tuesday 06 February 2024 16:19:03 +0000 (0:00:00.058) 0:01:50.206 ****** ok: [sut] => { "ansible_facts": { "__podman_quadlet_path": "/etc/containers/systemd" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get kube yaml contents] *************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77 Tuesday 06 February 2024 16:19:03 +0000 (0:00:00.088) 0:01:50.295 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_state != \"absent\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87 Tuesday 06 February 2024 16:19:04 +0000 (0:00:00.022) 0:01:50.317 ****** ok: [sut] => { "ansible_facts": { "__podman_images": [], "__podman_quadlet_file": "/etc/containers/systemd/envoy-proxy-configmap.yml", "__podman_volumes": [] }, "changed": false } TASK [fedora.linux_system_roles.podman : Cleanup quadlets] ********************* task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:103 Tuesday 06 February 2024 16:19:04 +0000 (0:00:00.049) 0:01:50.367 ****** included: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml for sut TASK [fedora.linux_system_roles.podman : Stat XDG_RUNTIME_DIR] ***************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:4 Tuesday 06 February 2024 16:19:04 +0000 (0:00:00.043) 0:01:50.410 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Stop and disable service] ************* task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:12 Tuesday 06 February 2024 16:19:04 +0000 (0:00:00.019) 0:01:50.430 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_service_name | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Remove quadlet file] ****************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:32 Tuesday 06 February 2024 16:19:04 +0000 (0:00:00.022) 0:01:50.453 ****** ok: [sut] => { "changed": false, "path": "/etc/containers/systemd/envoy-proxy-configmap.yml", "state": "absent" } TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:38 Tuesday 06 February 2024 16:19:04 +0000 (0:00:00.207) 0:01:50.660 ****** included: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for sut TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Tuesday 06 February 2024 16:19:04 +0000 (0:00:00.030) 0:01:50.690 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Tuesday 06 February 2024 16:19:04 +0000 (0:00:00.019) 0:01:50.709 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Tuesday 06 February 2024 16:19:04 +0000 (0:00:00.020) 0:01:50.730 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Reload systemctl] ********************* task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:46 Tuesday 06 February 2024 16:19:04 +0000 (0:00:00.021) 0:01:50.751 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_file_removed is changed", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Prune images no longer in use] ******** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:55 Tuesday 06 February 2024 16:19:04 +0000 (0:00:00.022) 0:01:50.774 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_file_removed is changed", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Remove network] *********************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:63 Tuesday 06 February 2024 16:19:04 +0000 (0:00:00.019) 0:01:50.794 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_file_removed is changed", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create and update quadlets] *********** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:107 Tuesday 06 February 2024 16:19:04 +0000 (0:00:00.020) 0:01:50.814 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_state != \"absent\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14 Tuesday 06 February 2024 16:19:04 +0000 (0:00:00.018) 0:01:50.833 ****** ok: [sut] => { "ansible_facts": { "__podman_quadlet_file_src": "", "__podman_quadlet_spec": {}, "__podman_quadlet_str": "---\napiVersion: v1\nkind: PersistentVolumeClaim\nmetadata:\n name: wp-pv-claim\n labels:\n app: wordpress\nspec:\n accessModes:\n - ReadWriteOnce\n resources:\n requests:\n storage: 20Gi\n---\napiVersion: v1\nkind: Pod\nmetadata:\n name: quadlet-demo\nspec:\n containers:\n - name: wordpress\n image: quay.io/linux-system-roles/wordpress:4.8-apache\n env:\n - name: WORDPRESS_DB_HOST\n value: quadlet-demo-mysql\n - name: WORDPRESS_DB_PASSWORD\n valueFrom:\n secretKeyRef:\n name: mysql-root-password-kube\n key: password\n volumeMounts:\n - name: wordpress-persistent-storage\n mountPath: /var/www/html\n resources:\n requests:\n memory: \"64Mi\"\n cpu: \"250m\"\n limits:\n memory: \"128Mi\"\n cpu: \"500m\"\n - name: envoy\n image: quay.io/linux-system-roles/envoyproxy:v1.25.0\n volumeMounts:\n - name: config-volume\n mountPath: /etc/envoy\n - name: certificates\n mountPath: /etc/envoy-certificates\n env:\n - name: ENVOY_UID\n value: \"0\"\n resources:\n requests:\n memory: \"64Mi\"\n cpu: \"250m\"\n limits:\n memory: \"128Mi\"\n cpu: \"500m\"\n volumes:\n - name: config-volume\n configMap:\n name: envoy-proxy-config\n - name: certificates\n secret:\n secretName: envoy-certificates\n - name: wordpress-persistent-storage\n persistentVolumeClaim:\n claimName: wp-pv-claim\n - name: www # not used - for testing hostpath\n hostPath:\n path: /tmp/httpd3\n - name: create # not used - for testing hostpath\n hostPath:\n path: /tmp/httpd3-create\n", "__podman_quadlet_template_src": "quadlet-demo.yml.j2" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25 Tuesday 06 February 2024 16:19:04 +0000 (0:00:00.063) 0:01:50.897 ****** ok: [sut] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_pull_image": true, "__podman_state": "absent", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] ***** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35 Tuesday 06 February 2024 16:19:04 +0000 (0:00:00.044) 0:01:50.942 ****** skipping: [sut] => { "changed": false, "false_condition": "not __podman_quadlet_str", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48 Tuesday 06 February 2024 16:19:04 +0000 (0:00:00.027) 0:01:50.969 ****** ok: [sut] => { "ansible_facts": { "__podman_quadlet_name": "quadlet-demo", "__podman_quadlet_type": "yml", "__podman_rootless": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57 Tuesday 06 February 2024 16:19:04 +0000 (0:00:00.059) 0:01:51.029 ****** included: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for sut TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Tuesday 06 February 2024 16:19:04 +0000 (0:00:00.136) 0:01:51.165 ****** skipping: [sut] => { "changed": false, "false_condition": "'getent_passwd' not in ansible_facts or __podman_user not in ansible_facts['getent_passwd']", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Tuesday 06 February 2024 16:19:04 +0000 (0:00:00.025) 0:01:51.191 ****** skipping: [sut] => { "changed": false, "false_condition": "not ansible_facts[\"getent_passwd\"][__podman_user]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Tuesday 06 February 2024 16:19:04 +0000 (0:00:00.022) 0:01:51.214 ****** ok: [sut] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get group information] **************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:27 Tuesday 06 February 2024 16:19:04 +0000 (0:00:00.030) 0:01:51.245 ****** ok: [sut] => { "ansible_facts": { "getent_group": { "root": [ "x", "0", "" ] } }, "changed": false } TASK [fedora.linux_system_roles.podman : Set group name] *********************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:34 Tuesday 06 February 2024 16:19:05 +0000 (0:00:00.206) 0:01:51.452 ****** ok: [sut] => { "ansible_facts": { "__podman_group_name": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:38 Tuesday 06 February 2024 16:19:05 +0000 (0:00:00.030) 0:01:51.482 ****** ok: [sut] => { "changed": false, "stat": { "atime": 1707236301.2420683, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "076f89863b474504550ab1b5c11216a77ed7123c", "ctime": 1707236269.7653096, "dev": 51713, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 4471560, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-pie-executable", "mode": "0755", "mtime": 1689166618.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 15496, "uid": 0, "version": "396954581", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************ task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:49 Tuesday 06 February 2024 16:19:05 +0000 (0:00:00.206) 0:01:51.689 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check group with getsubids] *********** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:53 Tuesday 06 February 2024 16:19:05 +0000 (0:00:00.021) 0:01:51.711 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if user is in subuid file] ****** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:63 Tuesday 06 February 2024 16:19:05 +0000 (0:00:00.020) 0:01:51.731 ****** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:71 Tuesday 06 February 2024 16:19:05 +0000 (0:00:00.023) 0:01:51.755 ****** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if group is in subgid file] ***** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:78 Tuesday 06 February 2024 16:19:05 +0000 (0:00:00.022) 0:01:51.777 ****** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] ***** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:86 Tuesday 06 February 2024 16:19:05 +0000 (0:00:00.022) 0:01:51.799 ****** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62 Tuesday 06 February 2024 16:19:05 +0000 (0:00:00.021) 0:01:51.821 ****** ok: [sut] => { "ansible_facts": { "__podman_activate_systemd_unit": true, "__podman_images_found": [], "__podman_kube_yamls_raw": "", "__podman_service_name": "", "__podman_systemd_scope": "system", "__podman_user_home_dir": "/root", "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73 Tuesday 06 February 2024 16:19:05 +0000 (0:00:00.039) 0:01:51.860 ****** ok: [sut] => { "ansible_facts": { "__podman_quadlet_path": "/etc/containers/systemd" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get kube yaml contents] *************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77 Tuesday 06 February 2024 16:19:05 +0000 (0:00:00.025) 0:01:51.885 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_state != \"absent\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87 Tuesday 06 February 2024 16:19:05 +0000 (0:00:00.023) 0:01:51.909 ****** ok: [sut] => { "ansible_facts": { "__podman_images": [], "__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo.yml", "__podman_volumes": [] }, "changed": false } TASK [fedora.linux_system_roles.podman : Cleanup quadlets] ********************* task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:103 Tuesday 06 February 2024 16:19:05 +0000 (0:00:00.051) 0:01:51.960 ****** included: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml for sut TASK [fedora.linux_system_roles.podman : Stat XDG_RUNTIME_DIR] ***************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:4 Tuesday 06 February 2024 16:19:05 +0000 (0:00:00.042) 0:01:52.003 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Stop and disable service] ************* task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:12 Tuesday 06 February 2024 16:19:05 +0000 (0:00:00.022) 0:01:52.025 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_service_name | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Remove quadlet file] ****************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:32 Tuesday 06 February 2024 16:19:05 +0000 (0:00:00.023) 0:01:52.048 ****** ok: [sut] => { "changed": false, "path": "/etc/containers/systemd/quadlet-demo.yml", "state": "absent" } TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:38 Tuesday 06 February 2024 16:19:06 +0000 (0:00:00.272) 0:01:52.320 ****** included: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for sut TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Tuesday 06 February 2024 16:19:06 +0000 (0:00:00.034) 0:01:52.355 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Tuesday 06 February 2024 16:19:06 +0000 (0:00:00.022) 0:01:52.377 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Tuesday 06 February 2024 16:19:06 +0000 (0:00:00.020) 0:01:52.398 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Reload systemctl] ********************* task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:46 Tuesday 06 February 2024 16:19:06 +0000 (0:00:00.020) 0:01:52.418 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_file_removed is changed", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Prune images no longer in use] ******** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:55 Tuesday 06 February 2024 16:19:06 +0000 (0:00:00.023) 0:01:52.442 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_file_removed is changed", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Remove network] *********************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:63 Tuesday 06 February 2024 16:19:06 +0000 (0:00:00.022) 0:01:52.464 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_file_removed is changed", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create and update quadlets] *********** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:107 Tuesday 06 February 2024 16:19:06 +0000 (0:00:00.022) 0:01:52.486 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_state != \"absent\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14 Tuesday 06 February 2024 16:19:06 +0000 (0:00:00.020) 0:01:52.506 ****** ok: [sut] => { "ansible_facts": { "__podman_quadlet_file_src": "quadlet-demo.kube", "__podman_quadlet_spec": {}, "__podman_quadlet_str": "[Install]\nWantedBy=default.target\n\n[Unit]\nRequires=quadlet-demo-mysql.service\nAfter=quadlet-demo-mysql.service\n\n[Kube]\n# Point to the yaml file in the same directory\nYaml=quadlet-demo.yml\n# Use the quadlet-demo network\nNetwork=quadlet-demo.network\n# Publish the envoy proxy data port\nPublishPort=8000:8080\n# Publish the envoy proxy admin port\nPublishPort=9000:9901\n# Use the envoy proxy config map in the same directory\nConfigMap=envoy-proxy-configmap.yml", "__podman_quadlet_template_src": "" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25 Tuesday 06 February 2024 16:19:06 +0000 (0:00:00.032) 0:01:52.539 ****** ok: [sut] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_pull_image": true, "__podman_state": "absent", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] ***** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35 Tuesday 06 February 2024 16:19:06 +0000 (0:00:00.027) 0:01:52.566 ****** skipping: [sut] => { "changed": false, "false_condition": "not __podman_quadlet_file_src", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48 Tuesday 06 February 2024 16:19:06 +0000 (0:00:00.022) 0:01:52.588 ****** ok: [sut] => { "ansible_facts": { "__podman_quadlet_name": "quadlet-demo", "__podman_quadlet_type": "kube", "__podman_rootless": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57 Tuesday 06 February 2024 16:19:06 +0000 (0:00:00.032) 0:01:52.621 ****** included: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for sut TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Tuesday 06 February 2024 16:19:06 +0000 (0:00:00.040) 0:01:52.662 ****** skipping: [sut] => { "changed": false, "false_condition": "'getent_passwd' not in ansible_facts or __podman_user not in ansible_facts['getent_passwd']", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Tuesday 06 February 2024 16:19:06 +0000 (0:00:00.025) 0:01:52.687 ****** skipping: [sut] => { "changed": false, "false_condition": "not ansible_facts[\"getent_passwd\"][__podman_user]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Tuesday 06 February 2024 16:19:06 +0000 (0:00:00.023) 0:01:52.711 ****** ok: [sut] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get group information] **************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:27 Tuesday 06 February 2024 16:19:06 +0000 (0:00:00.030) 0:01:52.741 ****** ok: [sut] => { "ansible_facts": { "getent_group": { "root": [ "x", "0", "" ] } }, "changed": false } TASK [fedora.linux_system_roles.podman : Set group name] *********************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:34 Tuesday 06 February 2024 16:19:06 +0000 (0:00:00.208) 0:01:52.950 ****** ok: [sut] => { "ansible_facts": { "__podman_group_name": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:38 Tuesday 06 February 2024 16:19:06 +0000 (0:00:00.030) 0:01:52.980 ****** ok: [sut] => { "changed": false, "stat": { "atime": 1707236301.2420683, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "076f89863b474504550ab1b5c11216a77ed7123c", "ctime": 1707236269.7653096, "dev": 51713, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 4471560, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-pie-executable", "mode": "0755", "mtime": 1689166618.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 15496, "uid": 0, "version": "396954581", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************ task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:49 Tuesday 06 February 2024 16:19:06 +0000 (0:00:00.209) 0:01:53.189 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check group with getsubids] *********** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:53 Tuesday 06 February 2024 16:19:06 +0000 (0:00:00.084) 0:01:53.274 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if user is in subuid file] ****** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:63 Tuesday 06 February 2024 16:19:06 +0000 (0:00:00.021) 0:01:53.296 ****** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:71 Tuesday 06 February 2024 16:19:07 +0000 (0:00:00.022) 0:01:53.319 ****** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if group is in subgid file] ***** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:78 Tuesday 06 February 2024 16:19:07 +0000 (0:00:00.020) 0:01:53.339 ****** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] ***** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:86 Tuesday 06 February 2024 16:19:07 +0000 (0:00:00.020) 0:01:53.360 ****** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62 Tuesday 06 February 2024 16:19:07 +0000 (0:00:00.020) 0:01:53.381 ****** ok: [sut] => { "ansible_facts": { "__podman_activate_systemd_unit": true, "__podman_images_found": [], "__podman_kube_yamls_raw": [ "quadlet-demo.yml" ], "__podman_service_name": "quadlet-demo.service", "__podman_systemd_scope": "system", "__podman_user_home_dir": "/root", "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73 Tuesday 06 February 2024 16:19:07 +0000 (0:00:00.041) 0:01:53.422 ****** ok: [sut] => { "ansible_facts": { "__podman_quadlet_path": "/etc/containers/systemd" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get kube yaml contents] *************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77 Tuesday 06 February 2024 16:19:07 +0000 (0:00:00.022) 0:01:53.445 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_state != \"absent\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87 Tuesday 06 February 2024 16:19:07 +0000 (0:00:00.023) 0:01:53.468 ****** ok: [sut] => { "ansible_facts": { "__podman_images": [], "__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo.kube", "__podman_volumes": [] }, "changed": false } TASK [fedora.linux_system_roles.podman : Cleanup quadlets] ********************* task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:103 Tuesday 06 February 2024 16:19:07 +0000 (0:00:00.048) 0:01:53.517 ****** included: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml for sut TASK [fedora.linux_system_roles.podman : Stat XDG_RUNTIME_DIR] ***************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:4 Tuesday 06 February 2024 16:19:07 +0000 (0:00:00.042) 0:01:53.560 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Stop and disable service] ************* task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:12 Tuesday 06 February 2024 16:19:07 +0000 (0:00:00.022) 0:01:53.582 ****** ok: [sut] => { "changed": false, "failed_when_result": false } MSG: Could not find the requested service quadlet-demo.service: host TASK [fedora.linux_system_roles.podman : Remove quadlet file] ****************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:32 Tuesday 06 February 2024 16:19:07 +0000 (0:00:00.355) 0:01:53.938 ****** ok: [sut] => { "changed": false, "path": "/etc/containers/systemd/quadlet-demo.kube", "state": "absent" } TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:38 Tuesday 06 February 2024 16:19:07 +0000 (0:00:00.212) 0:01:54.150 ****** included: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for sut TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Tuesday 06 February 2024 16:19:07 +0000 (0:00:00.035) 0:01:54.185 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Tuesday 06 February 2024 16:19:07 +0000 (0:00:00.019) 0:01:54.204 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Tuesday 06 February 2024 16:19:07 +0000 (0:00:00.018) 0:01:54.223 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Reload systemctl] ********************* task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:46 Tuesday 06 February 2024 16:19:07 +0000 (0:00:00.020) 0:01:54.243 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_file_removed is changed", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Prune images no longer in use] ******** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:55 Tuesday 06 February 2024 16:19:07 +0000 (0:00:00.021) 0:01:54.265 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_file_removed is changed", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Remove network] *********************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:63 Tuesday 06 February 2024 16:19:07 +0000 (0:00:00.021) 0:01:54.286 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_file_removed is changed", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create and update quadlets] *********** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:107 Tuesday 06 February 2024 16:19:08 +0000 (0:00:00.078) 0:01:54.364 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_state != \"absent\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14 Tuesday 06 February 2024 16:19:08 +0000 (0:00:00.020) 0:01:54.385 ****** ok: [sut] => { "ansible_facts": { "__podman_quadlet_file_src": "quadlet-demo.network", "__podman_quadlet_spec": {}, "__podman_quadlet_str": "[Network]\nSubnet=192.168.30.0/24\nGateway=192.168.30.1\nLabel=app=wordpress", "__podman_quadlet_template_src": "" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25 Tuesday 06 February 2024 16:19:08 +0000 (0:00:00.031) 0:01:54.417 ****** ok: [sut] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_pull_image": true, "__podman_state": "absent", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] ***** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35 Tuesday 06 February 2024 16:19:08 +0000 (0:00:00.030) 0:01:54.447 ****** skipping: [sut] => { "changed": false, "false_condition": "not __podman_quadlet_file_src", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48 Tuesday 06 February 2024 16:19:08 +0000 (0:00:00.022) 0:01:54.469 ****** ok: [sut] => { "ansible_facts": { "__podman_quadlet_name": "quadlet-demo", "__podman_quadlet_type": "network", "__podman_rootless": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57 Tuesday 06 February 2024 16:19:08 +0000 (0:00:00.032) 0:01:54.502 ****** included: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for sut TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Tuesday 06 February 2024 16:19:08 +0000 (0:00:00.038) 0:01:54.540 ****** skipping: [sut] => { "changed": false, "false_condition": "'getent_passwd' not in ansible_facts or __podman_user not in ansible_facts['getent_passwd']", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Tuesday 06 February 2024 16:19:08 +0000 (0:00:00.024) 0:01:54.565 ****** skipping: [sut] => { "changed": false, "false_condition": "not ansible_facts[\"getent_passwd\"][__podman_user]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Tuesday 06 February 2024 16:19:08 +0000 (0:00:00.024) 0:01:54.589 ****** ok: [sut] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get group information] **************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:27 Tuesday 06 February 2024 16:19:08 +0000 (0:00:00.027) 0:01:54.617 ****** ok: [sut] => { "ansible_facts": { "getent_group": { "root": [ "x", "0", "" ] } }, "changed": false } TASK [fedora.linux_system_roles.podman : Set group name] *********************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:34 Tuesday 06 February 2024 16:19:08 +0000 (0:00:00.207) 0:01:54.825 ****** ok: [sut] => { "ansible_facts": { "__podman_group_name": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:38 Tuesday 06 February 2024 16:19:08 +0000 (0:00:00.028) 0:01:54.854 ****** ok: [sut] => { "changed": false, "stat": { "atime": 1707236301.2420683, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "076f89863b474504550ab1b5c11216a77ed7123c", "ctime": 1707236269.7653096, "dev": 51713, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 4471560, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-pie-executable", "mode": "0755", "mtime": 1689166618.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 15496, "uid": 0, "version": "396954581", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************ task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:49 Tuesday 06 February 2024 16:19:08 +0000 (0:00:00.204) 0:01:55.059 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check group with getsubids] *********** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:53 Tuesday 06 February 2024 16:19:08 +0000 (0:00:00.020) 0:01:55.079 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if user is in subuid file] ****** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:63 Tuesday 06 February 2024 16:19:08 +0000 (0:00:00.020) 0:01:55.099 ****** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:71 Tuesday 06 February 2024 16:19:08 +0000 (0:00:00.019) 0:01:55.119 ****** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if group is in subgid file] ***** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:78 Tuesday 06 February 2024 16:19:08 +0000 (0:00:00.019) 0:01:55.139 ****** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] ***** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:86 Tuesday 06 February 2024 16:19:08 +0000 (0:00:00.019) 0:01:55.159 ****** skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62 Tuesday 06 February 2024 16:19:08 +0000 (0:00:00.019) 0:01:55.179 ****** ok: [sut] => { "ansible_facts": { "__podman_activate_systemd_unit": true, "__podman_images_found": [], "__podman_kube_yamls_raw": "", "__podman_service_name": "", "__podman_systemd_scope": "system", "__podman_user_home_dir": "/root", "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73 Tuesday 06 February 2024 16:19:08 +0000 (0:00:00.034) 0:01:55.214 ****** ok: [sut] => { "ansible_facts": { "__podman_quadlet_path": "/etc/containers/systemd" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get kube yaml contents] *************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77 Tuesday 06 February 2024 16:19:08 +0000 (0:00:00.085) 0:01:55.299 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_state != \"absent\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87 Tuesday 06 February 2024 16:19:09 +0000 (0:00:00.019) 0:01:55.318 ****** ok: [sut] => { "ansible_facts": { "__podman_images": [], "__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo.network", "__podman_volumes": [] }, "changed": false } TASK [fedora.linux_system_roles.podman : Cleanup quadlets] ********************* task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:103 Tuesday 06 February 2024 16:19:09 +0000 (0:00:00.050) 0:01:55.369 ****** included: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml for sut TASK [fedora.linux_system_roles.podman : Stat XDG_RUNTIME_DIR] ***************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:4 Tuesday 06 February 2024 16:19:09 +0000 (0:00:00.039) 0:01:55.409 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Stop and disable service] ************* task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:12 Tuesday 06 February 2024 16:19:09 +0000 (0:00:00.018) 0:01:55.428 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_service_name | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Remove quadlet file] ****************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:32 Tuesday 06 February 2024 16:19:09 +0000 (0:00:00.022) 0:01:55.450 ****** changed: [sut] => { "changed": true, "path": "/etc/containers/systemd/quadlet-demo.network", "state": "absent" } TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:38 Tuesday 06 February 2024 16:19:09 +0000 (0:00:00.210) 0:01:55.661 ****** included: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for sut TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Tuesday 06 February 2024 16:19:09 +0000 (0:00:00.033) 0:01:55.695 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Tuesday 06 February 2024 16:19:09 +0000 (0:00:00.019) 0:01:55.714 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Tuesday 06 February 2024 16:19:09 +0000 (0:00:00.019) 0:01:55.734 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Reload systemctl] ********************* task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:46 Tuesday 06 February 2024 16:19:09 +0000 (0:00:00.021) 0:01:55.755 ****** ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.podman : Prune images no longer in use] ******** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:55 Tuesday 06 February 2024 16:19:09 +0000 (0:00:00.523) 0:01:56.279 ****** changed: [sut] => { "changed": true, "cmd": [ "podman", "image", "prune", "-f" ], "delta": "0:00:00.038114", "end": "2024-02-06 16:19:10.192578", "rc": 0, "start": "2024-02-06 16:19:10.154464" } TASK [fedora.linux_system_roles.podman : Remove network] *********************** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:63 Tuesday 06 February 2024 16:19:10 +0000 (0:00:00.244) 0:01:56.524 ****** changed: [sut] => { "changed": true, "cmd": [ "podman", "network", "rm", "systemd-quadlet-demo" ], "delta": "0:00:00.038706", "end": "2024-02-06 16:19:10.441252", "rc": 0, "start": "2024-02-06 16:19:10.402546" } STDOUT: systemd-quadlet-demo TASK [fedora.linux_system_roles.podman : Create and update quadlets] *********** task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:107 Tuesday 06 February 2024 16:19:10 +0000 (0:00:00.248) 0:01:56.772 ****** skipping: [sut] => { "changed": false, "false_condition": "__podman_state != \"absent\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Cancel linger] ************************ task path: /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:122 Tuesday 06 February 2024 16:19:10 +0000 (0:00:00.021) 0:01:56.793 ****** skipping: [sut] => { "changed": false, "skipped_reason": "No items in the list" } to retry, use: --limit @/tmp/tests_quadlet_demo.retry PLAY RECAP ********************************************************************* sut : ok=269 changed=31 unreachable=0 failed=1 skipped=261 rescued=1 ignored=0 Tuesday 06 February 2024 16:19:10 +0000 (0:00:00.019) 0:01:56.812 ****** =============================================================================== fedora.linux_system_roles.podman : Ensure required packages are installed -- 50.67s /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:14 fedora.linux_system_roles.podman : Ensure container images are present --- 8.51s /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:20 fedora.linux_system_roles.certificate : Ensure certificate role dependencies are installed --- 5.94s /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:5 fedora.linux_system_roles.certificate : Ensure provider packages are installed --- 3.32s /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:23 fedora.linux_system_roles.firewall : Configure firewall ----------------- 1.72s /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:71 fedora.linux_system_roles.firewall : Enable and start firewalld service --- 1.72s /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:28 fedora.linux_system_roles.podman : Gather the package facts ------------- 1.14s /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:6 fedora.linux_system_roles.podman : Gather the package facts ------------- 0.97s /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:6 fedora.linux_system_roles.certificate : Ensure certificate requests ----- 0.93s /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:101 fedora.linux_system_roles.firewall : Install firewalld ------------------ 0.92s /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:19 fedora.linux_system_roles.firewall : Install firewalld ------------------ 0.91s /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:19 Gathering Facts --------------------------------------------------------- 0.84s /WORKDIR/git-fix-linger-handlingih6nvedg/tests/tests_quadlet_demo.yml:9 ------- fedora.linux_system_roles.certificate : Ensure provider service is running --- 0.76s /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:90 fedora.linux_system_roles.firewall : Configure firewall ----------------- 0.72s /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:71 fedora.linux_system_roles.certificate : Slurp the contents of the files --- 0.62s /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:152 fedora.linux_system_roles.certificate : Remove files -------------------- 0.59s /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:181 fedora.linux_system_roles.podman : Reload systemctl --------------------- 0.58s /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:46 fedora.linux_system_roles.podman : Ensure quadlet file is copied -------- 0.58s /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:46 fedora.linux_system_roles.podman : Stop and disable service ------------- 0.58s /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:12 fedora.linux_system_roles.podman : Stop and disable service ------------- 0.57s /WORKDIR/git-fix-linger-handlingih6nvedg/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:12 ---^---^---^---^---^--- # STDERR: ---v---v---v---v---v--- [DEPRECATION WARNING]: ANSIBLE_COLLECTIONS_PATHS option, does not fit var naming standard, use the singular form ANSIBLE_COLLECTIONS_PATH instead. This feature will be removed from ansible-core in version 2.19. Deprecation warnings can be disabled by setting deprecation_warnings=False in ansible.cfg. [WARNING]: Using a variable for a task's 'args' is unsafe in some situations (see https://docs.ansible.com/ansible/devel/reference_appendices/faq.html#argsplat- unsafe) ---^---^---^---^---^---