# STDOUT: ---v---v---v---v---v--- ansible-playbook 2.9.27 config file = /etc/ansible/ansible.cfg configured module search path = ['/home/jenkins/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules'] ansible python module location = /opt/ansible-2.9/lib/python3.9/site-packages/ansible executable location = /opt/ansible-2.9/bin/ansible-playbook python version = 3.9.18 (main, Sep 7 2023, 00:00:00) [GCC 11.4.1 20230605 (Red Hat 11.4.1-2)] Using /etc/ansible/ansible.cfg as config file Skipping callback 'actionable', as we already have a stdout callback. Skipping callback 'counter_enabled', as we already have a stdout callback. Skipping callback 'debug', as we already have a stdout callback. Skipping callback 'dense', as we already have a stdout callback. Skipping callback 'dense', as we already have a stdout callback. Skipping callback 'full_skip', as we already have a stdout callback. Skipping callback 'json', as we already have a stdout callback. Skipping callback 'minimal', as we already have a stdout callback. Skipping callback 'null', as we already have a stdout callback. Skipping callback 'oneline', as we already have a stdout callback. Skipping callback 'selective', as we already have a stdout callback. Skipping callback 'skippy', as we already have a stdout callback. Skipping callback 'stderr', as we already have a stdout callback. Skipping callback 'unixy', as we already have a stdout callback. Skipping callback 'yaml', as we already have a stdout callback. PLAYBOOK: tests_quadlet_demo.yml *********************************************** 2 plays in /WORKDIR/git-fix-linger-handling7nz_rqha/tests/tests_quadlet_demo.yml PLAY [all] ********************************************************************* META: ran handlers TASK [Include vault variables] ************************************************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/tests_quadlet_demo.yml:4 Sunday 04 February 2024 20:05:45 +0000 (0:00:00.016) 0:00:00.016 ******* ok: [sut] => { "ansible_facts": { "mysql_container_root_password": { "__ansible_vault": "$ANSIBLE_VAULT;1.1;AES256\n61333932373230333539663035366431326163363166363036323963623131363530326231303634\n6635326161643165363366323062333334363730376631660a393566366139353861656364656661\n38653463363837336639363032646433666361646535366137303464623261313663643336306465\n6264663730656337310a343962353137386238383064646533366433333437303566656433386233\n34343235326665646661623131643335313236313131353661386338343366316261643634653633\n3832313034366536616531323963333234326461353130303532\n" } }, "ansible_included_var_files": [ "/WORKDIR/git-fix-linger-handling7nz_rqha/tests/vars/vault-variables.yml" ], "changed": false } META: ran handlers META: ran handlers PLAY [Deploy the quadlet demo app] ********************************************* TASK [Gathering Facts] ********************************************************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/tests_quadlet_demo.yml:9 Sunday 04 February 2024 20:05:45 +0000 (0:00:00.015) 0:00:00.032 ******* ok: [sut] META: ran handlers TASK [Generate certificates] *************************************************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/tests_quadlet_demo.yml:39 Sunday 04 February 2024 20:05:46 +0000 (0:00:00.695) 0:00:00.727 ******* TASK [fedora.linux_system_roles.certificate : Set version specific variables] *** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:2 Sunday 04 February 2024 20:05:46 +0000 (0:00:00.043) 0:00:00.771 ******* included: /WORKDIR/git-fix-linger-handling7nz_rqha/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/set_vars.yml for sut TASK [fedora.linux_system_roles.certificate : Ensure ansible_facts used by role] *** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/set_vars.yml:2 Sunday 04 February 2024 20:05:46 +0000 (0:00:00.016) 0:00:00.788 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.certificate : Check if system is ostree] ******* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/set_vars.yml:10 Sunday 04 February 2024 20:05:46 +0000 (0:00:00.021) 0:00:00.810 ******* ok: [sut] => { "changed": false, "stat": { "exists": false } } TASK [fedora.linux_system_roles.certificate : Set flag to indicate system is ostree] *** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/set_vars.yml:15 Sunday 04 February 2024 20:05:46 +0000 (0:00:00.280) 0:00:01.090 ******* ok: [sut] => { "ansible_facts": { "__certificate_is_ostree": false }, "changed": false } TASK [fedora.linux_system_roles.certificate : Set platform/version specific variables] *** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/set_vars.yml:19 Sunday 04 February 2024 20:05:46 +0000 (0:00:00.025) 0:00:01.115 ******* skipping: [sut] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=CentOS.yml) => { "ansible_loop_var": "item", "changed": false, "item": "CentOS.yml", "skip_reason": "Conditional result was False" } ok: [sut] => (item=CentOS_9.yml) => { "ansible_facts": { "__certificate_certmonger_packages": [ "certmonger", "python3-packaging" ] }, "ansible_included_var_files": [ "/WORKDIR/git-fix-linger-handling7nz_rqha/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/vars/CentOS_9.yml" ], "ansible_loop_var": "item", "changed": false, "item": "CentOS_9.yml" } ok: [sut] => (item=CentOS_9.yml) => { "ansible_facts": { "__certificate_certmonger_packages": [ "certmonger", "python3-packaging" ] }, "ansible_included_var_files": [ "/WORKDIR/git-fix-linger-handling7nz_rqha/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/vars/CentOS_9.yml" ], "ansible_loop_var": "item", "changed": false, "item": "CentOS_9.yml" } TASK [fedora.linux_system_roles.certificate : Ensure certificate role dependencies are installed] *** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:5 Sunday 04 February 2024 20:05:46 +0000 (0:00:00.041) 0:00:01.157 ******* changed: [sut] => { "changed": true, "rc": 0, "results": [ "Installed: python3-pyasn1-0.4.8-6.el9.noarch", "Installed: python3-cffi-1.14.5-5.el9.x86_64", "Installed: python3-ply-3.11-14.el9.noarch", "Installed: python3-pycparser-2.20-6.el9.noarch", "Installed: python3-cryptography-36.0.1-4.el9.x86_64" ] } TASK [fedora.linux_system_roles.certificate : Ensure provider packages are installed] *** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:23 Sunday 04 February 2024 20:05:52 +0000 (0:00:05.668) 0:00:06.825 ******* changed: [sut] => (item=certmonger) => { "__certificate_provider": "certmonger", "ansible_loop_var": "__certificate_provider", "changed": true, "rc": 0, "results": [ "Installed: certmonger-0.79.17-1.el9.x86_64", "Installed: python3-packaging-20.9-5.el9.noarch", "Installed: python3-pyparsing-2.4.7-9.el9.noarch" ] } TASK [fedora.linux_system_roles.certificate : Ensure pre-scripts hooks directory exists] *** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:35 Sunday 04 February 2024 20:05:55 +0000 (0:00:03.274) 0:00:10.100 ******* changed: [sut] => (item=certmonger) => { "__certificate_provider": "certmonger", "ansible_loop_var": "__certificate_provider", "changed": true, "gid": 0, "group": "root", "mode": "0700", "owner": "root", "path": "/etc/certmonger//pre-scripts", "secontext": "unconfined_u:object_r:etc_t:s0", "size": 6, "state": "directory", "uid": 0 } TASK [fedora.linux_system_roles.certificate : Ensure post-scripts hooks directory exists] *** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:61 Sunday 04 February 2024 20:05:55 +0000 (0:00:00.290) 0:00:10.391 ******* changed: [sut] => (item=certmonger) => { "__certificate_provider": "certmonger", "ansible_loop_var": "__certificate_provider", "changed": true, "gid": 0, "group": "root", "mode": "0700", "owner": "root", "path": "/etc/certmonger//post-scripts", "secontext": "unconfined_u:object_r:etc_t:s0", "size": 6, "state": "directory", "uid": 0 } TASK [fedora.linux_system_roles.certificate : Ensure provider service is running] *** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:90 Sunday 04 February 2024 20:05:56 +0000 (0:00:00.210) 0:00:10.601 ******* changed: [sut] => (item=certmonger) => { "__certificate_provider": "certmonger", "ansible_loop_var": "__certificate_provider", "changed": true, "enabled": true, "name": "certmonger", "state": "started", "status": { "AccessSELinuxContext": "system_u:object_r:certmonger_unit_file_t:s0", "ActiveEnterTimestampMonotonic": "0", "ActiveExitTimestampMonotonic": "0", "ActiveState": "inactive", "After": "dbus-broker.service basic.target dbus.socket system.slice network.target syslog.target sysinit.target systemd-journald.socket", "AllowIsolate": "no", "AssertResult": "no", "AssertTimestampMonotonic": "0", "Before": "shutdown.target", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "BusName": "org.fedorahosted.certmonger", "CPUAccounting": "yes", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "[not set]", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "no", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore", "CleanResult": "success", "CollectMode": "inactive", "ConditionResult": "no", "ConditionTimestampMonotonic": "0", "ConfigurationDirectoryMode": "0755", "Conflicts": "shutdown.target", "ControlGroupId": "0", "ControlPID": "0", "CoredumpFilter": "0x33", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "Delegate": "no", "Description": "Certificate monitoring and PKI enrollment", "DevicePolicy": "auto", "DynamicUser": "no", "EnvironmentFiles": "/etc/sysconfig/certmonger (ignore_errors=yes)", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "0", "ExecMainStartTimestampMonotonic": "0", "ExecMainStatus": "0", "ExecStart": "{ path=/usr/sbin/certmonger ; argv[]=/usr/sbin/certmonger -S -p /run/certmonger.pid -n $OPTS ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStartEx": "{ path=/usr/sbin/certmonger ; argv[]=/usr/sbin/certmonger -S -p /run/certmonger.pid -n $OPTS ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExitType": "main", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FinalKillSignal": "9", "FragmentPath": "/usr/lib/systemd/system/certmonger.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOReadBytes": "18446744073709551615", "IOReadOperations": "18446744073709551615", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "IOWeight": "[not set]", "IOWriteBytes": "18446744073709551615", "IOWriteOperations": "18446744073709551615", "IPAccounting": "no", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "Id": "certmonger.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestampMonotonic": "0", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "control-group", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "0", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "8388608", "LimitMEMLOCKSoft": "8388608", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "524288", "LimitNOFILESoft": "1024", "LimitNPROC": "13974", "LimitNPROCSoft": "13974", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "13974", "LimitSIGPENDINGSoft": "13974", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "0", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "ManagedOOMSwap": "auto", "MemoryAccounting": "yes", "MemoryAvailable": "infinity", "MemoryCurrent": "[not set]", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemorySwapMax": "infinity", "MountAPIVFS": "no", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAPolicy": "n/a", "Names": "certmonger.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMPolicy": "stop", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "OnSuccessJobMode": "fail", "PIDFile": "/run/certmonger.pid", "PartOf": "dbus-broker.service", "Perpetual": "no", "PrivateDevices": "no", "PrivateIPC": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProcSubset": "all", "ProtectClock": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectHostname": "no", "ProtectKernelLogs": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectProc": "default", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "ReloadResult": "success", "ReloadSignal": "1", "RemainAfterExit": "no", "RemoveIPC": "no", "Requires": "system.slice sysinit.target dbus.socket", "Restart": "no", "RestartKillSignal": "15", "RestartUSec": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "Slice": "system.slice", "StandardError": "inherit", "StandardInput": "null", "StandardOutput": "journal", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StateChangeTimestampMonotonic": "0", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "dead", "SuccessAction": "none", "SyslogFacility": "3", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "2147483646", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "[not set]", "TasksMax": "22359", "TimeoutAbortUSec": "1min 30s", "TimeoutCleanUSec": "infinity", "TimeoutStartFailureMode": "terminate", "TimeoutStartUSec": "1min 30s", "TimeoutStopFailureMode": "terminate", "TimeoutStopUSec": "1min 30s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "dbus", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "disabled", "UnitFileState": "disabled", "UtmpMode": "init", "WatchdogSignal": "6", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "infinity" } } TASK [fedora.linux_system_roles.certificate : Ensure certificate requests] ***** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:101 Sunday 04 February 2024 20:05:56 +0000 (0:00:00.738) 0:00:11.340 ******* changed: [sut] => (item={'name': 'quadlet_demo', 'dns': ['localhost'], 'ca': 'self-sign'}) => { "ansible_loop_var": "item", "changed": true, "item": { "ca": "self-sign", "dns": [ "localhost" ], "name": "quadlet_demo" } } MSG: Certificate requested (new). TASK [fedora.linux_system_roles.certificate : Slurp the contents of the files] *** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:152 Sunday 04 February 2024 20:05:57 +0000 (0:00:00.852) 0:00:12.192 ******* ok: [sut] => (item=['cert', {'name': 'quadlet_demo', 'dns': ['localhost'], 'ca': 'self-sign'}]) => { "ansible_loop_var": "item", "changed": false, "content": "LS0tLS1CRUdJTiBDRVJUSUZJQ0FURS0tLS0tCk1JSURnakNDQW1xZ0F3SUJBZ0lRUDRFQzJ0cXJUbjZuaW11SDNRTi9EakFOQmdrcWhraUc5dzBCQVFzRkFEQlEKTVNBd0hnWURWUVFEREJkTWIyTmhiQ0JUYVdkdWFXNW5JRUYxZEdodmNtbDBlVEVzTUNvR0ExVUVBd3dqTTJZNApNVEF5WkdFdFpHRmhZalJsTjJVdFlUYzRZVFppT0RjdFpHUXdNemRtTUdRd0hoY05NalF3TWpBME1qQXdOVFUzCldoY05NalV3TWpBME1qQXdOVFUzV2pBVU1SSXdFQVlEVlFRREV3bHNiMk5oYkdodmMzUXdnZ0VpTUEwR0NTcUcKU0liM0RRRUJBUVVBQTRJQkR3QXdnZ0VLQW9JQkFRQ3ZDUnV0MHFQRUUvUkpsSS9ya2wxdDJjaFBSMU5VZkd0KwpCZDFQQ0o1R0hDUkV2eXpGQlV3N0daYkljQXYzcXFDOGZVdldPZmVlbU9FS1RGWldidFNlc2dZd3pQb1E3YlhXCkhJdDZheXVuc3ZiMzBaRTQ1ZXA3VEt0aTlXN0RYQVVZRVNZSjY5ODZjSmtTN3VGSzloWnB6M2oxVEI1QTFSTzEKcUZGWVFnclFsamxQTDdqRFpRSUsvVzQ0bjBUelNlM2g0NkFCSGlEdHFMVWNVeFErUDBvdnNReDFXcm5Ia1hqbwp3NGFaSkUxdnVLOTYrbVVyN3ZESVlpekM5SE9scHN0TDBOUzhmdkl0U3ZtNlFtVHpjZlMwK0h2M3RVeDhDaXMzCmpJbGRuV1ZOSjNocXhXT0NNeDkyam1ITS96eExYMTFNbWxNT1p3dlNaenZrbDF6b2JlWUhBZ01CQUFHamdaTXcKZ1pBd0N3WURWUjBQQkFRREFnV2dNQlFHQTFVZEVRUU5NQXVDQ1d4dlkyRnNhRzl6ZERBZEJnTlZIU1VFRmpBVQpCZ2dyQmdFRkJRY0RBUVlJS3dZQkJRVUhBd0l3REFZRFZSMFRBUUgvQkFJd0FEQWRCZ05WSFE0RUZnUVVGN2ZiClphRUd3VlRhV1ZNcHhyWmlQZllLdi9Bd0h3WURWUjBqQkJnd0ZvQVV1Q0tHZDR4TGNTSTQ4cWZYTUg2WFY0RTkKN0Y0d0RRWUpLb1pJaHZjTkFRRUxCUUFEZ2dFQkFHQ1RVMWVkb2ZIRHhkcnlURmVQWDQzV25tQ3dWeVhTUFlWNApvYzkxL1RSMmtVZFd0Tk9CazYzTkI1eHhxV1EwRjNpRU0zSk9mRUNKOEJLNzJsVFJWOFhHUU9xZWVjMjNtVnJvCkUzcmdDd0l4VEc2OFFPaWwrUHpNT3JPeVJ2QWZDWXV4TTZCVWt6YXRpcXRyRzY1ZmpZTWNNZFljMGV3QzhnZk0KcUFjanJHYkRYZ1Y1QWZ5VHdkYldLMjh5czZnL2N4R1pWamUyOGUvaUVtbzFUY21nNjU0TTltd3FsMldmSUNtdApoc010ckFyT0I0dG1pbWp0anF2clV6bXgvczNTTTJSWkNGL1BaN3UwcGdPSTFnZ2M5YTQ2RFpKV1JQUUVTTUdtCmswZ3NoOXdZOFhOcWRueTVadmwrK2g5ZXJ2cnFPTWkyQTFsWmtCNXBNNXZiZjJ2SGlSOD0KLS0tLS1FTkQgQ0VSVElGSUNBVEUtLS0tLQo=", "encoding": "base64", "item": [ "cert", { "ca": "self-sign", "dns": [ "localhost" ], "name": "quadlet_demo" } ], "source": "/etc/pki/tls/certs/quadlet_demo.crt" } ok: [sut] => (item=['key', {'name': 'quadlet_demo', 'dns': ['localhost'], 'ca': 'self-sign'}]) => { "ansible_loop_var": "item", "changed": false, "content": "LS0tLS1CRUdJTiBQUklWQVRFIEtFWS0tLS0tCk1JSUV2UUlCQURBTkJna3Foa2lHOXcwQkFRRUZBQVNDQktjd2dnU2pBZ0VBQW9JQkFRQ3ZDUnV0MHFQRUUvUkoKbEkvcmtsMXQyY2hQUjFOVWZHdCtCZDFQQ0o1R0hDUkV2eXpGQlV3N0daYkljQXYzcXFDOGZVdldPZmVlbU9FSwpURlpXYnRTZXNnWXd6UG9RN2JYV0hJdDZheXVuc3ZiMzBaRTQ1ZXA3VEt0aTlXN0RYQVVZRVNZSjY5ODZjSmtTCjd1Rks5aFpwejNqMVRCNUExUk8xcUZGWVFnclFsamxQTDdqRFpRSUsvVzQ0bjBUelNlM2g0NkFCSGlEdHFMVWMKVXhRK1Awb3ZzUXgxV3JuSGtYam93NGFaSkUxdnVLOTYrbVVyN3ZESVlpekM5SE9scHN0TDBOUzhmdkl0U3ZtNgpRbVR6Y2ZTMCtIdjN0VXg4Q2lzM2pJbGRuV1ZOSjNocXhXT0NNeDkyam1ITS96eExYMTFNbWxNT1p3dlNaenZrCmwxem9iZVlIQWdNQkFBRUNnZ0VBQkZLYXd1T3hWSzNUZWgrM29DODRySFpVNnBSTjNBZklsYytxUkFIbityNUcKTFFaMW9oWDc2SGU4MEExb1ZrRDZBRDJ0NmtmN05BajlnY1laMmIrWjg2TmM0TWFuSkZUaWlUdWlQQzg2WUVWYwpzUWtWWjR6U0NpSitGdzhOb2JoN2pia2FzZU1IMjhjZEdLa2Fta0hTa3VGalVOZUU1Yi85WVI2Ymg2ejEzR25XCjFrTGsyaWF5YjZxWjFDSkNjOFV2aE5qOG85c0wrK1lJZVhVTnpqWDg1NzFsSEx6czB3dnowb0xOVFhPa3gvVmoKRDhqeXcrSzlvU2hPN09lZFhxZGtKS282a3JPY3BsTEtnSGlYcDh3SnBFQVU1THFVMzMxWlBudzhSWUV2YXVmOApMa09CdjZHcUxlTzlTUEJKbjl4T2JOcWJhcVU5MWlZUFJCWlpRcUh6b1FLQmdRREFHNE9BSlM2MVJXTlRjOUxxCjZXNWt6S1ljSko1VFBlU1g0NVNGR2FpQ1l0RFJIdU5BWFJXVXVNeXI4RFA5NEc0WFFoZ2hKRW5mNWJpYnZSSU0Ka0RISzh6dVUrWnhiT3g5T09mTFdFSWYydUZEeXNrcTVmcHh6YUlqT2IxZU51VHBZWWxYeS9QZE5ISE5aQk1zKwpJMGxUSHR3VWdrbzJyVEQ1THlqWnlDZVBad0tCZ1FEcFFBMTZmVG5FRGZXYmU2RlF4K0RlR0Q2ZnphNUFod2pICno0NWF2eE5tcGE1YUtQeVF3Q3MrcXNPM0hFRUZOQUFEZzU0K2hTY1VpaGlyS3FjZnlRVGozY2lKdDZBdVo3TjgKekowS3dIc3o5UEtsRDJlaHhJZnYvRlYyek5PVFpFRElydUY2bnh1K0xUS25QY3E0VFBOdkpVYWZ5Rm5QZC9PdgpISjdIVFFEd1lRS0JnSDNNaHB4UHZQcHZhYytvTlNJbm9tYlhySjhaUzZSdkQxWjNiVGFMeXRFM0xFOEV6d1BDCjhQTG00UUJtMTJVVXBvTU1nUWlTM3QyRGJSUnZCM2E2cGlDa2p3dkxPUngwNXRjdkUxb2lrU1o3SkdjSlllMSsKZjYwRGN6TGJycFhjQ2R6KzZCR3FuYmNkUGx1NHkrN3c1SDlONGdoOEdXVWNuTWxRbkl1Z2ltZFpBb0dCQUo1Lwp6aWpzSzNMVVlZKzBaUncxbHRLVSt3a3doVGFJRENsckd6c0RHTFhwNUI2VU4zeE5lWlZ1Q096SElFL2UxQnVQCk50c1pORU02ekNTVUtWanFDbzZsU1A1OGJ6YlpkUmhiR21RU0JDWjhUdDVuSjI2LzAvZjZLa0UrVW8rWDBvWU0KZ0dBSStwc3lyM2NDQUYxWXZSdEw0ei8wcjBMU0toOVdmWGJGRnR4QkFvR0Fkb1pCSDRiNWd3eU13MU10Qm5DNQp0cHc5ZHNIeUQycXF1TzdQYUJKUk5CS3FSM1I3U0xEREs3dEdmTFl3UWZMQ0lJeDRabzlNa3lERDZPem1GV25LCkhUUW5OcnV4Vjh6VWNTMmFHMWpVeXdSR0liZWQyNUYzallSNEViU05PdHRkcnpkT212VEduN0xVY1dpUWt4ZEQKZFhJV3puZGlEZlBFdTZDeWN3K1NIanM9Ci0tLS0tRU5EIFBSSVZBVEUgS0VZLS0tLS0K", "encoding": "base64", "item": [ "key", { "ca": "self-sign", "dns": [ "localhost" ], "name": "quadlet_demo" } ], "source": "/etc/pki/tls/private/quadlet_demo.key" } ok: [sut] => (item=['ca', {'name': 'quadlet_demo', 'dns': ['localhost'], 'ca': 'self-sign'}]) => { "ansible_loop_var": "item", "changed": false, "content": "LS0tLS1CRUdJTiBDRVJUSUZJQ0FURS0tLS0tCk1JSURnakNDQW1xZ0F3SUJBZ0lRUDRFQzJ0cXJUbjZuaW11SDNRTi9EakFOQmdrcWhraUc5dzBCQVFzRkFEQlEKTVNBd0hnWURWUVFEREJkTWIyTmhiQ0JUYVdkdWFXNW5JRUYxZEdodmNtbDBlVEVzTUNvR0ExVUVBd3dqTTJZNApNVEF5WkdFdFpHRmhZalJsTjJVdFlUYzRZVFppT0RjdFpHUXdNemRtTUdRd0hoY05NalF3TWpBME1qQXdOVFUzCldoY05NalV3TWpBME1qQXdOVFUzV2pBVU1SSXdFQVlEVlFRREV3bHNiMk5oYkdodmMzUXdnZ0VpTUEwR0NTcUcKU0liM0RRRUJBUVVBQTRJQkR3QXdnZ0VLQW9JQkFRQ3ZDUnV0MHFQRUUvUkpsSS9ya2wxdDJjaFBSMU5VZkd0KwpCZDFQQ0o1R0hDUkV2eXpGQlV3N0daYkljQXYzcXFDOGZVdldPZmVlbU9FS1RGWldidFNlc2dZd3pQb1E3YlhXCkhJdDZheXVuc3ZiMzBaRTQ1ZXA3VEt0aTlXN0RYQVVZRVNZSjY5ODZjSmtTN3VGSzloWnB6M2oxVEI1QTFSTzEKcUZGWVFnclFsamxQTDdqRFpRSUsvVzQ0bjBUelNlM2g0NkFCSGlEdHFMVWNVeFErUDBvdnNReDFXcm5Ia1hqbwp3NGFaSkUxdnVLOTYrbVVyN3ZESVlpekM5SE9scHN0TDBOUzhmdkl0U3ZtNlFtVHpjZlMwK0h2M3RVeDhDaXMzCmpJbGRuV1ZOSjNocXhXT0NNeDkyam1ITS96eExYMTFNbWxNT1p3dlNaenZrbDF6b2JlWUhBZ01CQUFHamdaTXcKZ1pBd0N3WURWUjBQQkFRREFnV2dNQlFHQTFVZEVRUU5NQXVDQ1d4dlkyRnNhRzl6ZERBZEJnTlZIU1VFRmpBVQpCZ2dyQmdFRkJRY0RBUVlJS3dZQkJRVUhBd0l3REFZRFZSMFRBUUgvQkFJd0FEQWRCZ05WSFE0RUZnUVVGN2ZiClphRUd3VlRhV1ZNcHhyWmlQZllLdi9Bd0h3WURWUjBqQkJnd0ZvQVV1Q0tHZDR4TGNTSTQ4cWZYTUg2WFY0RTkKN0Y0d0RRWUpLb1pJaHZjTkFRRUxCUUFEZ2dFQkFHQ1RVMWVkb2ZIRHhkcnlURmVQWDQzV25tQ3dWeVhTUFlWNApvYzkxL1RSMmtVZFd0Tk9CazYzTkI1eHhxV1EwRjNpRU0zSk9mRUNKOEJLNzJsVFJWOFhHUU9xZWVjMjNtVnJvCkUzcmdDd0l4VEc2OFFPaWwrUHpNT3JPeVJ2QWZDWXV4TTZCVWt6YXRpcXRyRzY1ZmpZTWNNZFljMGV3QzhnZk0KcUFjanJHYkRYZ1Y1QWZ5VHdkYldLMjh5czZnL2N4R1pWamUyOGUvaUVtbzFUY21nNjU0TTltd3FsMldmSUNtdApoc010ckFyT0I0dG1pbWp0anF2clV6bXgvczNTTTJSWkNGL1BaN3UwcGdPSTFnZ2M5YTQ2RFpKV1JQUUVTTUdtCmswZ3NoOXdZOFhOcWRueTVadmwrK2g5ZXJ2cnFPTWkyQTFsWmtCNXBNNXZiZjJ2SGlSOD0KLS0tLS1FTkQgQ0VSVElGSUNBVEUtLS0tLQo=", "encoding": "base64", "item": [ "ca", { "ca": "self-sign", "dns": [ "localhost" ], "name": "quadlet_demo" } ], "source": "/etc/pki/tls/certs/quadlet_demo.crt" } TASK [fedora.linux_system_roles.certificate : Create return data] ************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:160 Sunday 04 February 2024 20:05:58 +0000 (0:00:00.613) 0:00:12.806 ******* ok: [sut] => { "ansible_facts": { "certificate_test_certs": { "quadlet_demo": { "ca": "/etc/pki/tls/certs/quadlet_demo.crt", "ca_content": "-----BEGIN CERTIFICATE-----\nMIIDgjCCAmqgAwIBAgIQP4EC2tqrTn6nimuH3QN/DjANBgkqhkiG9w0BAQsFADBQ\nMSAwHgYDVQQDDBdMb2NhbCBTaWduaW5nIEF1dGhvcml0eTEsMCoGA1UEAwwjM2Y4\nMTAyZGEtZGFhYjRlN2UtYTc4YTZiODctZGQwMzdmMGQwHhcNMjQwMjA0MjAwNTU3\nWhcNMjUwMjA0MjAwNTU3WjAUMRIwEAYDVQQDEwlsb2NhbGhvc3QwggEiMA0GCSqG\nSIb3DQEBAQUAA4IBDwAwggEKAoIBAQCvCRut0qPEE/RJlI/rkl1t2chPR1NUfGt+\nBd1PCJ5GHCREvyzFBUw7GZbIcAv3qqC8fUvWOfeemOEKTFZWbtSesgYwzPoQ7bXW\nHIt6ayunsvb30ZE45ep7TKti9W7DXAUYESYJ6986cJkS7uFK9hZpz3j1TB5A1RO1\nqFFYQgrQljlPL7jDZQIK/W44n0TzSe3h46ABHiDtqLUcUxQ+P0ovsQx1WrnHkXjo\nw4aZJE1vuK96+mUr7vDIYizC9HOlpstL0NS8fvItSvm6QmTzcfS0+Hv3tUx8Cis3\njIldnWVNJ3hqxWOCMx92jmHM/zxLX11MmlMOZwvSZzvkl1zobeYHAgMBAAGjgZMw\ngZAwCwYDVR0PBAQDAgWgMBQGA1UdEQQNMAuCCWxvY2FsaG9zdDAdBgNVHSUEFjAU\nBggrBgEFBQcDAQYIKwYBBQUHAwIwDAYDVR0TAQH/BAIwADAdBgNVHQ4EFgQUF7fb\nZaEGwVTaWVMpxrZiPfYKv/AwHwYDVR0jBBgwFoAUuCKGd4xLcSI48qfXMH6XV4E9\n7F4wDQYJKoZIhvcNAQELBQADggEBAGCTU1edofHDxdryTFePX43WnmCwVyXSPYV4\noc91/TR2kUdWtNOBk63NB5xxqWQ0F3iEM3JOfECJ8BK72lTRV8XGQOqeec23mVro\nE3rgCwIxTG68QOil+PzMOrOyRvAfCYuxM6BUkzatiqtrG65fjYMcMdYc0ewC8gfM\nqAcjrGbDXgV5AfyTwdbWK28ys6g/cxGZVje28e/iEmo1Tcmg654M9mwql2WfICmt\nhsMtrArOB4tmimjtjqvrUzmx/s3SM2RZCF/PZ7u0pgOI1ggc9a46DZJWRPQESMGm\nk0gsh9wY8XNqdny5Zvl++h9ervrqOMi2A1lZkB5pM5vbf2vHiR8=\n-----END CERTIFICATE-----\n", "cert": "/etc/pki/tls/certs/quadlet_demo.crt", "cert_content": "-----BEGIN CERTIFICATE-----\nMIIDgjCCAmqgAwIBAgIQP4EC2tqrTn6nimuH3QN/DjANBgkqhkiG9w0BAQsFADBQ\nMSAwHgYDVQQDDBdMb2NhbCBTaWduaW5nIEF1dGhvcml0eTEsMCoGA1UEAwwjM2Y4\nMTAyZGEtZGFhYjRlN2UtYTc4YTZiODctZGQwMzdmMGQwHhcNMjQwMjA0MjAwNTU3\nWhcNMjUwMjA0MjAwNTU3WjAUMRIwEAYDVQQDEwlsb2NhbGhvc3QwggEiMA0GCSqG\nSIb3DQEBAQUAA4IBDwAwggEKAoIBAQCvCRut0qPEE/RJlI/rkl1t2chPR1NUfGt+\nBd1PCJ5GHCREvyzFBUw7GZbIcAv3qqC8fUvWOfeemOEKTFZWbtSesgYwzPoQ7bXW\nHIt6ayunsvb30ZE45ep7TKti9W7DXAUYESYJ6986cJkS7uFK9hZpz3j1TB5A1RO1\nqFFYQgrQljlPL7jDZQIK/W44n0TzSe3h46ABHiDtqLUcUxQ+P0ovsQx1WrnHkXjo\nw4aZJE1vuK96+mUr7vDIYizC9HOlpstL0NS8fvItSvm6QmTzcfS0+Hv3tUx8Cis3\njIldnWVNJ3hqxWOCMx92jmHM/zxLX11MmlMOZwvSZzvkl1zobeYHAgMBAAGjgZMw\ngZAwCwYDVR0PBAQDAgWgMBQGA1UdEQQNMAuCCWxvY2FsaG9zdDAdBgNVHSUEFjAU\nBggrBgEFBQcDAQYIKwYBBQUHAwIwDAYDVR0TAQH/BAIwADAdBgNVHQ4EFgQUF7fb\nZaEGwVTaWVMpxrZiPfYKv/AwHwYDVR0jBBgwFoAUuCKGd4xLcSI48qfXMH6XV4E9\n7F4wDQYJKoZIhvcNAQELBQADggEBAGCTU1edofHDxdryTFePX43WnmCwVyXSPYV4\noc91/TR2kUdWtNOBk63NB5xxqWQ0F3iEM3JOfECJ8BK72lTRV8XGQOqeec23mVro\nE3rgCwIxTG68QOil+PzMOrOyRvAfCYuxM6BUkzatiqtrG65fjYMcMdYc0ewC8gfM\nqAcjrGbDXgV5AfyTwdbWK28ys6g/cxGZVje28e/iEmo1Tcmg654M9mwql2WfICmt\nhsMtrArOB4tmimjtjqvrUzmx/s3SM2RZCF/PZ7u0pgOI1ggc9a46DZJWRPQESMGm\nk0gsh9wY8XNqdny5Zvl++h9ervrqOMi2A1lZkB5pM5vbf2vHiR8=\n-----END CERTIFICATE-----\n", "key": "/etc/pki/tls/private/quadlet_demo.key", "key_content": "-----BEGIN PRIVATE KEY-----\nMIIEvQIBADANBgkqhkiG9w0BAQEFAASCBKcwggSjAgEAAoIBAQCvCRut0qPEE/RJ\nlI/rkl1t2chPR1NUfGt+Bd1PCJ5GHCREvyzFBUw7GZbIcAv3qqC8fUvWOfeemOEK\nTFZWbtSesgYwzPoQ7bXWHIt6ayunsvb30ZE45ep7TKti9W7DXAUYESYJ6986cJkS\n7uFK9hZpz3j1TB5A1RO1qFFYQgrQljlPL7jDZQIK/W44n0TzSe3h46ABHiDtqLUc\nUxQ+P0ovsQx1WrnHkXjow4aZJE1vuK96+mUr7vDIYizC9HOlpstL0NS8fvItSvm6\nQmTzcfS0+Hv3tUx8Cis3jIldnWVNJ3hqxWOCMx92jmHM/zxLX11MmlMOZwvSZzvk\nl1zobeYHAgMBAAECggEABFKawuOxVK3Teh+3oC84rHZU6pRN3AfIlc+qRAHn+r5G\nLQZ1ohX76He80A1oVkD6AD2t6kf7NAj9gcYZ2b+Z86Nc4ManJFTiiTuiPC86YEVc\nsQkVZ4zSCiJ+Fw8Nobh7jbkaseMH28cdGKkamkHSkuFjUNeE5b/9YR6bh6z13GnW\n1kLk2iayb6qZ1CJCc8UvhNj8o9sL++YIeXUNzjX8571lHLzs0wvz0oLNTXOkx/Vj\nD8jyw+K9oShO7OedXqdkJKo6krOcplLKgHiXp8wJpEAU5LqU331ZPnw8RYEvauf8\nLkOBv6GqLeO9SPBJn9xObNqbaqU91iYPRBZZQqHzoQKBgQDAG4OAJS61RWNTc9Lq\n6W5kzKYcJJ5TPeSX45SFGaiCYtDRHuNAXRWUuMyr8DP94G4XQhghJEnf5bibvRIM\nkDHK8zuU+ZxbOx9OOfLWEIf2uFDyskq5fpxzaIjOb1eNuTpYYlXy/PdNHHNZBMs+\nI0lTHtwUgko2rTD5LyjZyCePZwKBgQDpQA16fTnEDfWbe6FQx+DeGD6fza5AhwjH\nz45avxNmpa5aKPyQwCs+qsO3HEEFNAADg54+hScUihirKqcfyQTj3ciJt6AuZ7N8\nzJ0KwHsz9PKlD2ehxIfv/FV2zNOTZEDIruF6nxu+LTKnPcq4TPNvJUafyFnPd/Ov\nHJ7HTQDwYQKBgH3MhpxPvPpvac+oNSInombXrJ8ZS6RvD1Z3bTaLytE3LE8EzwPC\n8PLm4QBm12UUpoMMgQiS3t2DbRRvB3a6piCkjwvLORx05tcvE1oikSZ7JGcJYe1+\nf60DczLbrpXcCdz+6BGqnbcdPlu4y+7w5H9N4gh8GWUcnMlQnIugimdZAoGBAJ5/\nzijsK3LUYY+0ZRw1ltKU+wkwhTaIDClrGzsDGLXp5B6UN3xNeZVuCOzHIE/e1BuP\nNtsZNEM6zCSUKVjqCo6lSP58bzbZdRhbGmQSBCZ8Tt5nJ26/0/f6KkE+Uo+X0oYM\ngGAI+psyr3cCAF1YvRtL4z/0r0LSKh9WfXbFFtxBAoGAdoZBH4b5gwyMw1MtBnC5\ntpw9dsHyD2qquO7PaBJRNBKqR3R7SLDDK7tGfLYwQfLCIIx4Zo9MkyDD6OzmFWnK\nHTQnNruxV8zUcS2aG1jUywRGIbed25F3jYR4EbSNOttdrzdOmvTGn7LUcWiQkxdD\ndXIWzndiDfPEu6Cycw+SHjs=\n-----END PRIVATE KEY-----\n" } } }, "changed": false } TASK [fedora.linux_system_roles.certificate : Stop tracking certificates] ****** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:176 Sunday 04 February 2024 20:05:58 +0000 (0:00:00.027) 0:00:12.834 ******* ok: [sut] => (item={'cert': '/etc/pki/tls/certs/quadlet_demo.crt', 'cert_content': '-----BEGIN CERTIFICATE-----\nMIIDgjCCAmqgAwIBAgIQP4EC2tqrTn6nimuH3QN/DjANBgkqhkiG9w0BAQsFADBQ\nMSAwHgYDVQQDDBdMb2NhbCBTaWduaW5nIEF1dGhvcml0eTEsMCoGA1UEAwwjM2Y4\nMTAyZGEtZGFhYjRlN2UtYTc4YTZiODctZGQwMzdmMGQwHhcNMjQwMjA0MjAwNTU3\nWhcNMjUwMjA0MjAwNTU3WjAUMRIwEAYDVQQDEwlsb2NhbGhvc3QwggEiMA0GCSqG\nSIb3DQEBAQUAA4IBDwAwggEKAoIBAQCvCRut0qPEE/RJlI/rkl1t2chPR1NUfGt+\nBd1PCJ5GHCREvyzFBUw7GZbIcAv3qqC8fUvWOfeemOEKTFZWbtSesgYwzPoQ7bXW\nHIt6ayunsvb30ZE45ep7TKti9W7DXAUYESYJ6986cJkS7uFK9hZpz3j1TB5A1RO1\nqFFYQgrQljlPL7jDZQIK/W44n0TzSe3h46ABHiDtqLUcUxQ+P0ovsQx1WrnHkXjo\nw4aZJE1vuK96+mUr7vDIYizC9HOlpstL0NS8fvItSvm6QmTzcfS0+Hv3tUx8Cis3\njIldnWVNJ3hqxWOCMx92jmHM/zxLX11MmlMOZwvSZzvkl1zobeYHAgMBAAGjgZMw\ngZAwCwYDVR0PBAQDAgWgMBQGA1UdEQQNMAuCCWxvY2FsaG9zdDAdBgNVHSUEFjAU\nBggrBgEFBQcDAQYIKwYBBQUHAwIwDAYDVR0TAQH/BAIwADAdBgNVHQ4EFgQUF7fb\nZaEGwVTaWVMpxrZiPfYKv/AwHwYDVR0jBBgwFoAUuCKGd4xLcSI48qfXMH6XV4E9\n7F4wDQYJKoZIhvcNAQELBQADggEBAGCTU1edofHDxdryTFePX43WnmCwVyXSPYV4\noc91/TR2kUdWtNOBk63NB5xxqWQ0F3iEM3JOfECJ8BK72lTRV8XGQOqeec23mVro\nE3rgCwIxTG68QOil+PzMOrOyRvAfCYuxM6BUkzatiqtrG65fjYMcMdYc0ewC8gfM\nqAcjrGbDXgV5AfyTwdbWK28ys6g/cxGZVje28e/iEmo1Tcmg654M9mwql2WfICmt\nhsMtrArOB4tmimjtjqvrUzmx/s3SM2RZCF/PZ7u0pgOI1ggc9a46DZJWRPQESMGm\nk0gsh9wY8XNqdny5Zvl++h9ervrqOMi2A1lZkB5pM5vbf2vHiR8=\n-----END CERTIFICATE-----\n', 'key': '/etc/pki/tls/private/quadlet_demo.key', 'key_content': '-----BEGIN PRIVATE KEY-----\nMIIEvQIBADANBgkqhkiG9w0BAQEFAASCBKcwggSjAgEAAoIBAQCvCRut0qPEE/RJ\nlI/rkl1t2chPR1NUfGt+Bd1PCJ5GHCREvyzFBUw7GZbIcAv3qqC8fUvWOfeemOEK\nTFZWbtSesgYwzPoQ7bXWHIt6ayunsvb30ZE45ep7TKti9W7DXAUYESYJ6986cJkS\n7uFK9hZpz3j1TB5A1RO1qFFYQgrQljlPL7jDZQIK/W44n0TzSe3h46ABHiDtqLUc\nUxQ+P0ovsQx1WrnHkXjow4aZJE1vuK96+mUr7vDIYizC9HOlpstL0NS8fvItSvm6\nQmTzcfS0+Hv3tUx8Cis3jIldnWVNJ3hqxWOCMx92jmHM/zxLX11MmlMOZwvSZzvk\nl1zobeYHAgMBAAECggEABFKawuOxVK3Teh+3oC84rHZU6pRN3AfIlc+qRAHn+r5G\nLQZ1ohX76He80A1oVkD6AD2t6kf7NAj9gcYZ2b+Z86Nc4ManJFTiiTuiPC86YEVc\nsQkVZ4zSCiJ+Fw8Nobh7jbkaseMH28cdGKkamkHSkuFjUNeE5b/9YR6bh6z13GnW\n1kLk2iayb6qZ1CJCc8UvhNj8o9sL++YIeXUNzjX8571lHLzs0wvz0oLNTXOkx/Vj\nD8jyw+K9oShO7OedXqdkJKo6krOcplLKgHiXp8wJpEAU5LqU331ZPnw8RYEvauf8\nLkOBv6GqLeO9SPBJn9xObNqbaqU91iYPRBZZQqHzoQKBgQDAG4OAJS61RWNTc9Lq\n6W5kzKYcJJ5TPeSX45SFGaiCYtDRHuNAXRWUuMyr8DP94G4XQhghJEnf5bibvRIM\nkDHK8zuU+ZxbOx9OOfLWEIf2uFDyskq5fpxzaIjOb1eNuTpYYlXy/PdNHHNZBMs+\nI0lTHtwUgko2rTD5LyjZyCePZwKBgQDpQA16fTnEDfWbe6FQx+DeGD6fza5AhwjH\nz45avxNmpa5aKPyQwCs+qsO3HEEFNAADg54+hScUihirKqcfyQTj3ciJt6AuZ7N8\nzJ0KwHsz9PKlD2ehxIfv/FV2zNOTZEDIruF6nxu+LTKnPcq4TPNvJUafyFnPd/Ov\nHJ7HTQDwYQKBgH3MhpxPvPpvac+oNSInombXrJ8ZS6RvD1Z3bTaLytE3LE8EzwPC\n8PLm4QBm12UUpoMMgQiS3t2DbRRvB3a6piCkjwvLORx05tcvE1oikSZ7JGcJYe1+\nf60DczLbrpXcCdz+6BGqnbcdPlu4y+7w5H9N4gh8GWUcnMlQnIugimdZAoGBAJ5/\nzijsK3LUYY+0ZRw1ltKU+wkwhTaIDClrGzsDGLXp5B6UN3xNeZVuCOzHIE/e1BuP\nNtsZNEM6zCSUKVjqCo6lSP58bzbZdRhbGmQSBCZ8Tt5nJ26/0/f6KkE+Uo+X0oYM\ngGAI+psyr3cCAF1YvRtL4z/0r0LSKh9WfXbFFtxBAoGAdoZBH4b5gwyMw1MtBnC5\ntpw9dsHyD2qquO7PaBJRNBKqR3R7SLDDK7tGfLYwQfLCIIx4Zo9MkyDD6OzmFWnK\nHTQnNruxV8zUcS2aG1jUywRGIbed25F3jYR4EbSNOttdrzdOmvTGn7LUcWiQkxdD\ndXIWzndiDfPEu6Cycw+SHjs=\n-----END PRIVATE KEY-----\n', 'ca': '/etc/pki/tls/certs/quadlet_demo.crt', 'ca_content': '-----BEGIN CERTIFICATE-----\nMIIDgjCCAmqgAwIBAgIQP4EC2tqrTn6nimuH3QN/DjANBgkqhkiG9w0BAQsFADBQ\nMSAwHgYDVQQDDBdMb2NhbCBTaWduaW5nIEF1dGhvcml0eTEsMCoGA1UEAwwjM2Y4\nMTAyZGEtZGFhYjRlN2UtYTc4YTZiODctZGQwMzdmMGQwHhcNMjQwMjA0MjAwNTU3\nWhcNMjUwMjA0MjAwNTU3WjAUMRIwEAYDVQQDEwlsb2NhbGhvc3QwggEiMA0GCSqG\nSIb3DQEBAQUAA4IBDwAwggEKAoIBAQCvCRut0qPEE/RJlI/rkl1t2chPR1NUfGt+\nBd1PCJ5GHCREvyzFBUw7GZbIcAv3qqC8fUvWOfeemOEKTFZWbtSesgYwzPoQ7bXW\nHIt6ayunsvb30ZE45ep7TKti9W7DXAUYESYJ6986cJkS7uFK9hZpz3j1TB5A1RO1\nqFFYQgrQljlPL7jDZQIK/W44n0TzSe3h46ABHiDtqLUcUxQ+P0ovsQx1WrnHkXjo\nw4aZJE1vuK96+mUr7vDIYizC9HOlpstL0NS8fvItSvm6QmTzcfS0+Hv3tUx8Cis3\njIldnWVNJ3hqxWOCMx92jmHM/zxLX11MmlMOZwvSZzvkl1zobeYHAgMBAAGjgZMw\ngZAwCwYDVR0PBAQDAgWgMBQGA1UdEQQNMAuCCWxvY2FsaG9zdDAdBgNVHSUEFjAU\nBggrBgEFBQcDAQYIKwYBBQUHAwIwDAYDVR0TAQH/BAIwADAdBgNVHQ4EFgQUF7fb\nZaEGwVTaWVMpxrZiPfYKv/AwHwYDVR0jBBgwFoAUuCKGd4xLcSI48qfXMH6XV4E9\n7F4wDQYJKoZIhvcNAQELBQADggEBAGCTU1edofHDxdryTFePX43WnmCwVyXSPYV4\noc91/TR2kUdWtNOBk63NB5xxqWQ0F3iEM3JOfECJ8BK72lTRV8XGQOqeec23mVro\nE3rgCwIxTG68QOil+PzMOrOyRvAfCYuxM6BUkzatiqtrG65fjYMcMdYc0ewC8gfM\nqAcjrGbDXgV5AfyTwdbWK28ys6g/cxGZVje28e/iEmo1Tcmg654M9mwql2WfICmt\nhsMtrArOB4tmimjtjqvrUzmx/s3SM2RZCF/PZ7u0pgOI1ggc9a46DZJWRPQESMGm\nk0gsh9wY8XNqdny5Zvl++h9ervrqOMi2A1lZkB5pM5vbf2vHiR8=\n-----END CERTIFICATE-----\n'}) => { "ansible_loop_var": "item", "changed": false, "cmd": [ "getcert", "stop-tracking", "-f", "/etc/pki/tls/certs/quadlet_demo.crt" ], "delta": "0:00:00.031727", "end": "2024-02-04 20:05:58.622917", "item": { "ca": "/etc/pki/tls/certs/quadlet_demo.crt", "ca_content": "-----BEGIN CERTIFICATE-----\nMIIDgjCCAmqgAwIBAgIQP4EC2tqrTn6nimuH3QN/DjANBgkqhkiG9w0BAQsFADBQ\nMSAwHgYDVQQDDBdMb2NhbCBTaWduaW5nIEF1dGhvcml0eTEsMCoGA1UEAwwjM2Y4\nMTAyZGEtZGFhYjRlN2UtYTc4YTZiODctZGQwMzdmMGQwHhcNMjQwMjA0MjAwNTU3\nWhcNMjUwMjA0MjAwNTU3WjAUMRIwEAYDVQQDEwlsb2NhbGhvc3QwggEiMA0GCSqG\nSIb3DQEBAQUAA4IBDwAwggEKAoIBAQCvCRut0qPEE/RJlI/rkl1t2chPR1NUfGt+\nBd1PCJ5GHCREvyzFBUw7GZbIcAv3qqC8fUvWOfeemOEKTFZWbtSesgYwzPoQ7bXW\nHIt6ayunsvb30ZE45ep7TKti9W7DXAUYESYJ6986cJkS7uFK9hZpz3j1TB5A1RO1\nqFFYQgrQljlPL7jDZQIK/W44n0TzSe3h46ABHiDtqLUcUxQ+P0ovsQx1WrnHkXjo\nw4aZJE1vuK96+mUr7vDIYizC9HOlpstL0NS8fvItSvm6QmTzcfS0+Hv3tUx8Cis3\njIldnWVNJ3hqxWOCMx92jmHM/zxLX11MmlMOZwvSZzvkl1zobeYHAgMBAAGjgZMw\ngZAwCwYDVR0PBAQDAgWgMBQGA1UdEQQNMAuCCWxvY2FsaG9zdDAdBgNVHSUEFjAU\nBggrBgEFBQcDAQYIKwYBBQUHAwIwDAYDVR0TAQH/BAIwADAdBgNVHQ4EFgQUF7fb\nZaEGwVTaWVMpxrZiPfYKv/AwHwYDVR0jBBgwFoAUuCKGd4xLcSI48qfXMH6XV4E9\n7F4wDQYJKoZIhvcNAQELBQADggEBAGCTU1edofHDxdryTFePX43WnmCwVyXSPYV4\noc91/TR2kUdWtNOBk63NB5xxqWQ0F3iEM3JOfECJ8BK72lTRV8XGQOqeec23mVro\nE3rgCwIxTG68QOil+PzMOrOyRvAfCYuxM6BUkzatiqtrG65fjYMcMdYc0ewC8gfM\nqAcjrGbDXgV5AfyTwdbWK28ys6g/cxGZVje28e/iEmo1Tcmg654M9mwql2WfICmt\nhsMtrArOB4tmimjtjqvrUzmx/s3SM2RZCF/PZ7u0pgOI1ggc9a46DZJWRPQESMGm\nk0gsh9wY8XNqdny5Zvl++h9ervrqOMi2A1lZkB5pM5vbf2vHiR8=\n-----END CERTIFICATE-----\n", "cert": "/etc/pki/tls/certs/quadlet_demo.crt", "cert_content": "-----BEGIN CERTIFICATE-----\nMIIDgjCCAmqgAwIBAgIQP4EC2tqrTn6nimuH3QN/DjANBgkqhkiG9w0BAQsFADBQ\nMSAwHgYDVQQDDBdMb2NhbCBTaWduaW5nIEF1dGhvcml0eTEsMCoGA1UEAwwjM2Y4\nMTAyZGEtZGFhYjRlN2UtYTc4YTZiODctZGQwMzdmMGQwHhcNMjQwMjA0MjAwNTU3\nWhcNMjUwMjA0MjAwNTU3WjAUMRIwEAYDVQQDEwlsb2NhbGhvc3QwggEiMA0GCSqG\nSIb3DQEBAQUAA4IBDwAwggEKAoIBAQCvCRut0qPEE/RJlI/rkl1t2chPR1NUfGt+\nBd1PCJ5GHCREvyzFBUw7GZbIcAv3qqC8fUvWOfeemOEKTFZWbtSesgYwzPoQ7bXW\nHIt6ayunsvb30ZE45ep7TKti9W7DXAUYESYJ6986cJkS7uFK9hZpz3j1TB5A1RO1\nqFFYQgrQljlPL7jDZQIK/W44n0TzSe3h46ABHiDtqLUcUxQ+P0ovsQx1WrnHkXjo\nw4aZJE1vuK96+mUr7vDIYizC9HOlpstL0NS8fvItSvm6QmTzcfS0+Hv3tUx8Cis3\njIldnWVNJ3hqxWOCMx92jmHM/zxLX11MmlMOZwvSZzvkl1zobeYHAgMBAAGjgZMw\ngZAwCwYDVR0PBAQDAgWgMBQGA1UdEQQNMAuCCWxvY2FsaG9zdDAdBgNVHSUEFjAU\nBggrBgEFBQcDAQYIKwYBBQUHAwIwDAYDVR0TAQH/BAIwADAdBgNVHQ4EFgQUF7fb\nZaEGwVTaWVMpxrZiPfYKv/AwHwYDVR0jBBgwFoAUuCKGd4xLcSI48qfXMH6XV4E9\n7F4wDQYJKoZIhvcNAQELBQADggEBAGCTU1edofHDxdryTFePX43WnmCwVyXSPYV4\noc91/TR2kUdWtNOBk63NB5xxqWQ0F3iEM3JOfECJ8BK72lTRV8XGQOqeec23mVro\nE3rgCwIxTG68QOil+PzMOrOyRvAfCYuxM6BUkzatiqtrG65fjYMcMdYc0ewC8gfM\nqAcjrGbDXgV5AfyTwdbWK28ys6g/cxGZVje28e/iEmo1Tcmg654M9mwql2WfICmt\nhsMtrArOB4tmimjtjqvrUzmx/s3SM2RZCF/PZ7u0pgOI1ggc9a46DZJWRPQESMGm\nk0gsh9wY8XNqdny5Zvl++h9ervrqOMi2A1lZkB5pM5vbf2vHiR8=\n-----END CERTIFICATE-----\n", "key": "/etc/pki/tls/private/quadlet_demo.key", "key_content": "-----BEGIN PRIVATE KEY-----\nMIIEvQIBADANBgkqhkiG9w0BAQEFAASCBKcwggSjAgEAAoIBAQCvCRut0qPEE/RJ\nlI/rkl1t2chPR1NUfGt+Bd1PCJ5GHCREvyzFBUw7GZbIcAv3qqC8fUvWOfeemOEK\nTFZWbtSesgYwzPoQ7bXWHIt6ayunsvb30ZE45ep7TKti9W7DXAUYESYJ6986cJkS\n7uFK9hZpz3j1TB5A1RO1qFFYQgrQljlPL7jDZQIK/W44n0TzSe3h46ABHiDtqLUc\nUxQ+P0ovsQx1WrnHkXjow4aZJE1vuK96+mUr7vDIYizC9HOlpstL0NS8fvItSvm6\nQmTzcfS0+Hv3tUx8Cis3jIldnWVNJ3hqxWOCMx92jmHM/zxLX11MmlMOZwvSZzvk\nl1zobeYHAgMBAAECggEABFKawuOxVK3Teh+3oC84rHZU6pRN3AfIlc+qRAHn+r5G\nLQZ1ohX76He80A1oVkD6AD2t6kf7NAj9gcYZ2b+Z86Nc4ManJFTiiTuiPC86YEVc\nsQkVZ4zSCiJ+Fw8Nobh7jbkaseMH28cdGKkamkHSkuFjUNeE5b/9YR6bh6z13GnW\n1kLk2iayb6qZ1CJCc8UvhNj8o9sL++YIeXUNzjX8571lHLzs0wvz0oLNTXOkx/Vj\nD8jyw+K9oShO7OedXqdkJKo6krOcplLKgHiXp8wJpEAU5LqU331ZPnw8RYEvauf8\nLkOBv6GqLeO9SPBJn9xObNqbaqU91iYPRBZZQqHzoQKBgQDAG4OAJS61RWNTc9Lq\n6W5kzKYcJJ5TPeSX45SFGaiCYtDRHuNAXRWUuMyr8DP94G4XQhghJEnf5bibvRIM\nkDHK8zuU+ZxbOx9OOfLWEIf2uFDyskq5fpxzaIjOb1eNuTpYYlXy/PdNHHNZBMs+\nI0lTHtwUgko2rTD5LyjZyCePZwKBgQDpQA16fTnEDfWbe6FQx+DeGD6fza5AhwjH\nz45avxNmpa5aKPyQwCs+qsO3HEEFNAADg54+hScUihirKqcfyQTj3ciJt6AuZ7N8\nzJ0KwHsz9PKlD2ehxIfv/FV2zNOTZEDIruF6nxu+LTKnPcq4TPNvJUafyFnPd/Ov\nHJ7HTQDwYQKBgH3MhpxPvPpvac+oNSInombXrJ8ZS6RvD1Z3bTaLytE3LE8EzwPC\n8PLm4QBm12UUpoMMgQiS3t2DbRRvB3a6piCkjwvLORx05tcvE1oikSZ7JGcJYe1+\nf60DczLbrpXcCdz+6BGqnbcdPlu4y+7w5H9N4gh8GWUcnMlQnIugimdZAoGBAJ5/\nzijsK3LUYY+0ZRw1ltKU+wkwhTaIDClrGzsDGLXp5B6UN3xNeZVuCOzHIE/e1BuP\nNtsZNEM6zCSUKVjqCo6lSP58bzbZdRhbGmQSBCZ8Tt5nJ26/0/f6KkE+Uo+X0oYM\ngGAI+psyr3cCAF1YvRtL4z/0r0LSKh9WfXbFFtxBAoGAdoZBH4b5gwyMw1MtBnC5\ntpw9dsHyD2qquO7PaBJRNBKqR3R7SLDDK7tGfLYwQfLCIIx4Zo9MkyDD6OzmFWnK\nHTQnNruxV8zUcS2aG1jUywRGIbed25F3jYR4EbSNOttdrzdOmvTGn7LUcWiQkxdD\ndXIWzndiDfPEu6Cycw+SHjs=\n-----END PRIVATE KEY-----\n" }, "rc": 0, "start": "2024-02-04 20:05:58.591190" } STDOUT: Request "20240204200557" removed. TASK [fedora.linux_system_roles.certificate : Remove files] ******************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:181 Sunday 04 February 2024 20:05:58 +0000 (0:00:00.300) 0:00:13.135 ******* changed: [sut] => (item=/etc/pki/tls/certs/quadlet_demo.crt) => { "ansible_loop_var": "item", "changed": true, "item": "/etc/pki/tls/certs/quadlet_demo.crt", "path": "/etc/pki/tls/certs/quadlet_demo.crt", "state": "absent" } changed: [sut] => (item=/etc/pki/tls/private/quadlet_demo.key) => { "ansible_loop_var": "item", "changed": true, "item": "/etc/pki/tls/private/quadlet_demo.key", "path": "/etc/pki/tls/private/quadlet_demo.key", "state": "absent" } ok: [sut] => (item=/etc/pki/tls/certs/quadlet_demo.crt) => { "ansible_loop_var": "item", "changed": false, "item": "/etc/pki/tls/certs/quadlet_demo.crt", "path": "/etc/pki/tls/certs/quadlet_demo.crt", "state": "absent" } TASK [Run the role] ************************************************************ task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/tests_quadlet_demo.yml:50 Sunday 04 February 2024 20:05:59 +0000 (0:00:00.561) 0:00:13.696 ******* TASK [linux-system-roles.podman : Set platform/version specific variables] ***** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/main.yml:3 Sunday 04 February 2024 20:05:59 +0000 (0:00:00.062) 0:00:13.758 ******* included: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/set_vars.yml for sut TASK [linux-system-roles.podman : Ensure ansible_facts used by role] *********** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/set_vars.yml:3 Sunday 04 February 2024 20:05:59 +0000 (0:00:00.017) 0:00:13.776 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Check if system is ostree] ******************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/set_vars.yml:12 Sunday 04 February 2024 20:05:59 +0000 (0:00:00.021) 0:00:13.798 ******* ok: [sut] => { "changed": false, "stat": { "exists": false } } TASK [linux-system-roles.podman : Set flag to indicate system is ostree] ******* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/set_vars.yml:17 Sunday 04 February 2024 20:05:59 +0000 (0:00:00.188) 0:00:13.986 ******* ok: [sut] => { "ansible_facts": { "__podman_is_ostree": false }, "changed": false } TASK [linux-system-roles.podman : Set platform/version specific variables] ***** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/set_vars.yml:21 Sunday 04 February 2024 20:05:59 +0000 (0:00:00.024) 0:00:14.010 ******* ok: [sut] => (item=RedHat.yml) => { "ansible_facts": { "__podman_packages": [ "podman", "shadow-utils-subid" ] }, "ansible_included_var_files": [ "/WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/vars/RedHat.yml" ], "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml" } skipping: [sut] => (item=CentOS.yml) => { "ansible_loop_var": "item", "changed": false, "item": "CentOS.yml", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=CentOS_9.yml) => { "ansible_loop_var": "item", "changed": false, "item": "CentOS_9.yml", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=CentOS_9.yml) => { "ansible_loop_var": "item", "changed": false, "item": "CentOS_9.yml", "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Gather the package facts] ******************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/main.yml:6 Sunday 04 February 2024 20:05:59 +0000 (0:00:00.039) 0:00:14.050 ******* ok: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [linux-system-roles.podman : Enable copr if requested] ******************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/main.yml:10 Sunday 04 February 2024 20:06:00 +0000 (0:00:01.125) 0:00:15.175 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Ensure required packages are installed] ****** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/main.yml:14 Sunday 04 February 2024 20:06:00 +0000 (0:00:00.033) 0:00:15.209 ******* changed: [sut] => { "changed": true, "rc": 0, "results": [ "Installed: libslirp-4.4.0-7.el9.x86_64", "Installed: podman-2:4.8.1-1.el9.x86_64", "Installed: netavark-2:1.9.0-1.el9.x86_64", "Installed: slirp4netns-1.2.2-1.el9.x86_64", "Installed: shadow-utils-subid-2:4.9-8.el9.x86_64", "Installed: criu-3.19-1.el9.x86_64", "Installed: yajl-2.1.0-22.el9.x86_64", "Installed: criu-libs-3.19-1.el9.x86_64", "Installed: crun-1.14-1.el9.x86_64", "Installed: conmon-2:2.1.10-1.el9.x86_64", "Installed: aardvark-dns-2:1.9.0-1.el9.x86_64", "Installed: protobuf-c-1.3.3-13.el9.x86_64", "Installed: fuse-overlayfs-1.12-1.el9.x86_64", "Installed: fuse3-3.10.2-6.el9.x86_64", "Installed: fuse3-libs-3.10.2-6.el9.x86_64", "Installed: container-selinux-3:2.229.0-1.el9.noarch", "Installed: libnet-1.2-6.el9.x86_64", "Installed: containers-common-2:1-55.el9.x86_64", "Installed: fuse-common-3.10.2-6.el9.x86_64" ] } TASK [linux-system-roles.podman : Get podman version] ************************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/main.yml:23 Sunday 04 February 2024 20:06:50 +0000 (0:00:49.703) 0:01:04.912 ******* ok: [sut] => { "changed": false, "cmd": [ "podman", "--version" ], "delta": "0:00:00.033178", "end": "2024-02-04 20:06:50.615555", "rc": 0, "start": "2024-02-04 20:06:50.582377" } STDOUT: podman version 4.8.1 TASK [linux-system-roles.podman : Set podman version] ************************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/main.yml:28 Sunday 04 February 2024 20:06:50 +0000 (0:00:00.241) 0:01:05.153 ******* ok: [sut] => { "ansible_facts": { "podman_version": "4.8.1" }, "changed": false } TASK [linux-system-roles.podman : Podman package version must be 4.2 or later] *** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/main.yml:32 Sunday 04 February 2024 20:06:50 +0000 (0:00:00.033) 0:01:05.187 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Podman package version must be 4.4 or later for quadlet, secrets] *** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/main.yml:39 Sunday 04 February 2024 20:06:50 +0000 (0:00:00.033) 0:01:05.220 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } META: end_host conditional evaluated to false, continuing execution for sut TASK [linux-system-roles.podman : Check user and group information] ************ task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/main.yml:56 Sunday 04 February 2024 20:06:50 +0000 (0:00:00.097) 0:01:05.318 ******* included: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml for sut TASK [linux-system-roles.podman : Get user information] ************************ task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:2 Sunday 04 February 2024 20:06:50 +0000 (0:00:00.044) 0:01:05.362 ******* ok: [sut] => { "ansible_facts": { "getent_passwd": { "root": [ "x", "0", "0", "root", "/root", "/bin/bash" ] } }, "changed": false } TASK [linux-system-roles.podman : Fail if user does not exist] ***************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:10 Sunday 04 February 2024 20:06:51 +0000 (0:00:00.256) 0:01:05.619 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Set group for podman user] ******************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:17 Sunday 04 February 2024 20:06:51 +0000 (0:00:00.027) 0:01:05.646 ******* ok: [sut] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [linux-system-roles.podman : Get group information] *********************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:28 Sunday 04 February 2024 20:06:51 +0000 (0:00:00.028) 0:01:05.675 ******* ok: [sut] => { "ansible_facts": { "getent_group": { "root": [ "x", "0", "" ] } }, "changed": false } TASK [linux-system-roles.podman : Set group name] ****************************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:36 Sunday 04 February 2024 20:06:51 +0000 (0:00:00.197) 0:01:05.872 ******* ok: [sut] => { "ansible_facts": { "__podman_group_name": "root" }, "changed": false } TASK [linux-system-roles.podman : See if getsubids exists] ********************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:41 Sunday 04 February 2024 20:06:51 +0000 (0:00:00.026) 0:01:05.899 ******* ok: [sut] => { "changed": false, "stat": { "atime": 1689166618.0, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "076f89863b474504550ab1b5c11216a77ed7123c", "ctime": 1707077179.8497021, "dev": 51713, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 4380701, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-pie-executable", "mode": "0755", "mtime": 1689166618.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 15496, "uid": 0, "version": "2573075764", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [linux-system-roles.podman : Check user with getsubids] ******************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:52 Sunday 04 February 2024 20:06:51 +0000 (0:00:00.195) 0:01:06.095 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Check group with getsubids] ****************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:56 Sunday 04 February 2024 20:06:51 +0000 (0:00:00.022) 0:01:06.117 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Check if user is in subuid file] ************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:66 Sunday 04 February 2024 20:06:51 +0000 (0:00:00.020) 0:01:06.138 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Fail if user not in subuid file] ************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:74 Sunday 04 February 2024 20:06:51 +0000 (0:00:00.021) 0:01:06.159 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Check if group is in subgid file] ************ task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:81 Sunday 04 February 2024 20:06:51 +0000 (0:00:00.021) 0:01:06.181 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Fail if group not in subgid file] ************ task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:89 Sunday 04 February 2024 20:06:51 +0000 (0:00:00.041) 0:01:06.223 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Set config file paths] *********************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/main.yml:62 Sunday 04 February 2024 20:06:51 +0000 (0:00:00.021) 0:01:06.244 ******* ok: [sut] => { "ansible_facts": { "__podman_container_conf_file": "/etc/containers/containers.conf.d/50-systemroles.conf", "__podman_policy_json_file": "/etc/containers/policy.json", "__podman_registries_conf_file": "/etc/containers/registries.conf.d/50-systemroles.conf", "__podman_storage_conf_file": "/etc/containers/storage.conf" }, "changed": false } TASK [linux-system-roles.podman : Handle container.conf.d] ********************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/main.yml:80 Sunday 04 February 2024 20:06:51 +0000 (0:00:00.047) 0:01:06.292 ******* included: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_container_conf_d.yml for sut TASK [linux-system-roles.podman : Ensure containers.d exists] ****************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_container_conf_d.yml:5 Sunday 04 February 2024 20:06:51 +0000 (0:00:00.044) 0:01:06.337 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Update container config file] **************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_container_conf_d.yml:14 Sunday 04 February 2024 20:06:51 +0000 (0:00:00.020) 0:01:06.358 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Handle registries.conf.d] ******************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/main.yml:83 Sunday 04 February 2024 20:06:51 +0000 (0:00:00.021) 0:01:06.379 ******* included: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_registries_conf_d.yml for sut TASK [linux-system-roles.podman : Ensure registries.d exists] ****************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_registries_conf_d.yml:5 Sunday 04 February 2024 20:06:51 +0000 (0:00:00.044) 0:01:06.424 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Update registries config file] *************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_registries_conf_d.yml:14 Sunday 04 February 2024 20:06:51 +0000 (0:00:00.021) 0:01:06.446 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Handle storage.conf] ************************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/main.yml:86 Sunday 04 February 2024 20:06:51 +0000 (0:00:00.021) 0:01:06.467 ******* included: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_storage_conf.yml for sut TASK [linux-system-roles.podman : Ensure storage.conf parent dir exists] ******* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_storage_conf.yml:5 Sunday 04 February 2024 20:06:52 +0000 (0:00:00.045) 0:01:06.513 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Update storage config file] ****************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_storage_conf.yml:14 Sunday 04 February 2024 20:06:52 +0000 (0:00:00.021) 0:01:06.534 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Handle policy.json] ************************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/main.yml:89 Sunday 04 February 2024 20:06:52 +0000 (0:00:00.021) 0:01:06.556 ******* included: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_policy_json.yml for sut TASK [linux-system-roles.podman : Ensure policy.json parent dir exists] ******** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_policy_json.yml:6 Sunday 04 February 2024 20:06:52 +0000 (0:00:00.044) 0:01:06.601 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Stat the policy.json file] ******************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_policy_json.yml:15 Sunday 04 February 2024 20:06:52 +0000 (0:00:00.022) 0:01:06.623 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Get the existing policy.json] **************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_policy_json.yml:20 Sunday 04 February 2024 20:06:52 +0000 (0:00:00.023) 0:01:06.647 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Write new policy.json file] ****************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_policy_json.yml:26 Sunday 04 February 2024 20:06:52 +0000 (0:00:00.022) 0:01:06.669 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Manage firewall for specified ports] ************************************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/main.yml:95 Sunday 04 February 2024 20:06:52 +0000 (0:00:00.023) 0:01:06.692 ******* TASK [fedora.linux_system_roles.firewall : Setup firewalld] ******************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:2 Sunday 04 February 2024 20:06:52 +0000 (0:00:00.099) 0:01:06.791 ******* included: /WORKDIR/git-fix-linger-handling7nz_rqha/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml for sut TASK [fedora.linux_system_roles.firewall : Ensure ansible_facts used by role] *** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:2 Sunday 04 February 2024 20:06:52 +0000 (0:00:00.041) 0:01:06.833 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Check if system is ostree] ********** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:10 Sunday 04 February 2024 20:06:52 +0000 (0:00:00.027) 0:01:06.860 ******* ok: [sut] => { "changed": false, "stat": { "exists": false } } TASK [fedora.linux_system_roles.firewall : Set flag to indicate system is ostree] *** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:15 Sunday 04 February 2024 20:06:52 +0000 (0:00:00.192) 0:01:07.053 ******* ok: [sut] => { "ansible_facts": { "__firewall_is_ostree": false }, "changed": false } TASK [fedora.linux_system_roles.firewall : Install firewalld] ****************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:19 Sunday 04 February 2024 20:06:52 +0000 (0:00:00.025) 0:01:07.078 ******* ok: [sut] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [fedora.linux_system_roles.firewall : Collect service facts] ************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:5 Sunday 04 February 2024 20:06:53 +0000 (0:00:00.887) 0:01:07.966 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Attempt to stop and disable conflicting services] *** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:9 Sunday 04 February 2024 20:06:53 +0000 (0:00:00.022) 0:01:07.988 ******* skipping: [sut] => (item=nftables) => { "ansible_loop_var": "item", "changed": false, "item": "nftables", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=iptables) => { "ansible_loop_var": "item", "changed": false, "item": "iptables", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=ufw) => { "ansible_loop_var": "item", "changed": false, "item": "ufw", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Unmask firewalld service] *********** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:22 Sunday 04 February 2024 20:06:53 +0000 (0:00:00.030) 0:01:08.018 ******* ok: [sut] => { "changed": false, "name": "firewalld", "status": { "AccessSELinuxContext": "system_u:object_r:firewalld_unit_file_t:s0", "ActiveEnterTimestampMonotonic": "0", "ActiveExitTimestampMonotonic": "0", "ActiveState": "inactive", "After": "polkit.service basic.target system.slice dbus-broker.service sysinit.target dbus.socket", "AllowIsolate": "no", "AssertResult": "no", "AssertTimestampMonotonic": "0", "Before": "shutdown.target network-pre.target", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "BusName": "org.fedoraproject.FirewallD1", "CPUAccounting": "yes", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "[not set]", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "yes", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore", "CleanResult": "success", "CollectMode": "inactive", "ConditionResult": "no", "ConditionTimestampMonotonic": "0", "ConfigurationDirectoryMode": "0755", "Conflicts": "iptables.service ip6tables.service ebtables.service ipset.service nftables.service shutdown.target", "ControlGroupId": "0", "ControlPID": "0", "CoredumpFilter": "0x33", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "Delegate": "no", "Description": "firewalld - dynamic firewall daemon", "DevicePolicy": "auto", "Documentation": "\"man:firewalld(1)\"", "DynamicUser": "no", "EnvironmentFiles": "/etc/sysconfig/firewalld (ignore_errors=yes)", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "0", "ExecMainStartTimestampMonotonic": "0", "ExecMainStatus": "0", "ExecReload": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecReloadEx": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStart": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStartEx": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExitType": "main", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FinalKillSignal": "9", "FragmentPath": "/usr/lib/systemd/system/firewalld.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOReadBytes": "18446744073709551615", "IOReadOperations": "18446744073709551615", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "IOWeight": "[not set]", "IOWriteBytes": "18446744073709551615", "IOWriteOperations": "18446744073709551615", "IPAccounting": "no", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "Id": "firewalld.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestampMonotonic": "0", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "mixed", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "0", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "8388608", "LimitMEMLOCKSoft": "8388608", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "524288", "LimitNOFILESoft": "1024", "LimitNPROC": "13974", "LimitNPROCSoft": "13974", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "13974", "LimitSIGPENDINGSoft": "13974", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "0", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "ManagedOOMSwap": "auto", "MemoryAccounting": "yes", "MemoryAvailable": "infinity", "MemoryCurrent": "[not set]", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemorySwapMax": "infinity", "MountAPIVFS": "no", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAPolicy": "n/a", "Names": "firewalld.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMPolicy": "stop", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "OnSuccessJobMode": "fail", "Perpetual": "no", "PrivateDevices": "no", "PrivateIPC": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProcSubset": "all", "ProtectClock": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectHostname": "no", "ProtectKernelLogs": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectProc": "default", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "ReloadResult": "success", "ReloadSignal": "1", "RemainAfterExit": "no", "RemoveIPC": "no", "Requires": "dbus.socket system.slice sysinit.target", "Restart": "no", "RestartKillSignal": "15", "RestartUSec": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "Slice": "system.slice", "StandardError": "null", "StandardInput": "null", "StandardOutput": "null", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StateChangeTimestampMonotonic": "0", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "dead", "SuccessAction": "none", "SyslogFacility": "3", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "2147483646", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "[not set]", "TasksMax": "22359", "TimeoutAbortUSec": "1min 30s", "TimeoutCleanUSec": "infinity", "TimeoutStartFailureMode": "terminate", "TimeoutStartUSec": "1min 30s", "TimeoutStopFailureMode": "terminate", "TimeoutStopUSec": "1min 30s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "dbus", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "enabled", "UnitFileState": "disabled", "UtmpMode": "init", "Wants": "network-pre.target", "WatchdogSignal": "6", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "infinity" } } TASK [fedora.linux_system_roles.firewall : Enable and start firewalld service] *** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:28 Sunday 04 February 2024 20:06:53 +0000 (0:00:00.346) 0:01:08.364 ******* changed: [sut] => { "changed": true, "enabled": true, "name": "firewalld", "state": "started", "status": { "AccessSELinuxContext": "system_u:object_r:firewalld_unit_file_t:s0", "ActiveEnterTimestampMonotonic": "0", "ActiveExitTimestampMonotonic": "0", "ActiveState": "inactive", "After": "basic.target dbus.socket system.slice polkit.service dbus-broker.service sysinit.target", "AllowIsolate": "no", "AssertResult": "no", "AssertTimestampMonotonic": "0", "Before": "shutdown.target network-pre.target", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "BusName": "org.fedoraproject.FirewallD1", "CPUAccounting": "yes", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "[not set]", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "yes", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore", "CleanResult": "success", "CollectMode": "inactive", "ConditionResult": "no", "ConditionTimestampMonotonic": "0", "ConfigurationDirectoryMode": "0755", "Conflicts": "iptables.service shutdown.target nftables.service ip6tables.service ipset.service ebtables.service", "ControlGroupId": "0", "ControlPID": "0", "CoredumpFilter": "0x33", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "Delegate": "no", "Description": "firewalld - dynamic firewall daemon", "DevicePolicy": "auto", "Documentation": "\"man:firewalld(1)\"", "DynamicUser": "no", "EnvironmentFiles": "/etc/sysconfig/firewalld (ignore_errors=yes)", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "0", "ExecMainStartTimestampMonotonic": "0", "ExecMainStatus": "0", "ExecReload": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecReloadEx": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStart": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStartEx": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExitType": "main", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FinalKillSignal": "9", "FragmentPath": "/usr/lib/systemd/system/firewalld.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOReadBytes": "18446744073709551615", "IOReadOperations": "18446744073709551615", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "IOWeight": "[not set]", "IOWriteBytes": "18446744073709551615", "IOWriteOperations": "18446744073709551615", "IPAccounting": "no", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "Id": "firewalld.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestampMonotonic": "0", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "mixed", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "0", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "8388608", "LimitMEMLOCKSoft": "8388608", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "524288", "LimitNOFILESoft": "1024", "LimitNPROC": "13974", "LimitNPROCSoft": "13974", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "13974", "LimitSIGPENDINGSoft": "13974", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "0", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "ManagedOOMSwap": "auto", "MemoryAccounting": "yes", "MemoryAvailable": "infinity", "MemoryCurrent": "[not set]", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemorySwapMax": "infinity", "MountAPIVFS": "no", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAPolicy": "n/a", "Names": "firewalld.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMPolicy": "stop", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "OnSuccessJobMode": "fail", "Perpetual": "no", "PrivateDevices": "no", "PrivateIPC": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProcSubset": "all", "ProtectClock": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectHostname": "no", "ProtectKernelLogs": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectProc": "default", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "ReloadResult": "success", "ReloadSignal": "1", "RemainAfterExit": "no", "RemoveIPC": "no", "Requires": "system.slice dbus.socket sysinit.target", "Restart": "no", "RestartKillSignal": "15", "RestartUSec": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "Slice": "system.slice", "StandardError": "null", "StandardInput": "null", "StandardOutput": "null", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StateChangeTimestampMonotonic": "0", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "dead", "SuccessAction": "none", "SyslogFacility": "3", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "2147483646", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "[not set]", "TasksMax": "22359", "TimeoutAbortUSec": "1min 30s", "TimeoutCleanUSec": "infinity", "TimeoutStartFailureMode": "terminate", "TimeoutStartUSec": "1min 30s", "TimeoutStopFailureMode": "terminate", "TimeoutStopUSec": "1min 30s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "dbus", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "enabled", "UnitFileState": "disabled", "UtmpMode": "init", "Wants": "network-pre.target", "WatchdogSignal": "6", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "infinity" } } TASK [fedora.linux_system_roles.firewall : Check if previous replaced is defined] *** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:34 Sunday 04 February 2024 20:06:55 +0000 (0:00:01.678) 0:01:10.043 ******* ok: [sut] => { "ansible_facts": { "__firewall_previous_replaced": false, "__firewall_python_cmd": "/usr/bin/python3", "__firewall_report_changed": true }, "changed": false } TASK [fedora.linux_system_roles.firewall : Get config files, checksums before and remove] *** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:43 Sunday 04 February 2024 20:06:55 +0000 (0:00:00.027) 0:01:10.071 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Tell firewall module it is able to report changed] *** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:55 Sunday 04 February 2024 20:06:55 +0000 (0:00:00.023) 0:01:10.095 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Configure firewall] ***************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:71 Sunday 04 February 2024 20:06:55 +0000 (0:00:00.022) 0:01:10.117 ******* changed: [sut] => (item={'port': '8000/tcp', 'state': 'enabled'}) => { "__firewall_changed": true, "ansible_loop_var": "item", "changed": true, "item": { "port": "8000/tcp", "state": "enabled" } } changed: [sut] => (item={'port': '9000/tcp', 'state': 'enabled'}) => { "__firewall_changed": true, "ansible_loop_var": "item", "changed": true, "item": { "port": "9000/tcp", "state": "enabled" } } TASK [fedora.linux_system_roles.firewall : Gather firewall config information] *** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:120 Sunday 04 February 2024 20:06:57 +0000 (0:00:01.625) 0:01:11.742 ******* skipping: [sut] => (item={'port': '8000/tcp', 'state': 'enabled'}) => { "ansible_loop_var": "item", "changed": false, "item": { "port": "8000/tcp", "state": "enabled" }, "skip_reason": "Conditional result was False" } skipping: [sut] => (item={'port': '9000/tcp', 'state': 'enabled'}) => { "ansible_loop_var": "item", "changed": false, "item": { "port": "9000/tcp", "state": "enabled" }, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Update firewalld_config fact] ******* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:130 Sunday 04 February 2024 20:06:57 +0000 (0:00:00.038) 0:01:11.781 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Gather firewall config if no arguments] *** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:139 Sunday 04 February 2024 20:06:57 +0000 (0:00:00.045) 0:01:11.826 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Update firewalld_config fact] ******* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:144 Sunday 04 February 2024 20:06:57 +0000 (0:00:00.023) 0:01:11.850 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Get config files, checksums after] *** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:153 Sunday 04 February 2024 20:06:57 +0000 (0:00:00.022) 0:01:11.872 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Calculate what has changed] ********* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:163 Sunday 04 February 2024 20:06:57 +0000 (0:00:00.025) 0:01:11.897 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Show diffs] ************************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:169 Sunday 04 February 2024 20:06:57 +0000 (0:00:00.022) 0:01:11.919 ******* skipping: [sut] => {} TASK [Manage selinux for specified ports] ************************************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/main.yml:102 Sunday 04 February 2024 20:06:57 +0000 (0:00:00.023) 0:01:11.943 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Keep track of users that need to cancel linger] *** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/main.yml:109 Sunday 04 February 2024 20:06:57 +0000 (0:00:00.022) 0:01:11.965 ******* ok: [sut] => { "ansible_facts": { "__podman_cancel_user_linger": [] }, "changed": false } TASK [linux-system-roles.podman : Handle secrets] ****************************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/main.yml:113 Sunday 04 February 2024 20:06:57 +0000 (0:00:00.022) 0:01:11.988 ******* included: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_secret.yml for sut included: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_secret.yml for sut included: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_secret.yml for sut TASK [linux-system-roles.podman : Set variables part 0] ************************ task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_secret.yml:3 Sunday 04 February 2024 20:06:57 +0000 (0:00:00.084) 0:01:12.073 ******* ok: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [linux-system-roles.podman : Set variables part 1] ************************ task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_secret.yml:12 Sunday 04 February 2024 20:06:57 +0000 (0:00:00.026) 0:01:12.099 ******* ok: [sut] => { "ansible_facts": { "__podman_user": "root" }, "changed": false } TASK [linux-system-roles.podman : Set variables part 2] ************************ task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_secret.yml:17 Sunday 04 February 2024 20:06:57 +0000 (0:00:00.022) 0:01:12.122 ******* ok: [sut] => { "ansible_facts": { "__podman_rootless": false, "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [linux-system-roles.podman : Manage linger] ******************************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_secret.yml:23 Sunday 04 February 2024 20:06:57 +0000 (0:00:00.029) 0:01:12.152 ******* included: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/manage_linger.yml for sut TASK [linux-system-roles.podman : Enable linger if needed] ********************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/manage_linger.yml:12 Sunday 04 February 2024 20:06:57 +0000 (0:00:00.037) 0:01:12.189 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Mark user as not yet needing to cancel linger] *** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/manage_linger.yml:18 Sunday 04 February 2024 20:06:57 +0000 (0:00:00.022) 0:01:12.211 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Mark user for possible linger cancel] ******** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/manage_linger.yml:22 Sunday 04 February 2024 20:06:57 +0000 (0:00:00.021) 0:01:12.233 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Manage each secret] ************************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_secret.yml:28 Sunday 04 February 2024 20:06:57 +0000 (0:00:00.022) 0:01:12.256 ******* ok: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [linux-system-roles.podman : Set variables part 0] ************************ task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_secret.yml:3 Sunday 04 February 2024 20:06:58 +0000 (0:00:00.333) 0:01:12.590 ******* ok: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [linux-system-roles.podman : Set variables part 1] ************************ task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_secret.yml:12 Sunday 04 February 2024 20:06:58 +0000 (0:00:00.048) 0:01:12.639 ******* ok: [sut] => { "ansible_facts": { "__podman_user": "root" }, "changed": false } TASK [linux-system-roles.podman : Set variables part 2] ************************ task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_secret.yml:17 Sunday 04 February 2024 20:06:58 +0000 (0:00:00.025) 0:01:12.664 ******* ok: [sut] => { "ansible_facts": { "__podman_rootless": false, "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [linux-system-roles.podman : Manage linger] ******************************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_secret.yml:23 Sunday 04 February 2024 20:06:58 +0000 (0:00:00.028) 0:01:12.693 ******* included: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/manage_linger.yml for sut TASK [linux-system-roles.podman : Enable linger if needed] ********************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/manage_linger.yml:12 Sunday 04 February 2024 20:06:58 +0000 (0:00:00.037) 0:01:12.731 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Mark user as not yet needing to cancel linger] *** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/manage_linger.yml:18 Sunday 04 February 2024 20:06:58 +0000 (0:00:00.021) 0:01:12.753 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Mark user for possible linger cancel] ******** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/manage_linger.yml:22 Sunday 04 February 2024 20:06:58 +0000 (0:00:00.021) 0:01:12.774 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Manage each secret] ************************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_secret.yml:28 Sunday 04 February 2024 20:06:58 +0000 (0:00:00.023) 0:01:12.798 ******* ok: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [linux-system-roles.podman : Set variables part 0] ************************ task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_secret.yml:3 Sunday 04 February 2024 20:06:58 +0000 (0:00:00.234) 0:01:13.033 ******* ok: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [linux-system-roles.podman : Set variables part 1] ************************ task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_secret.yml:12 Sunday 04 February 2024 20:06:58 +0000 (0:00:00.024) 0:01:13.057 ******* ok: [sut] => { "ansible_facts": { "__podman_user": "root" }, "changed": false } TASK [linux-system-roles.podman : Set variables part 2] ************************ task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_secret.yml:17 Sunday 04 February 2024 20:06:58 +0000 (0:00:00.023) 0:01:13.081 ******* ok: [sut] => { "ansible_facts": { "__podman_rootless": false, "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [linux-system-roles.podman : Manage linger] ******************************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_secret.yml:23 Sunday 04 February 2024 20:06:58 +0000 (0:00:00.028) 0:01:13.109 ******* included: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/manage_linger.yml for sut TASK [linux-system-roles.podman : Enable linger if needed] ********************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/manage_linger.yml:12 Sunday 04 February 2024 20:06:58 +0000 (0:00:00.036) 0:01:13.146 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Mark user as not yet needing to cancel linger] *** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/manage_linger.yml:18 Sunday 04 February 2024 20:06:58 +0000 (0:00:00.021) 0:01:13.167 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Mark user for possible linger cancel] ******** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/manage_linger.yml:22 Sunday 04 February 2024 20:06:58 +0000 (0:00:00.021) 0:01:13.188 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Manage each secret] ************************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_secret.yml:28 Sunday 04 February 2024 20:06:58 +0000 (0:00:00.021) 0:01:13.210 ******* ok: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [linux-system-roles.podman : Handle Kubernetes specifications] ************ task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/main.yml:120 Sunday 04 February 2024 20:06:58 +0000 (0:00:00.236) 0:01:13.446 ******* TASK [linux-system-roles.podman : Handle Quadlet specifications] *************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/main.yml:126 Sunday 04 February 2024 20:06:58 +0000 (0:00:00.020) 0:01:13.467 ******* included: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml for sut included: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml for sut included: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml for sut included: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml for sut included: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml for sut included: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml for sut TASK [linux-system-roles.podman : Set per-container variables part 0] ********** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:14 Sunday 04 February 2024 20:06:59 +0000 (0:00:00.141) 0:01:13.608 ******* ok: [sut] => { "ansible_facts": { "__podman_quadlet_file_src": "quadlet-demo.network", "__podman_quadlet_spec": {}, "__podman_quadlet_str": "[Network]\nSubnet=192.168.30.0/24\nGateway=192.168.30.1\nLabel=app=wordpress", "__podman_quadlet_template_src": "" }, "changed": false } TASK [linux-system-roles.podman : Set per-container variables part 1] ********** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:40 Sunday 04 February 2024 20:06:59 +0000 (0:00:00.032) 0:01:13.641 ******* ok: [sut] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_pull_image": true, "__podman_state": "created", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [linux-system-roles.podman : Fail if no quadlet spec is given] ************ task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:57 Sunday 04 February 2024 20:06:59 +0000 (0:00:00.027) 0:01:13.668 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Set per-container variables part 2] ********** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:70 Sunday 04 February 2024 20:06:59 +0000 (0:00:00.024) 0:01:13.692 ******* ok: [sut] => { "ansible_facts": { "__podman_quadlet_name": "quadlet-demo", "__podman_quadlet_type": "network", "__podman_rootless": false }, "changed": false } TASK [linux-system-roles.podman : Check user and group information] ************ task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:97 Sunday 04 February 2024 20:06:59 +0000 (0:00:00.032) 0:01:13.725 ******* included: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml for sut TASK [linux-system-roles.podman : Get user information] ************************ task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:2 Sunday 04 February 2024 20:06:59 +0000 (0:00:00.041) 0:01:13.766 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Fail if user does not exist] ***************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:10 Sunday 04 February 2024 20:06:59 +0000 (0:00:00.026) 0:01:13.792 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Set group for podman user] ******************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:17 Sunday 04 February 2024 20:06:59 +0000 (0:00:00.027) 0:01:13.819 ******* ok: [sut] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [linux-system-roles.podman : Get group information] *********************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:28 Sunday 04 February 2024 20:06:59 +0000 (0:00:00.030) 0:01:13.850 ******* ok: [sut] => { "ansible_facts": { "getent_group": { "root": [ "x", "0", "" ] } }, "changed": false } TASK [linux-system-roles.podman : Set group name] ****************************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:36 Sunday 04 February 2024 20:06:59 +0000 (0:00:00.199) 0:01:14.049 ******* ok: [sut] => { "ansible_facts": { "__podman_group_name": "root" }, "changed": false } TASK [linux-system-roles.podman : See if getsubids exists] ********************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:41 Sunday 04 February 2024 20:06:59 +0000 (0:00:00.027) 0:01:14.077 ******* ok: [sut] => { "changed": false, "stat": { "atime": 1707077211.5674798, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "076f89863b474504550ab1b5c11216a77ed7123c", "ctime": 1707077179.8497021, "dev": 51713, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 4380701, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-pie-executable", "mode": "0755", "mtime": 1689166618.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 15496, "uid": 0, "version": "2573075764", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [linux-system-roles.podman : Check user with getsubids] ******************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:52 Sunday 04 February 2024 20:06:59 +0000 (0:00:00.196) 0:01:14.274 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Check group with getsubids] ****************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:56 Sunday 04 February 2024 20:06:59 +0000 (0:00:00.022) 0:01:14.296 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Check if user is in subuid file] ************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:66 Sunday 04 February 2024 20:06:59 +0000 (0:00:00.022) 0:01:14.319 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Fail if user not in subuid file] ************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:74 Sunday 04 February 2024 20:06:59 +0000 (0:00:00.049) 0:01:14.368 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Check if group is in subgid file] ************ task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:81 Sunday 04 February 2024 20:06:59 +0000 (0:00:00.022) 0:01:14.391 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Fail if group not in subgid file] ************ task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:89 Sunday 04 February 2024 20:06:59 +0000 (0:00:00.023) 0:01:14.414 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Set per-container variables part 3] ********** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:102 Sunday 04 February 2024 20:06:59 +0000 (0:00:00.021) 0:01:14.436 ******* ok: [sut] => { "ansible_facts": { "__podman_activate_systemd_unit": true, "__podman_images_found": [], "__podman_kube_yamls_raw": "", "__podman_service_name": "", "__podman_systemd_scope": "system", "__podman_user_home_dir": "/root", "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [linux-system-roles.podman : Set per-container variables part 4] ********** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:134 Sunday 04 February 2024 20:06:59 +0000 (0:00:00.040) 0:01:14.477 ******* ok: [sut] => { "ansible_facts": { "__podman_quadlet_path": "/etc/containers/systemd" }, "changed": false } TASK [linux-system-roles.podman : Get kube yaml contents] ********************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:140 Sunday 04 February 2024 20:07:00 +0000 (0:00:00.022) 0:01:14.500 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Set per-container variables part 5] ********** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:151 Sunday 04 February 2024 20:07:00 +0000 (0:00:00.023) 0:01:14.523 ******* ok: [sut] => { "ansible_facts": { "__podman_images": [], "__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo.network", "__podman_volumes": [] }, "changed": false } TASK [linux-system-roles.podman : Cleanup quadlets] **************************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:205 Sunday 04 February 2024 20:07:00 +0000 (0:00:00.056) 0:01:14.579 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Create and update quadlets] ****************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:209 Sunday 04 February 2024 20:07:00 +0000 (0:00:00.021) 0:01:14.601 ******* included: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/create_update_quadlet_spec.yml for sut TASK [linux-system-roles.podman : Manage linger] ******************************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/create_update_quadlet_spec.yml:2 Sunday 04 February 2024 20:07:00 +0000 (0:00:00.051) 0:01:14.653 ******* included: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/manage_linger.yml for sut TASK [linux-system-roles.podman : Enable linger if needed] ********************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/manage_linger.yml:12 Sunday 04 February 2024 20:07:00 +0000 (0:00:00.041) 0:01:14.694 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Mark user as not yet needing to cancel linger] *** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/manage_linger.yml:18 Sunday 04 February 2024 20:07:00 +0000 (0:00:00.022) 0:01:14.717 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Mark user for possible linger cancel] ******** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/manage_linger.yml:22 Sunday 04 February 2024 20:07:00 +0000 (0:00:00.021) 0:01:14.739 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Create host directories] ********************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/create_update_quadlet_spec.yml:7 Sunday 04 February 2024 20:07:00 +0000 (0:00:00.022) 0:01:14.761 ******* TASK [linux-system-roles.podman : Ensure container images are present] ********* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/create_update_quadlet_spec.yml:25 Sunday 04 February 2024 20:07:00 +0000 (0:00:00.019) 0:01:14.781 ******* TASK [linux-system-roles.podman : Ensure the quadlet directory is present] ***** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/create_update_quadlet_spec.yml:42 Sunday 04 February 2024 20:07:00 +0000 (0:00:00.022) 0:01:14.803 ******* ok: [sut] => { "changed": false, "gid": 0, "group": "root", "mode": "0755", "owner": "root", "path": "/etc/containers/systemd", "secontext": "system_u:object_r:etc_t:s0", "size": 6, "state": "directory", "uid": 0 } TASK [linux-system-roles.podman : Ensure quadlet file is copied] *************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/create_update_quadlet_spec.yml:51 Sunday 04 February 2024 20:07:00 +0000 (0:00:00.205) 0:01:15.009 ******* changed: [sut] => { "changed": true, "checksum": "e57c08d49aff4bae8daab138d913aeddaa8682a0", "dest": "/etc/containers/systemd/quadlet-demo.network", "gid": 0, "group": "root", "md5sum": "061f3cf318cbd8ab5794bb1173831fb8", "mode": "0644", "owner": "root", "secontext": "system_u:object_r:etc_t:s0", "size": 74, "src": "/root/.ansible/tmp/ansible-tmp-1707077220.554798-10119-150119709159900/source", "state": "file", "uid": 0 } TASK [linux-system-roles.podman : Ensure quadlet file content is present] ****** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/create_update_quadlet_spec.yml:61 Sunday 04 February 2024 20:07:01 +0000 (0:00:00.608) 0:01:15.617 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Ensure quadlet file is present] ************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/create_update_quadlet_spec.yml:73 Sunday 04 February 2024 20:07:01 +0000 (0:00:00.024) 0:01:15.642 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Reload systemctl] **************************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/create_update_quadlet_spec.yml:85 Sunday 04 February 2024 20:07:01 +0000 (0:00:00.023) 0:01:15.665 ******* ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.podman : Start service] ******************************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/create_update_quadlet_spec.yml:114 Sunday 04 February 2024 20:07:01 +0000 (0:00:00.505) 0:01:16.171 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Restart service] ***************************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/create_update_quadlet_spec.yml:130 Sunday 04 February 2024 20:07:01 +0000 (0:00:00.025) 0:01:16.197 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Set per-container variables part 0] ********** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:14 Sunday 04 February 2024 20:07:01 +0000 (0:00:00.026) 0:01:16.223 ******* ok: [sut] => { "ansible_facts": { "__podman_quadlet_file_src": "quadlet-demo-mysql.volume", "__podman_quadlet_spec": {}, "__podman_quadlet_str": "[Volume]", "__podman_quadlet_template_src": "" }, "changed": false } TASK [linux-system-roles.podman : Set per-container variables part 1] ********** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:40 Sunday 04 February 2024 20:07:01 +0000 (0:00:00.031) 0:01:16.255 ******* ok: [sut] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_pull_image": true, "__podman_state": "created", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [linux-system-roles.podman : Fail if no quadlet spec is given] ************ task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:57 Sunday 04 February 2024 20:07:01 +0000 (0:00:00.027) 0:01:16.282 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Set per-container variables part 2] ********** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:70 Sunday 04 February 2024 20:07:01 +0000 (0:00:00.022) 0:01:16.305 ******* ok: [sut] => { "ansible_facts": { "__podman_quadlet_name": "quadlet-demo-mysql", "__podman_quadlet_type": "volume", "__podman_rootless": false }, "changed": false } TASK [linux-system-roles.podman : Check user and group information] ************ task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:97 Sunday 04 February 2024 20:07:01 +0000 (0:00:00.032) 0:01:16.337 ******* included: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml for sut TASK [linux-system-roles.podman : Get user information] ************************ task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:2 Sunday 04 February 2024 20:07:01 +0000 (0:00:00.040) 0:01:16.377 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Fail if user does not exist] ***************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:10 Sunday 04 February 2024 20:07:01 +0000 (0:00:00.025) 0:01:16.403 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Set group for podman user] ******************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:17 Sunday 04 February 2024 20:07:01 +0000 (0:00:00.026) 0:01:16.430 ******* ok: [sut] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [linux-system-roles.podman : Get group information] *********************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:28 Sunday 04 February 2024 20:07:01 +0000 (0:00:00.029) 0:01:16.460 ******* ok: [sut] => { "ansible_facts": { "getent_group": { "root": [ "x", "0", "" ] } }, "changed": false } TASK [linux-system-roles.podman : Set group name] ****************************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:36 Sunday 04 February 2024 20:07:02 +0000 (0:00:00.205) 0:01:16.665 ******* ok: [sut] => { "ansible_facts": { "__podman_group_name": "root" }, "changed": false } TASK [linux-system-roles.podman : See if getsubids exists] ********************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:41 Sunday 04 February 2024 20:07:02 +0000 (0:00:00.074) 0:01:16.740 ******* ok: [sut] => { "changed": false, "stat": { "atime": 1707077211.5674798, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "076f89863b474504550ab1b5c11216a77ed7123c", "ctime": 1707077179.8497021, "dev": 51713, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 4380701, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-pie-executable", "mode": "0755", "mtime": 1689166618.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 15496, "uid": 0, "version": "2573075764", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [linux-system-roles.podman : Check user with getsubids] ******************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:52 Sunday 04 February 2024 20:07:02 +0000 (0:00:00.201) 0:01:16.942 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Check group with getsubids] ****************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:56 Sunday 04 February 2024 20:07:02 +0000 (0:00:00.023) 0:01:16.965 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Check if user is in subuid file] ************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:66 Sunday 04 February 2024 20:07:02 +0000 (0:00:00.022) 0:01:16.987 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Fail if user not in subuid file] ************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:74 Sunday 04 February 2024 20:07:02 +0000 (0:00:00.024) 0:01:17.011 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Check if group is in subgid file] ************ task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:81 Sunday 04 February 2024 20:07:02 +0000 (0:00:00.023) 0:01:17.035 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Fail if group not in subgid file] ************ task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:89 Sunday 04 February 2024 20:07:02 +0000 (0:00:00.021) 0:01:17.057 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Set per-container variables part 3] ********** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:102 Sunday 04 February 2024 20:07:02 +0000 (0:00:00.021) 0:01:17.078 ******* ok: [sut] => { "ansible_facts": { "__podman_activate_systemd_unit": true, "__podman_images_found": [], "__podman_kube_yamls_raw": "", "__podman_service_name": "quadlet-demo-mysql-volume.service", "__podman_systemd_scope": "system", "__podman_user_home_dir": "/root", "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [linux-system-roles.podman : Set per-container variables part 4] ********** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:134 Sunday 04 February 2024 20:07:02 +0000 (0:00:00.039) 0:01:17.118 ******* ok: [sut] => { "ansible_facts": { "__podman_quadlet_path": "/etc/containers/systemd" }, "changed": false } TASK [linux-system-roles.podman : Get kube yaml contents] ********************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:140 Sunday 04 February 2024 20:07:02 +0000 (0:00:00.022) 0:01:17.141 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Set per-container variables part 5] ********** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:151 Sunday 04 February 2024 20:07:02 +0000 (0:00:00.023) 0:01:17.164 ******* ok: [sut] => { "ansible_facts": { "__podman_images": [], "__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo-mysql.volume", "__podman_volumes": [] }, "changed": false } TASK [linux-system-roles.podman : Cleanup quadlets] **************************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:205 Sunday 04 February 2024 20:07:02 +0000 (0:00:00.054) 0:01:17.219 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Create and update quadlets] ****************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:209 Sunday 04 February 2024 20:07:02 +0000 (0:00:00.021) 0:01:17.241 ******* included: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/create_update_quadlet_spec.yml for sut TASK [linux-system-roles.podman : Manage linger] ******************************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/create_update_quadlet_spec.yml:2 Sunday 04 February 2024 20:07:02 +0000 (0:00:00.046) 0:01:17.288 ******* included: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/manage_linger.yml for sut TASK [linux-system-roles.podman : Enable linger if needed] ********************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/manage_linger.yml:12 Sunday 04 February 2024 20:07:02 +0000 (0:00:00.037) 0:01:17.326 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Mark user as not yet needing to cancel linger] *** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/manage_linger.yml:18 Sunday 04 February 2024 20:07:02 +0000 (0:00:00.023) 0:01:17.349 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Mark user for possible linger cancel] ******** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/manage_linger.yml:22 Sunday 04 February 2024 20:07:02 +0000 (0:00:00.022) 0:01:17.371 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Create host directories] ********************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/create_update_quadlet_spec.yml:7 Sunday 04 February 2024 20:07:02 +0000 (0:00:00.022) 0:01:17.394 ******* TASK [linux-system-roles.podman : Ensure container images are present] ********* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/create_update_quadlet_spec.yml:25 Sunday 04 February 2024 20:07:02 +0000 (0:00:00.047) 0:01:17.442 ******* TASK [linux-system-roles.podman : Ensure the quadlet directory is present] ***** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/create_update_quadlet_spec.yml:42 Sunday 04 February 2024 20:07:02 +0000 (0:00:00.022) 0:01:17.464 ******* ok: [sut] => { "changed": false, "gid": 0, "group": "root", "mode": "0755", "owner": "root", "path": "/etc/containers/systemd", "secontext": "system_u:object_r:etc_t:s0", "size": 34, "state": "directory", "uid": 0 } TASK [linux-system-roles.podman : Ensure quadlet file is copied] *************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/create_update_quadlet_spec.yml:51 Sunday 04 February 2024 20:07:03 +0000 (0:00:00.205) 0:01:17.670 ******* changed: [sut] => { "changed": true, "checksum": "585f8cbdf0ec73000f9227dcffbef71e9552ea4a", "dest": "/etc/containers/systemd/quadlet-demo-mysql.volume", "gid": 0, "group": "root", "md5sum": "5ddd03a022aeb4502d9bc8ce436b4233", "mode": "0644", "owner": "root", "secontext": "system_u:object_r:etc_t:s0", "size": 9, "src": "/root/.ansible/tmp/ansible-tmp-1707077223.2156909-10203-183059669130975/source", "state": "file", "uid": 0 } TASK [linux-system-roles.podman : Ensure quadlet file content is present] ****** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/create_update_quadlet_spec.yml:61 Sunday 04 February 2024 20:07:03 +0000 (0:00:00.509) 0:01:18.179 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Ensure quadlet file is present] ************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/create_update_quadlet_spec.yml:73 Sunday 04 February 2024 20:07:03 +0000 (0:00:00.023) 0:01:18.203 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Reload systemctl] **************************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/create_update_quadlet_spec.yml:85 Sunday 04 February 2024 20:07:03 +0000 (0:00:00.024) 0:01:18.227 ******* ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.podman : Start service] ******************************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/create_update_quadlet_spec.yml:114 Sunday 04 February 2024 20:07:04 +0000 (0:00:00.501) 0:01:18.728 ******* changed: [sut] => { "changed": true, "name": "quadlet-demo-mysql-volume.service", "state": "started", "status": { "AccessSELinuxContext": "system_u:object_r:systemd_unit_file_t:s0", "ActiveEnterTimestampMonotonic": "0", "ActiveExitTimestampMonotonic": "0", "ActiveState": "inactive", "After": "system.slice basic.target systemd-journald.socket sysinit.target -.mount", "AllowIsolate": "no", "AssertResult": "no", "AssertTimestampMonotonic": "0", "Before": "shutdown.target", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "CPUAccounting": "yes", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "[not set]", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "no", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore", "CleanResult": "success", "CollectMode": "inactive", "ConditionResult": "no", "ConditionTimestampMonotonic": "0", "ConfigurationDirectoryMode": "0755", "Conflicts": "shutdown.target", "ControlGroupId": "0", "ControlPID": "0", "CoredumpFilter": "0x33", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "Delegate": "no", "Description": "quadlet-demo-mysql-volume.service", "DevicePolicy": "auto", "DynamicUser": "no", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "0", "ExecMainStartTimestampMonotonic": "0", "ExecMainStatus": "0", "ExecStart": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman volume create --ignore systemd-quadlet-demo-mysql ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStartEx": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman volume create --ignore systemd-quadlet-demo-mysql ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExitType": "main", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FinalKillSignal": "9", "FragmentPath": "/run/systemd/generator/quadlet-demo-mysql-volume.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOReadBytes": "18446744073709551615", "IOReadOperations": "18446744073709551615", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "IOWeight": "[not set]", "IOWriteBytes": "18446744073709551615", "IOWriteOperations": "18446744073709551615", "IPAccounting": "no", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "Id": "quadlet-demo-mysql-volume.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestampMonotonic": "0", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "control-group", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "0", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "8388608", "LimitMEMLOCKSoft": "8388608", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "524288", "LimitNOFILESoft": "1024", "LimitNPROC": "13974", "LimitNPROCSoft": "13974", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "13974", "LimitSIGPENDINGSoft": "13974", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "0", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "ManagedOOMSwap": "auto", "MemoryAccounting": "yes", "MemoryAvailable": "infinity", "MemoryCurrent": "[not set]", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemorySwapMax": "infinity", "MountAPIVFS": "no", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAPolicy": "n/a", "Names": "quadlet-demo-mysql-volume.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMPolicy": "stop", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "OnSuccessJobMode": "fail", "Perpetual": "no", "PrivateDevices": "no", "PrivateIPC": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProcSubset": "all", "ProtectClock": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectHostname": "no", "ProtectKernelLogs": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectProc": "default", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "ReloadResult": "success", "ReloadSignal": "1", "RemainAfterExit": "yes", "RemoveIPC": "no", "Requires": "system.slice sysinit.target -.mount", "RequiresMountsFor": "/run/containers", "Restart": "no", "RestartKillSignal": "15", "RestartUSec": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "Slice": "system.slice", "StandardError": "inherit", "StandardInput": "null", "StandardOutput": "journal", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StateChangeTimestampMonotonic": "0", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "dead", "SuccessAction": "none", "SyslogFacility": "3", "SyslogIdentifier": "quadlet-demo-mysql-volume", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "2147483646", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "[not set]", "TasksMax": "22359", "TimeoutAbortUSec": "1min 30s", "TimeoutCleanUSec": "infinity", "TimeoutStartFailureMode": "terminate", "TimeoutStartUSec": "infinity", "TimeoutStopFailureMode": "terminate", "TimeoutStopUSec": "1min 30s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "oneshot", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "disabled", "UnitFileState": "generated", "UtmpMode": "init", "WatchdogSignal": "6", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "infinity" } } TASK [linux-system-roles.podman : Restart service] ***************************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/create_update_quadlet_spec.yml:130 Sunday 04 February 2024 20:07:04 +0000 (0:00:00.467) 0:01:19.196 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Set per-container variables part 0] ********** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:14 Sunday 04 February 2024 20:07:04 +0000 (0:00:00.024) 0:01:19.220 ******* ok: [sut] => { "ansible_facts": { "__podman_quadlet_file_src": "", "__podman_quadlet_spec": {}, "__podman_quadlet_str": "[Install]\nWantedBy=default.target\n\n[Container]\nImage=quay.io/linux-system-roles/mysql:5.6\nContainerName=quadlet-demo-mysql\nVolume=quadlet-demo-mysql.volume:/var/lib/mysql\nVolume=/tmp/quadlet_demo:/var/lib/quadlet_demo:Z\nNetwork=quadlet-demo.network\nSecret=mysql-root-password-container,type=env,target=MYSQL_ROOT_PASSWORD\nHealthCmd=/usr/bin/true\nHealthOnFailure=kill\n", "__podman_quadlet_template_src": "quadlet-demo-mysql.container.j2" }, "changed": false } TASK [linux-system-roles.podman : Set per-container variables part 1] ********** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:40 Sunday 04 February 2024 20:07:04 +0000 (0:00:00.074) 0:01:19.295 ******* ok: [sut] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_pull_image": true, "__podman_state": "created", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [linux-system-roles.podman : Fail if no quadlet spec is given] ************ task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:57 Sunday 04 February 2024 20:07:04 +0000 (0:00:00.027) 0:01:19.323 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Set per-container variables part 2] ********** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:70 Sunday 04 February 2024 20:07:04 +0000 (0:00:00.023) 0:01:19.346 ******* ok: [sut] => { "ansible_facts": { "__podman_quadlet_name": "quadlet-demo-mysql", "__podman_quadlet_type": "container", "__podman_rootless": false }, "changed": false } TASK [linux-system-roles.podman : Check user and group information] ************ task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:97 Sunday 04 February 2024 20:07:04 +0000 (0:00:00.056) 0:01:19.402 ******* included: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml for sut TASK [linux-system-roles.podman : Get user information] ************************ task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:2 Sunday 04 February 2024 20:07:04 +0000 (0:00:00.041) 0:01:19.443 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Fail if user does not exist] ***************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:10 Sunday 04 February 2024 20:07:04 +0000 (0:00:00.026) 0:01:19.470 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Set group for podman user] ******************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:17 Sunday 04 February 2024 20:07:05 +0000 (0:00:00.027) 0:01:19.498 ******* ok: [sut] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [linux-system-roles.podman : Get group information] *********************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:28 Sunday 04 February 2024 20:07:05 +0000 (0:00:00.030) 0:01:19.528 ******* ok: [sut] => { "ansible_facts": { "getent_group": { "root": [ "x", "0", "" ] } }, "changed": false } TASK [linux-system-roles.podman : Set group name] ****************************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:36 Sunday 04 February 2024 20:07:05 +0000 (0:00:00.201) 0:01:19.730 ******* ok: [sut] => { "ansible_facts": { "__podman_group_name": "root" }, "changed": false } TASK [linux-system-roles.podman : See if getsubids exists] ********************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:41 Sunday 04 February 2024 20:07:05 +0000 (0:00:00.030) 0:01:19.761 ******* ok: [sut] => { "changed": false, "stat": { "atime": 1707077211.5674798, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "076f89863b474504550ab1b5c11216a77ed7123c", "ctime": 1707077179.8497021, "dev": 51713, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 4380701, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-pie-executable", "mode": "0755", "mtime": 1689166618.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 15496, "uid": 0, "version": "2573075764", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [linux-system-roles.podman : Check user with getsubids] ******************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:52 Sunday 04 February 2024 20:07:05 +0000 (0:00:00.198) 0:01:19.959 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Check group with getsubids] ****************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:56 Sunday 04 February 2024 20:07:05 +0000 (0:00:00.022) 0:01:19.982 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Check if user is in subuid file] ************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:66 Sunday 04 February 2024 20:07:05 +0000 (0:00:00.023) 0:01:20.005 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Fail if user not in subuid file] ************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:74 Sunday 04 February 2024 20:07:05 +0000 (0:00:00.021) 0:01:20.027 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Check if group is in subgid file] ************ task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:81 Sunday 04 February 2024 20:07:05 +0000 (0:00:00.022) 0:01:20.049 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Fail if group not in subgid file] ************ task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:89 Sunday 04 February 2024 20:07:05 +0000 (0:00:00.021) 0:01:20.070 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Set per-container variables part 3] ********** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:102 Sunday 04 February 2024 20:07:05 +0000 (0:00:00.023) 0:01:20.093 ******* ok: [sut] => { "ansible_facts": { "__podman_activate_systemd_unit": true, "__podman_images_found": [ "quay.io/linux-system-roles/mysql:5.6" ], "__podman_kube_yamls_raw": "", "__podman_service_name": "quadlet-demo-mysql.service", "__podman_systemd_scope": "system", "__podman_user_home_dir": "/root", "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [linux-system-roles.podman : Set per-container variables part 4] ********** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:134 Sunday 04 February 2024 20:07:05 +0000 (0:00:00.039) 0:01:20.133 ******* ok: [sut] => { "ansible_facts": { "__podman_quadlet_path": "/etc/containers/systemd" }, "changed": false } TASK [linux-system-roles.podman : Get kube yaml contents] ********************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:140 Sunday 04 February 2024 20:07:05 +0000 (0:00:00.023) 0:01:20.156 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Set per-container variables part 5] ********** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:151 Sunday 04 February 2024 20:07:05 +0000 (0:00:00.022) 0:01:20.179 ******* ok: [sut] => { "ansible_facts": { "__podman_images": [ "quay.io/linux-system-roles/mysql:5.6" ], "__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo-mysql.container", "__podman_volumes": [ "/tmp/quadlet_demo" ] }, "changed": false } TASK [linux-system-roles.podman : Cleanup quadlets] **************************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:205 Sunday 04 February 2024 20:07:05 +0000 (0:00:00.056) 0:01:20.236 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Create and update quadlets] ****************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:209 Sunday 04 February 2024 20:07:05 +0000 (0:00:00.055) 0:01:20.291 ******* included: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/create_update_quadlet_spec.yml for sut TASK [linux-system-roles.podman : Manage linger] ******************************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/create_update_quadlet_spec.yml:2 Sunday 04 February 2024 20:07:05 +0000 (0:00:00.049) 0:01:20.341 ******* included: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/manage_linger.yml for sut TASK [linux-system-roles.podman : Enable linger if needed] ********************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/manage_linger.yml:12 Sunday 04 February 2024 20:07:05 +0000 (0:00:00.040) 0:01:20.381 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Mark user as not yet needing to cancel linger] *** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/manage_linger.yml:18 Sunday 04 February 2024 20:07:05 +0000 (0:00:00.022) 0:01:20.403 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Mark user for possible linger cancel] ******** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/manage_linger.yml:22 Sunday 04 February 2024 20:07:05 +0000 (0:00:00.023) 0:01:20.426 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Create host directories] ********************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/create_update_quadlet_spec.yml:7 Sunday 04 February 2024 20:07:05 +0000 (0:00:00.022) 0:01:20.449 ******* changed: [sut] => (item=/tmp/quadlet_demo) => { "ansible_loop_var": "item", "changed": true, "gid": 0, "group": "root", "item": "/tmp/quadlet_demo", "mode": "0777", "owner": "root", "path": "/tmp/quadlet_demo", "secontext": "unconfined_u:object_r:user_tmp_t:s0", "size": 6, "state": "directory", "uid": 0 } TASK [linux-system-roles.podman : Ensure container images are present] ********* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/create_update_quadlet_spec.yml:25 Sunday 04 February 2024 20:07:06 +0000 (0:00:00.217) 0:01:20.666 ******* changed: [sut] => (item=quay.io/linux-system-roles/mysql:5.6) => { "actions": [ "Pulled image quay.io/linux-system-roles/mysql:5.6" ], "ansible_loop_var": "item", "changed": true, "failed_when_result": false, "image": [ { "Annotations": {}, "Architecture": "amd64", "Author": "", "Comment": "", "Config": { "Cmd": [ "mysqld" ], "Entrypoint": [ "docker-entrypoint.sh" ], "Env": [ "PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin", "GOSU_VERSION=1.12", "MYSQL_MAJOR=5.6", "MYSQL_VERSION=5.6.51-1debian9" ], "ExposedPorts": { "3306/tcp": {} }, "Volumes": { "/var/lib/mysql": {} } }, "Created": "2021-12-21T02:57:59.587337844Z", "Digest": "sha256:2d0419de75076a55e0d0c0680b4f86ecdbb8a88de92054698a0028bcb218a3a8", "GraphDriver": { "Data": { "LowerDir": "/var/lib/containers/storage/overlay/63c8cde32bc583ca7971591c23194c1fd072197949f28fcad7f10780ad1af022/diff:/var/lib/containers/storage/overlay/4904a0ac4db0b6060c944fc7c882cba99b79b50f90151785350c0e4ba7f94390/diff:/var/lib/containers/storage/overlay/1822e40336d4b2b31d888f669f443965d4476dae35fd0a67e58eefd9c5ecb9e5/diff:/var/lib/containers/storage/overlay/988a3979591d296974121b0a893ec621ea5e9939ab3cc5e7c194dbe27f7b5677/diff:/var/lib/containers/storage/overlay/ba0315059f870fcb66ebd6517f455f714fb4eeeb4fdee9cefb3d7f59a7b0cdb7/diff:/var/lib/containers/storage/overlay/0dbdd9024cc38ecc226b2c4772718d32ff0abc288b41669d58d506274c16c67a/diff:/var/lib/containers/storage/overlay/f8ddb99f39386eac409b19f8b347a7a10ee91771cfb838c88793ad8947a233f5/diff:/var/lib/containers/storage/overlay/2ecb8e9889676e4b9b129bab56314644612905ce6e6ce1c92d5a86ac8f08f03b/diff:/var/lib/containers/storage/overlay/07e715cdf99b679f1e067f4819d1b7e1ba9ea24229ede4e6b45b9240fe5865f8/diff:/var/lib/containers/storage/overlay/2b83e5699838047f936a3875bcce7fe1b169983bf86785ae7519c5bc488558ae/diff", "UpperDir": "/var/lib/containers/storage/overlay/bb762056c5cde2e3213fe6ccf595176708249829a607b0afc5799075d4a852ec/diff", "WorkDir": "/var/lib/containers/storage/overlay/bb762056c5cde2e3213fe6ccf595176708249829a607b0afc5799075d4a852ec/work" }, "Name": "overlay" }, "History": [ { "created": "2021-12-21T01:24:41.626621944Z", "created_by": "/bin/sh -c #(nop) ADD file:b1650c19c1dd1f4cff71553f2bb0bc949944d0bc24b54c318b2880c14538648a in / " }, { "created": "2021-12-21T01:24:42.056427889Z", "created_by": "/bin/sh -c #(nop) CMD [\"bash\"]", "empty_layer": true }, { "created": "2021-12-21T02:56:59.289452902Z", "created_by": "/bin/sh -c groupadd -r mysql && useradd -r -g mysql mysql" }, { "created": "2021-12-21T02:57:05.692613282Z", "created_by": "/bin/sh -c apt-get update && apt-get install -y --no-install-recommends gnupg dirmngr && rm -rf /var/lib/apt/lists/*" }, { "created": "2021-12-21T02:57:05.921738618Z", "created_by": "/bin/sh -c #(nop) ENV GOSU_VERSION=1.12", "empty_layer": true }, { "created": "2021-12-21T02:57:18.981216995Z", "created_by": "/bin/sh -c set -eux; \tsavedAptMark=\"$(apt-mark showmanual)\"; \tapt-get update; \tapt-get install -y --no-install-recommends ca-certificates wget; \trm -rf /var/lib/apt/lists/*; \tdpkgArch=\"$(dpkg --print-architecture | awk -F- '{ print $NF }')\"; \twget -O /usr/local/bin/gosu \"https://github.com/tianon/gosu/releases/download/$GOSU_VERSION/gosu-$dpkgArch\"; \twget -O /usr/local/bin/gosu.asc \"https://github.com/tianon/gosu/releases/download/$GOSU_VERSION/gosu-$dpkgArch.asc\"; \texport GNUPGHOME=\"$(mktemp -d)\"; \tgpg --batch --keyserver hkps://keys.openpgp.org --recv-keys B42F6819007F00F88E364FD4036A9C25BF357DD4; \tgpg --batch --verify /usr/local/bin/gosu.asc /usr/local/bin/gosu; \tgpgconf --kill all; \trm -rf \"$GNUPGHOME\" /usr/local/bin/gosu.asc; \tapt-mark auto '.*' > /dev/null; \t[ -z \"$savedAptMark\" ] || apt-mark manual $savedAptMark > /dev/null; \tapt-get purge -y --auto-remove -o APT::AutoRemove::RecommendsImportant=false; \tchmod +x /usr/local/bin/gosu; \tgosu --version; \tgosu nobody true" }, { "created": "2021-12-21T02:57:19.781116502Z", "created_by": "/bin/sh -c mkdir /docker-entrypoint-initdb.d" }, { "created": "2021-12-21T02:57:25.555452451Z", "created_by": "/bin/sh -c apt-get update && apt-get install -y --no-install-recommends \t\tpwgen \t\tperl \t\txz-utils \t&& rm -rf /var/lib/apt/lists/*" }, { "created": "2021-12-21T02:57:37.684730941Z", "created_by": "/bin/sh -c set -ex; \tkey='A4A9406876FCBD3C456770C88C718D3B5072E1F5'; \texport GNUPGHOME=\"$(mktemp -d)\"; \tgpg --batch --keyserver keyserver.ubuntu.com --recv-keys \"$key\"; \tgpg --batch --export \"$key\" > /etc/apt/trusted.gpg.d/mysql.gpg; \tgpgconf --kill all; \trm -rf \"$GNUPGHOME\"; \tapt-key list > /dev/null" }, { "created": "2021-12-21T02:57:37.903115972Z", "created_by": "/bin/sh -c #(nop) ENV MYSQL_MAJOR=5.6", "empty_layer": true }, { "created": "2021-12-21T02:57:38.155528708Z", "created_by": "/bin/sh -c #(nop) ENV MYSQL_VERSION=5.6.51-1debian9", "empty_layer": true }, { "created": "2021-12-21T02:57:39.106901365Z", "created_by": "/bin/sh -c echo 'deb http://repo.mysql.com/apt/debian/ stretch mysql-5.6' > /etc/apt/sources.list.d/mysql.list" }, { "created": "2021-12-21T02:57:56.693791892Z", "created_by": "/bin/sh -c { \t\techo mysql-community-server mysql-community-server/data-dir select ''; \t\techo mysql-community-server mysql-community-server/root-pass password ''; \t\techo mysql-community-server mysql-community-server/re-root-pass password ''; \t\techo mysql-community-server mysql-community-server/remove-test-db select false; \t} | debconf-set-selections \t&& apt-get update \t&& apt-get install -y \t\tmysql-server=\"${MYSQL_VERSION}\" \t&& find /etc/mysql/ -name '*.cnf' -print0 \t\t| xargs -0 grep -lZE '^(bind-address|log)' \t\t| xargs -rt -0 sed -Ei 's/^(bind-address|log)/#&/' \t&& echo '[mysqld]\\nskip-host-cache\\nskip-name-resolve' > /etc/mysql/conf.d/docker.cnf \t&& rm -rf /var/lib/apt/lists/* \t&& rm -rf /var/lib/mysql && mkdir -p /var/lib/mysql /var/run/mysqld \t&& chown -R mysql:mysql /var/lib/mysql /var/run/mysqld \t&& chmod 1777 /var/run/mysqld /var/lib/mysql" }, { "created": "2021-12-21T02:57:57.29721744Z", "created_by": "/bin/sh -c #(nop) VOLUME [/var/lib/mysql]", "empty_layer": true }, { "created": "2021-12-21T02:57:57.591077869Z", "created_by": "/bin/sh -c #(nop) COPY file:345a22fe55d3e6783a17075612415413487e7dba27fbf1000a67c7870364b739 in /usr/local/bin/ " }, { "created": "2021-12-21T02:57:58.761392735Z", "created_by": "/bin/sh -c ln -s usr/local/bin/docker-entrypoint.sh /entrypoint.sh # backwards compat" }, { "created": "2021-12-21T02:57:59.04809683Z", "created_by": "/bin/sh -c #(nop) ENTRYPOINT [\"docker-entrypoint.sh\"]", "empty_layer": true }, { "created": "2021-12-21T02:57:59.31392792Z", "created_by": "/bin/sh -c #(nop) EXPOSE 3306", "empty_layer": true }, { "created": "2021-12-21T02:57:59.587337844Z", "created_by": "/bin/sh -c #(nop) CMD [\"mysqld\"]", "empty_layer": true } ], "Id": "dd3b2a5dcb48ff61113592ed5ddd762581be4387c7bc552375a2159422aa6bf5", "Labels": null, "ManifestType": "application/vnd.docker.distribution.manifest.v2+json", "NamesHistory": [ "quay.io/linux-system-roles/mysql:5.6" ], "Os": "linux", "Parent": "", "RepoDigests": [ "quay.io/linux-system-roles/mysql@sha256:2d0419de75076a55e0d0c0680b4f86ecdbb8a88de92054698a0028bcb218a3a8", "quay.io/linux-system-roles/mysql@sha256:4faa0a30c0055742c72daff1482976814f340d4bfb647f0fa2c3562fd1a7abda" ], "RepoTags": [ "quay.io/linux-system-roles/mysql:5.6" ], "RootFS": { "Layers": [ "sha256:2b83e5699838047f936a3875bcce7fe1b169983bf86785ae7519c5bc488558ae", "sha256:2e10295573911964a8acc37c738aed958d0bc385bc51df6f184b196fb592c975", "sha256:d414fdead0b9c934781e5611c016d7c237062c9470db810ef28b321a492dbb9c", "sha256:4085e588967d10dafb76ad52f66f661b41b9008cc16568ce02853de7cbbf7433", "sha256:7ea96a4e341b4d139f5e076e71d19229537292226c3b24ec2d588567667a1191", "sha256:e3dce1c82d4e4142e8e1bf2fb9c6fd2555196d40088167270ff95c9af37b88af", "sha256:2612088e90f60ecb4a13e68adbfb2b1366f82fc8fbf1f0970f8769cf3cd753ee", "sha256:eba393347f896c4028340bf1a40cc5d6b8f21eeb7abf3e79799d842a6babb93f", "sha256:7c5a5c1986b16b0d8ad4eaa8066627910ec7794a657467e8f18d223f1c0034fd", "sha256:49a1ca1cd2b80f84d6616df2b5123458f4bcb065bca28c43d9c3fd5a0480b907", "sha256:7137327a7221ddb123883146beb68003ee31d652ec46ab8bea3ba12c2df8df8f" ], "Type": "layers" }, "Size": 307607194, "User": "", "Version": "20.10.7", "VirtualSize": 307607194 } ], "item": "quay.io/linux-system-roles/mysql:5.6", "podman_actions": [ "/usr/bin/podman image ls quay.io/linux-system-roles/mysql:5.6 --format json", "/usr/bin/podman image exists quay.io/linux-system-roles/mysql:5.6", "/usr/bin/podman pull quay.io/linux-system-roles/mysql:5.6 -q", "/usr/bin/podman inspect dd3b2a5dcb48ff61113592ed5ddd762581be4387c7bc552375a2159422aa6bf5 --format json" ] } TASK [linux-system-roles.podman : Ensure the quadlet directory is present] ***** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/create_update_quadlet_spec.yml:42 Sunday 04 February 2024 20:07:14 +0000 (0:00:08.332) 0:01:28.999 ******* ok: [sut] => { "changed": false, "gid": 0, "group": "root", "mode": "0755", "owner": "root", "path": "/etc/containers/systemd", "secontext": "system_u:object_r:etc_t:s0", "size": 67, "state": "directory", "uid": 0 } TASK [linux-system-roles.podman : Ensure quadlet file is copied] *************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/create_update_quadlet_spec.yml:51 Sunday 04 February 2024 20:07:14 +0000 (0:00:00.206) 0:01:29.206 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Ensure quadlet file content is present] ****** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/create_update_quadlet_spec.yml:61 Sunday 04 February 2024 20:07:14 +0000 (0:00:00.047) 0:01:29.253 ******* changed: [sut] => { "changed": true, "checksum": "7bf9312d57274d4dabbd222b01717e2cc0cfa0ed", "dest": "/etc/containers/systemd/quadlet-demo-mysql.container", "gid": 0, "group": "root", "md5sum": "f1cd3875deb45b806389825bb4b5ad4a", "mode": "0644", "owner": "root", "secontext": "system_u:object_r:etc_t:s0", "size": 367, "src": "/root/.ansible/tmp/ansible-tmp-1707077234.8014777-10292-186010283769801/source", "state": "file", "uid": 0 } TASK [linux-system-roles.podman : Ensure quadlet file is present] ************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/create_update_quadlet_spec.yml:73 Sunday 04 February 2024 20:07:15 +0000 (0:00:00.514) 0:01:29.768 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Reload systemctl] **************************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/create_update_quadlet_spec.yml:85 Sunday 04 February 2024 20:07:15 +0000 (0:00:00.028) 0:01:29.796 ******* ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.podman : Start service] ******************************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/create_update_quadlet_spec.yml:114 Sunday 04 February 2024 20:07:15 +0000 (0:00:00.526) 0:01:30.323 ******* fatal: [sut]: FAILED! => { "changed": false } MSG: Unable to start service quadlet-demo-mysql.service: Job for quadlet-demo-mysql.service failed because the control process exited with error code. See "systemctl status quadlet-demo-mysql.service" and "journalctl -xeu quadlet-demo-mysql.service" for details. TASK [Dump journal] ************************************************************ task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/tests_quadlet_demo.yml:131 Sunday 04 February 2024 20:07:16 +0000 (0:00:00.495) 0:01:30.818 ******* ok: [sut] => { "changed": false, "cmd": [ "journalctl", "-ex" ], "delta": "0:00:00.036733", "end": "2024-02-04 20:07:16.525408", "rc": 0, "start": "2024-02-04 20:07:16.488675" } STDOUT: Feb 04 19:57:51 localhost systemd-journald[484]: Runtime Journal (/run/log/journal/75ecf5059a924499888af4d7df65fc87) is 8.0M, max 70.6M, 62.6M free. ░░ Subject: Disk space used by the journal ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ Runtime Journal (/run/log/journal/75ecf5059a924499888af4d7df65fc87) is currently using 8.0M. ░░ Maximum allowed usage is set to 70.6M. ░░ Leaving at least 35.3M free (of currently available 690.5M of disk space). ░░ Enforced usage limit is thus 70.6M, of which 62.6M are still available. ░░ ░░ The limits controlling how much disk space is used by the journal may ░░ be configured with SystemMaxUse=, SystemKeepFree=, SystemMaxFileSize=, ░░ RuntimeMaxUse=, RuntimeKeepFree=, RuntimeMaxFileSize= settings in ░░ /etc/systemd/journald.conf. See journald.conf(5) for details. Feb 04 19:57:51 localhost systemd-journald[484]: Received client request to flush runtime journal. Feb 04 19:57:51 localhost systemd[1]: Finished Flush Journal to Persistent Storage. ░░ Subject: A start job for unit systemd-journal-flush.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-journal-flush.service has finished successfully. ░░ ░░ The job identifier is 171. Feb 04 19:57:51 localhost systemd[1]: modprobe@configfs.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit modprobe@configfs.service has successfully entered the 'dead' state. Feb 04 19:57:51 localhost systemd[1]: Finished Load Kernel Module configfs. ░░ Subject: A start job for unit modprobe@configfs.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit modprobe@configfs.service has finished successfully. ░░ ░░ The job identifier is 150. Feb 04 19:57:51 localhost kernel: fuse: init (API version 7.36) Feb 04 19:57:51 localhost systemd[1]: modprobe@fuse.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit modprobe@fuse.service has successfully entered the 'dead' state. Feb 04 19:57:51 localhost systemd[1]: Finished Load Kernel Module fuse. ░░ Subject: A start job for unit modprobe@fuse.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit modprobe@fuse.service has finished successfully. ░░ ░░ The job identifier is 129. Feb 04 19:57:51 localhost kernel: ACPI: bus type drm_connector registered Feb 04 19:57:51 localhost systemd[1]: modprobe@drm.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit modprobe@drm.service has successfully entered the 'dead' state. Feb 04 19:57:51 localhost systemd[1]: Finished Load Kernel Module drm. ░░ Subject: A start job for unit modprobe@drm.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit modprobe@drm.service has finished successfully. ░░ ░░ The job identifier is 206. Feb 04 19:57:52 localhost systemd[1]: Mounting FUSE Control File System... ░░ Subject: A start job for unit sys-fs-fuse-connections.mount has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sys-fs-fuse-connections.mount has begun execution. ░░ ░░ The job identifier is 128. Feb 04 19:57:52 localhost systemd[1]: Mounting Kernel Configuration File System... ░░ Subject: A start job for unit sys-kernel-config.mount has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sys-kernel-config.mount has begun execution. ░░ ░░ The job identifier is 149. Feb 04 19:57:52 localhost systemd[1]: Mounted FUSE Control File System. ░░ Subject: A start job for unit sys-fs-fuse-connections.mount has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sys-fs-fuse-connections.mount has finished successfully. ░░ ░░ The job identifier is 128. Feb 04 19:57:52 localhost systemd[1]: Mounted Kernel Configuration File System. ░░ Subject: A start job for unit sys-kernel-config.mount has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sys-kernel-config.mount has finished successfully. ░░ ░░ The job identifier is 149. Feb 04 19:57:52 localhost systemd[1]: Finished Create Static Device Nodes in /dev. ░░ Subject: A start job for unit systemd-tmpfiles-setup-dev.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-tmpfiles-setup-dev.service has finished successfully. ░░ ░░ The job identifier is 138. Feb 04 19:57:52 localhost systemd[1]: Reached target Preparation for Local File Systems. ░░ Subject: A start job for unit local-fs-pre.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit local-fs-pre.target has finished successfully. ░░ ░░ The job identifier is 142. Feb 04 19:57:52 localhost systemd[1]: Reached target Local File Systems. ░░ Subject: A start job for unit local-fs.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit local-fs.target has finished successfully. ░░ ░░ The job identifier is 140. Feb 04 19:57:52 localhost systemd[1]: Rebuild Dynamic Linker Cache was skipped because no trigger condition checks were met. ░░ Subject: A start job for unit ldconfig.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit ldconfig.service has finished successfully. ░░ ░░ The job identifier is 147. Feb 04 19:57:52 localhost systemd[1]: Mark the need to relabel after reboot was skipped because of an unmet condition check (ConditionSecurity=!selinux). ░░ Subject: A start job for unit selinux-autorelabel-mark.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit selinux-autorelabel-mark.service has finished successfully. ░░ ░░ The job identifier is 173. Feb 04 19:57:52 localhost systemd[1]: Set Up Additional Binary Formats was skipped because no trigger condition checks were met. ░░ Subject: A start job for unit systemd-binfmt.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-binfmt.service has finished successfully. ░░ ░░ The job identifier is 164. Feb 04 19:57:52 localhost systemd[1]: Store a System Token in an EFI Variable was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/LoaderFeatures-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). ░░ Subject: A start job for unit systemd-boot-system-token.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-boot-system-token.service has finished successfully. ░░ ░░ The job identifier is 133. Feb 04 19:57:52 localhost systemd[1]: Starting Automatic Boot Loader Update... ░░ Subject: A start job for unit systemd-boot-update.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-boot-update.service has begun execution. ░░ ░░ The job identifier is 159. Feb 04 19:57:52 localhost systemd[1]: Starting Commit a transient machine-id on disk... ░░ Subject: A start job for unit systemd-machine-id-commit.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-machine-id-commit.service has begun execution. ░░ ░░ The job identifier is 163. Feb 04 19:57:52 localhost systemd[1]: Starting Create Volatile Files and Directories... ░░ Subject: A start job for unit systemd-tmpfiles-setup.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-tmpfiles-setup.service has begun execution. ░░ ░░ The job identifier is 154. Feb 04 19:57:52 localhost systemd[1]: Starting Rule-based Manager for Device Events and Files... ░░ Subject: A start job for unit systemd-udevd.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-udevd.service has begun execution. ░░ ░░ The job identifier is 122. Feb 04 19:57:52 localhost bootctl[497]: Couldn't find EFI system partition, skipping. Feb 04 19:57:52 localhost systemd[1]: Finished Automatic Boot Loader Update. ░░ Subject: A start job for unit systemd-boot-update.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-boot-update.service has finished successfully. ░░ ░░ The job identifier is 159. Feb 04 19:57:52 localhost systemd[1]: etc-machine\x2did.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit etc-machine\x2did.mount has successfully entered the 'dead' state. Feb 04 19:57:52 localhost systemd[1]: Finished Commit a transient machine-id on disk. ░░ Subject: A start job for unit systemd-machine-id-commit.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-machine-id-commit.service has finished successfully. ░░ ░░ The job identifier is 163. Feb 04 19:57:52 localhost systemd-udevd[500]: Using default interface naming scheme 'rhel-9.0'. Feb 04 19:57:52 localhost systemd[1]: Finished Create Volatile Files and Directories. ░░ Subject: A start job for unit systemd-tmpfiles-setup.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-tmpfiles-setup.service has finished successfully. ░░ ░░ The job identifier is 154. Feb 04 19:57:52 localhost systemd[1]: Mounting RPC Pipe File System... ░░ Subject: A start job for unit var-lib-nfs-rpc_pipefs.mount has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit var-lib-nfs-rpc_pipefs.mount has begun execution. ░░ ░░ The job identifier is 234. Feb 04 19:57:52 localhost systemd[1]: Starting Security Auditing Service... ░░ Subject: A start job for unit auditd.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit auditd.service has begun execution. ░░ ░░ The job identifier is 220. Feb 04 19:57:52 localhost systemd[1]: Starting RPC Bind... ░░ Subject: A start job for unit rpcbind.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit rpcbind.service has begun execution. ░░ ░░ The job identifier is 221. Feb 04 19:57:52 localhost systemd[1]: Rebuild Journal Catalog was skipped because of an unmet condition check (ConditionNeedsUpdate=/var). ░░ Subject: A start job for unit systemd-journal-catalog-update.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-journal-catalog-update.service has finished successfully. ░░ ░░ The job identifier is 135. Feb 04 19:57:52 localhost systemd[1]: Update is Completed was skipped because no trigger condition checks were met. ░░ Subject: A start job for unit systemd-update-done.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-update-done.service has finished successfully. ░░ ░░ The job identifier is 143. Feb 04 19:57:52 localhost kernel: RPC: Registered named UNIX socket transport module. Feb 04 19:57:52 localhost kernel: RPC: Registered udp transport module. Feb 04 19:57:52 localhost kernel: RPC: Registered tcp transport module. Feb 04 19:57:52 localhost kernel: RPC: Registered tcp-with-tls transport module. Feb 04 19:57:52 localhost kernel: RPC: Registered tcp NFSv4.1 backchannel transport module. Feb 04 19:57:52 localhost systemd[1]: Mounted RPC Pipe File System. ░░ Subject: A start job for unit var-lib-nfs-rpc_pipefs.mount has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit var-lib-nfs-rpc_pipefs.mount has finished successfully. ░░ ░░ The job identifier is 234. Feb 04 19:57:52 localhost systemd[1]: Reached target rpc_pipefs.target. ░░ Subject: A start job for unit rpc_pipefs.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit rpc_pipefs.target has finished successfully. ░░ ░░ The job identifier is 233. Feb 04 19:57:52 localhost systemd[1]: Started Rule-based Manager for Device Events and Files. ░░ Subject: A start job for unit systemd-udevd.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-udevd.service has finished successfully. ░░ ░░ The job identifier is 122. Feb 04 19:57:52 localhost systemd[1]: Starting Load Kernel Module configfs... ░░ Subject: A start job for unit modprobe@configfs.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit modprobe@configfs.service has begun execution. ░░ ░░ The job identifier is 257. Feb 04 19:57:52 localhost systemd[1]: modprobe@configfs.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit modprobe@configfs.service has successfully entered the 'dead' state. Feb 04 19:57:52 localhost systemd[1]: Finished Load Kernel Module configfs. ░░ Subject: A start job for unit modprobe@configfs.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit modprobe@configfs.service has finished successfully. ░░ ░░ The job identifier is 257. Feb 04 19:57:52 localhost kernel: input: PC Speaker as /devices/platform/pcspkr/input/input5 Feb 04 19:57:53 localhost kernel: piix4_smbus 0000:00:01.3: SMBus base address uninitialized - upgrade BIOS or use force_addr=0xaddr Feb 04 19:57:53 localhost kernel: RAPL PMU: API unit is 2^-32 Joules, 0 fixed counters, 655360 ms ovfl timer Feb 04 19:57:53 localhost systemd-udevd[525]: Network interface NamePolicy= disabled on kernel command line. Feb 04 19:57:53 localhost auditd[539]: No plugins found, not dispatching events Feb 04 19:57:53 localhost systemd[1]: Started RPC Bind. ░░ Subject: A start job for unit rpcbind.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit rpcbind.service has finished successfully. ░░ ░░ The job identifier is 221. Feb 04 19:57:53 localhost auditd[539]: Init complete, auditd 3.1.2 listening for events (startup state enable) Feb 04 19:57:53 localhost kernel: cirrus 0000:00:02.0: vgaarb: deactivate vga console Feb 04 19:57:53 localhost kernel: Console: switching to colour dummy device 80x25 Feb 04 19:57:53 localhost kernel: [drm] Initialized cirrus 2.0.0 2019 for 0000:00:02.0 on minor 0 Feb 04 19:57:53 localhost kernel: fbcon: cirrusdrmfb (fb0) is primary device Feb 04 19:57:53 localhost kernel: Console: switching to colour frame buffer device 128x48 Feb 04 19:57:53 localhost kernel: cirrus 0000:00:02.0: [drm] fb0: cirrusdrmfb frame buffer device Feb 04 19:57:53 localhost augenrules[542]: /sbin/augenrules: No change Feb 04 19:57:53 localhost augenrules[569]: No rules Feb 04 19:57:53 localhost augenrules[569]: enabled 1 Feb 04 19:57:53 localhost augenrules[569]: failure 1 Feb 04 19:57:53 localhost augenrules[569]: pid 539 Feb 04 19:57:53 localhost augenrules[569]: rate_limit 0 Feb 04 19:57:53 localhost augenrules[569]: backlog_limit 8192 Feb 04 19:57:53 localhost augenrules[569]: lost 0 Feb 04 19:57:53 localhost augenrules[569]: backlog 0 Feb 04 19:57:53 localhost augenrules[569]: backlog_wait_time 60000 Feb 04 19:57:53 localhost augenrules[569]: backlog_wait_time_actual 0 Feb 04 19:57:53 localhost augenrules[569]: enabled 1 Feb 04 19:57:53 localhost augenrules[569]: failure 1 Feb 04 19:57:53 localhost augenrules[569]: pid 539 Feb 04 19:57:53 localhost augenrules[569]: rate_limit 0 Feb 04 19:57:53 localhost augenrules[569]: backlog_limit 8192 Feb 04 19:57:53 localhost augenrules[569]: lost 0 Feb 04 19:57:53 localhost augenrules[569]: backlog 4 Feb 04 19:57:53 localhost augenrules[569]: backlog_wait_time 60000 Feb 04 19:57:53 localhost augenrules[569]: backlog_wait_time_actual 0 Feb 04 19:57:53 localhost augenrules[569]: enabled 1 Feb 04 19:57:53 localhost augenrules[569]: failure 1 Feb 04 19:57:53 localhost augenrules[569]: pid 539 Feb 04 19:57:53 localhost augenrules[569]: rate_limit 0 Feb 04 19:57:53 localhost augenrules[569]: backlog_limit 8192 Feb 04 19:57:53 localhost augenrules[569]: lost 0 Feb 04 19:57:53 localhost augenrules[569]: backlog 4 Feb 04 19:57:53 localhost augenrules[569]: backlog_wait_time 60000 Feb 04 19:57:53 localhost augenrules[569]: backlog_wait_time_actual 0 Feb 04 19:57:53 localhost systemd[1]: Started Security Auditing Service. ░░ Subject: A start job for unit auditd.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit auditd.service has finished successfully. ░░ ░░ The job identifier is 220. Feb 04 19:57:53 localhost systemd[1]: Starting Record System Boot/Shutdown in UTMP... ░░ Subject: A start job for unit systemd-update-utmp.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-update-utmp.service has begun execution. ░░ ░░ The job identifier is 223. Feb 04 19:57:53 localhost systemd[1]: Finished Record System Boot/Shutdown in UTMP. ░░ Subject: A start job for unit systemd-update-utmp.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-update-utmp.service has finished successfully. ░░ ░░ The job identifier is 223. Feb 04 19:57:53 localhost systemd[1]: Reached target System Initialization. ░░ Subject: A start job for unit sysinit.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sysinit.target has finished successfully. ░░ ░░ The job identifier is 118. Feb 04 19:57:53 localhost systemd[1]: Started dnf makecache --timer. ░░ Subject: A start job for unit dnf-makecache.timer has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit dnf-makecache.timer has finished successfully. ░░ ░░ The job identifier is 182. Feb 04 19:57:53 localhost systemd[1]: Started Daily rotation of log files. ░░ Subject: A start job for unit logrotate.timer has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit logrotate.timer has finished successfully. ░░ ░░ The job identifier is 181. Feb 04 19:57:53 localhost systemd[1]: Started Daily Cleanup of Temporary Directories. ░░ Subject: A start job for unit systemd-tmpfiles-clean.timer has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-tmpfiles-clean.timer has finished successfully. ░░ ░░ The job identifier is 180. Feb 04 19:57:53 localhost systemd[1]: Reached target Timer Units. ░░ Subject: A start job for unit timers.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit timers.target has finished successfully. ░░ ░░ The job identifier is 179. Feb 04 19:57:53 localhost systemd[1]: Listening on D-Bus System Message Bus Socket. ░░ Subject: A start job for unit dbus.socket has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit dbus.socket has finished successfully. ░░ ░░ The job identifier is 186. Feb 04 19:57:53 localhost systemd[1]: Listening on SSSD Kerberos Cache Manager responder socket. ░░ Subject: A start job for unit sssd-kcm.socket has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sssd-kcm.socket has finished successfully. ░░ ░░ The job identifier is 193. Feb 04 19:57:53 localhost systemd[1]: Reached target Socket Units. ░░ Subject: A start job for unit sockets.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sockets.target has finished successfully. ░░ ░░ The job identifier is 189. Feb 04 19:57:54 localhost systemd[1]: Starting D-Bus System Message Bus... ░░ Subject: A start job for unit dbus-broker.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit dbus-broker.service has begun execution. ░░ ░░ The job identifier is 187. Feb 04 19:57:54 localhost systemd[1]: TPM2 PCR Barrier (Initialization) was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/StubPcrKernelImage-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). ░░ Subject: A start job for unit systemd-pcrphase-sysinit.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-pcrphase-sysinit.service has finished successfully. ░░ ░░ The job identifier is 166. Feb 04 19:57:54 localhost systemd[1]: Started D-Bus System Message Bus. ░░ Subject: A start job for unit dbus-broker.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit dbus-broker.service has finished successfully. ░░ ░░ The job identifier is 187. Feb 04 19:57:54 localhost systemd[1]: Reached target Basic System. ░░ Subject: A start job for unit basic.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit basic.target has finished successfully. ░░ ░░ The job identifier is 117. Feb 04 19:57:54 localhost systemd[1]: Starting NTP client/server... ░░ Subject: A start job for unit chronyd.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit chronyd.service has begun execution. ░░ ░░ The job identifier is 239. Feb 04 19:57:54 localhost systemd[1]: Starting Initial cloud-init job (pre-networking)... ░░ Subject: A start job for unit cloud-init-local.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit cloud-init-local.service has begun execution. ░░ ░░ The job identifier is 219. Feb 04 19:57:54 localhost systemd[1]: Starting Restore /run/initramfs on shutdown... ░░ Subject: A start job for unit dracut-shutdown.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit dracut-shutdown.service has begun execution. ░░ ░░ The job identifier is 139. Feb 04 19:57:54 localhost systemd[1]: Started irqbalance daemon. ░░ Subject: A start job for unit irqbalance.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit irqbalance.service has finished successfully. ░░ ░░ The job identifier is 242. Feb 04 19:57:54 localhost systemd[1]: Load CPU microcode update was skipped because of an unmet condition check (ConditionPathExists=/sys/devices/system/cpu/microcode/reload). ░░ Subject: A start job for unit microcode.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit microcode.service has finished successfully. ░░ ░░ The job identifier is 197. Feb 04 19:57:54 localhost systemd[1]: Started Hardware RNG Entropy Gatherer Daemon. ░░ Subject: A start job for unit rngd.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit rngd.service has finished successfully. ░░ ░░ The job identifier is 225. Feb 04 19:57:54 localhost systemd[1]: OpenSSH ecdsa Server Key Generation was skipped because of an unmet condition check (ConditionPathExists=!/run/systemd/generator.early/multi-user.target.wants/cloud-init.target). ░░ Subject: A start job for unit sshd-keygen@ecdsa.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sshd-keygen@ecdsa.service has finished successfully. ░░ ░░ The job identifier is 217. Feb 04 19:57:54 localhost systemd[1]: OpenSSH ed25519 Server Key Generation was skipped because of an unmet condition check (ConditionPathExists=!/run/systemd/generator.early/multi-user.target.wants/cloud-init.target). ░░ Subject: A start job for unit sshd-keygen@ed25519.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sshd-keygen@ed25519.service has finished successfully. ░░ ░░ The job identifier is 215. Feb 04 19:57:54 localhost systemd[1]: OpenSSH rsa Server Key Generation was skipped because of an unmet condition check (ConditionPathExists=!/run/systemd/generator.early/multi-user.target.wants/cloud-init.target). ░░ Subject: A start job for unit sshd-keygen@rsa.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sshd-keygen@rsa.service has finished successfully. ░░ ░░ The job identifier is 218. Feb 04 19:57:54 localhost systemd[1]: Reached target sshd-keygen.target. ░░ Subject: A start job for unit sshd-keygen.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sshd-keygen.target has finished successfully. ░░ ░░ The job identifier is 214. Feb 04 19:57:54 localhost systemd[1]: System Security Services Daemon was skipped because no trigger condition checks were met. ░░ Subject: A start job for unit sssd.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sssd.service has finished successfully. ░░ ░░ The job identifier is 237. Feb 04 19:57:54 localhost systemd[1]: Reached target User and Group Name Lookups. ░░ Subject: A start job for unit nss-user-lookup.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit nss-user-lookup.target has finished successfully. ░░ ░░ The job identifier is 238. Feb 04 19:57:54 localhost systemd[1]: Starting User Login Management... ░░ Subject: A start job for unit systemd-logind.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-logind.service has begun execution. ░░ ░░ The job identifier is 205. Feb 04 19:57:54 localhost systemd-logind[584]: New seat seat0. ░░ Subject: A new seat seat0 is now available ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ Documentation: sd-login(3) ░░ ░░ A new seat seat0 has been configured and is now available. Feb 04 19:57:54 localhost systemd-logind[584]: Watching system buttons on /dev/input/event0 (Power Button) Feb 04 19:57:54 localhost systemd-logind[584]: Watching system buttons on /dev/input/event1 (Sleep Button) Feb 04 19:57:54 localhost systemd-logind[584]: Watching system buttons on /dev/input/event2 (AT Translated Set 2 keyboard) Feb 04 19:57:54 localhost dbus-broker-lau[577]: Ready Feb 04 19:57:54 localhost systemd[1]: Starting Rotate log files... ░░ Subject: A start job for unit logrotate.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit logrotate.service has begun execution. ░░ ░░ The job identifier is 263. Feb 04 19:57:54 localhost systemd[1]: Started User Login Management. ░░ Subject: A start job for unit systemd-logind.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-logind.service has finished successfully. ░░ ░░ The job identifier is 205. Feb 04 19:57:54 localhost systemd[1]: Finished Restore /run/initramfs on shutdown. ░░ Subject: A start job for unit dracut-shutdown.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit dracut-shutdown.service has finished successfully. ░░ ░░ The job identifier is 139. Feb 04 19:57:55 localhost systemd[1]: logrotate.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit logrotate.service has successfully entered the 'dead' state. Feb 04 19:57:55 localhost systemd[1]: Finished Rotate log files. ░░ Subject: A start job for unit logrotate.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit logrotate.service has finished successfully. ░░ ░░ The job identifier is 263. Feb 04 19:57:55 localhost rngd[583]: Disabling 7: PKCS11 Entropy generator (pkcs11) Feb 04 19:57:55 localhost rngd[583]: Disabling 5: NIST Network Entropy Beacon (nist) Feb 04 19:57:55 localhost rngd[583]: Disabling 9: Qrypt quantum entropy beacon (qrypt) Feb 04 19:57:55 localhost rngd[583]: Initializing available sources Feb 04 19:57:55 localhost rngd[583]: [hwrng ]: Initialization Failed Feb 04 19:57:55 localhost rngd[583]: [rdrand]: Enabling RDRAND rng support Feb 04 19:57:55 localhost rngd[583]: [rdrand]: Initialized Feb 04 19:57:55 localhost rngd[583]: [jitter]: JITTER timeout set to 5 sec Feb 04 19:57:55 localhost rngd[583]: [jitter]: Initializing AES buffer Feb 04 19:57:55 localhost chronyd[592]: chronyd version 4.3 starting (+CMDMON +NTP +REFCLOCK +RTC +PRIVDROP +SCFILTER +SIGND +ASYNCDNS +NTS +SECHASH +IPV6 +DEBUG) Feb 04 19:57:55 localhost chronyd[592]: Frequency 0.000 +/- 1000000.000 ppm read from /var/lib/chrony/drift Feb 04 19:57:55 localhost chronyd[592]: Using right/UTC timezone to obtain leap second data Feb 04 19:57:55 localhost chronyd[592]: Loaded seccomp filter (level 2) Feb 04 19:57:55 localhost systemd[1]: Started NTP client/server. ░░ Subject: A start job for unit chronyd.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit chronyd.service has finished successfully. ░░ ░░ The job identifier is 239. Feb 04 19:58:00 localhost rngd[583]: [jitter]: Unable to obtain AES key, disabling JITTER source Feb 04 19:58:00 localhost rngd[583]: [jitter]: Initialization Failed Feb 04 19:58:00 localhost rngd[583]: Process privileges have been dropped to 2:2 Feb 04 19:58:01 localhost cloud-init[596]: Cloud-init v. 23.4-2.el9 running 'init-local' at Sun, 04 Feb 2024 19:58:01 +0000. Up 23.94 seconds. Feb 04 19:58:01 localhost dhclient[599]: Internet Systems Consortium DHCP Client 4.4.2b1 Feb 04 19:58:01 localhost dhclient[599]: Copyright 2004-2019 Internet Systems Consortium. Feb 04 19:58:01 localhost dhclient[599]: All rights reserved. Feb 04 19:58:01 localhost dhclient[599]: For info, please visit https://www.isc.org/software/dhcp/ Feb 04 19:58:01 localhost dhclient[599]: Feb 04 19:58:01 localhost dhclient[599]: Listening on LPF/eth0/12:cd:2e:aa:47:c5 Feb 04 19:58:01 localhost dhclient[599]: Sending on LPF/eth0/12:cd:2e:aa:47:c5 Feb 04 19:58:01 localhost dhclient[599]: Sending on Socket/fallback Feb 04 19:58:01 localhost dhclient[599]: DHCPDISCOVER on eth0 to 255.255.255.255 port 67 interval 6 (xid=0xdabd774c) Feb 04 19:58:01 localhost dhclient[599]: DHCPOFFER of 10.31.9.63 from 10.31.8.1 Feb 04 19:58:01 localhost dhclient[599]: DHCPREQUEST for 10.31.9.63 on eth0 to 255.255.255.255 port 67 (xid=0xdabd774c) Feb 04 19:58:01 localhost dhclient[599]: DHCPACK of 10.31.9.63 from 10.31.8.1 (xid=0xdabd774c) Feb 04 19:58:01 localhost dhclient[599]: bound to 10.31.9.63 -- renewal in 1603 seconds. Feb 04 19:58:02 localhost systemd[1]: Starting Hostname Service... ░░ Subject: A start job for unit systemd-hostnamed.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-hostnamed.service has begun execution. ░░ ░░ The job identifier is 325. Feb 04 19:58:02 localhost systemd[1]: Started Hostname Service. ░░ Subject: A start job for unit systemd-hostnamed.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-hostnamed.service has finished successfully. ░░ ░░ The job identifier is 325. Feb 04 19:58:02 ip-10-31-9-63.us-east-1.aws.redhat.com systemd-hostnamed[614]: Hostname set to (static) Feb 04 19:58:02 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: Finished Initial cloud-init job (pre-networking). ░░ Subject: A start job for unit cloud-init-local.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit cloud-init-local.service has finished successfully. ░░ ░░ The job identifier is 219. Feb 04 19:58:02 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: Reached target Preparation for Network. ░░ Subject: A start job for unit network-pre.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit network-pre.target has finished successfully. ░░ ░░ The job identifier is 145. Feb 04 19:58:02 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: Starting Network Manager... ░░ Subject: A start job for unit NetworkManager.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit NetworkManager.service has begun execution. ░░ ░░ The job identifier is 185. Feb 04 19:58:02 ip-10-31-9-63.us-east-1.aws.redhat.com NetworkManager[618]: [1707076682.9270] NetworkManager (version 1.45.10-1.el9) is starting... (boot:6cbc375d-0231-4045-8667-be09e8f214b8) Feb 04 19:58:02 ip-10-31-9-63.us-east-1.aws.redhat.com NetworkManager[618]: [1707076682.9273] Read config: /etc/NetworkManager/NetworkManager.conf (run: 15-carrier-timeout.conf) Feb 04 19:58:03 ip-10-31-9-63.us-east-1.aws.redhat.com NetworkManager[618]: [1707076683.0359] manager[0x555eacded040]: monitoring kernel firmware directory '/lib/firmware'. Feb 04 19:58:03 ip-10-31-9-63.us-east-1.aws.redhat.com NetworkManager[618]: [1707076683.0385] hostname: hostname: using hostnamed Feb 04 19:58:03 ip-10-31-9-63.us-east-1.aws.redhat.com NetworkManager[618]: [1707076683.0386] hostname: static hostname changed from (none) to "ip-10-31-9-63.us-east-1.aws.redhat.com" Feb 04 19:58:03 ip-10-31-9-63.us-east-1.aws.redhat.com NetworkManager[618]: [1707076683.0400] dns-mgr: init: dns=default,systemd-resolved rc-manager=symlink (auto) Feb 04 19:58:03 ip-10-31-9-63.us-east-1.aws.redhat.com NetworkManager[618]: [1707076683.1082] manager[0x555eacded040]: rfkill: Wi-Fi hardware radio set enabled Feb 04 19:58:03 ip-10-31-9-63.us-east-1.aws.redhat.com NetworkManager[618]: [1707076683.1082] manager[0x555eacded040]: rfkill: WWAN hardware radio set enabled Feb 04 19:58:03 ip-10-31-9-63.us-east-1.aws.redhat.com NetworkManager[618]: [1707076683.1186] Loaded device plugin: NMTeamFactory (/usr/lib64/NetworkManager/1.45.10-1.el9/libnm-device-plugin-team.so) Feb 04 19:58:03 ip-10-31-9-63.us-east-1.aws.redhat.com NetworkManager[618]: [1707076683.1186] manager: rfkill: Wi-Fi enabled by radio killswitch; enabled by state file Feb 04 19:58:03 ip-10-31-9-63.us-east-1.aws.redhat.com NetworkManager[618]: [1707076683.1194] manager: rfkill: WWAN enabled by radio killswitch; enabled by state file Feb 04 19:58:03 ip-10-31-9-63.us-east-1.aws.redhat.com NetworkManager[618]: [1707076683.1196] manager: Networking is enabled by state file Feb 04 19:58:03 ip-10-31-9-63.us-east-1.aws.redhat.com NetworkManager[618]: [1707076683.1239] settings: Loaded settings plugin: keyfile (internal) Feb 04 19:58:03 ip-10-31-9-63.us-east-1.aws.redhat.com NetworkManager[618]: [1707076683.1276] settings: Loaded settings plugin: ifcfg-rh ("/usr/lib64/NetworkManager/1.45.10-1.el9/libnm-settings-plugin-ifcfg-rh.so") Feb 04 19:58:03 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: Listening on Load/Save RF Kill Switch Status /dev/rfkill Watch. ░░ Subject: A start job for unit systemd-rfkill.socket has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-rfkill.socket has finished successfully. ░░ ░░ The job identifier is 452. Feb 04 19:58:03 ip-10-31-9-63.us-east-1.aws.redhat.com NetworkManager[618]: [1707076683.1312] Warning: the ifcfg-rh plugin is deprecated, please migrate connections to the keyfile format using "nmcli connection migrate" Feb 04 19:58:03 ip-10-31-9-63.us-east-1.aws.redhat.com NetworkManager[618]: [1707076683.1329] dhcp: init: Using DHCP client 'internal' Feb 04 19:58:03 ip-10-31-9-63.us-east-1.aws.redhat.com NetworkManager[618]: [1707076683.1331] manager: (lo): new Loopback device (/org/freedesktop/NetworkManager/Devices/1) Feb 04 19:58:03 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: Starting Network Manager Script Dispatcher Service... ░░ Subject: A start job for unit NetworkManager-dispatcher.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit NetworkManager-dispatcher.service has begun execution. ░░ ░░ The job identifier is 389. Feb 04 19:58:03 ip-10-31-9-63.us-east-1.aws.redhat.com NetworkManager[618]: [1707076683.1357] device (lo): state change: unmanaged -> unavailable (reason 'connection-assumed', sys-iface-state: 'external') Feb 04 19:58:03 ip-10-31-9-63.us-east-1.aws.redhat.com NetworkManager[618]: [1707076683.1364] device (lo): state change: unavailable -> disconnected (reason 'connection-assumed', sys-iface-state: 'external') Feb 04 19:58:03 ip-10-31-9-63.us-east-1.aws.redhat.com NetworkManager[618]: [1707076683.1373] device (lo): Activation: starting connection 'lo' (b436c604-992f-4c75-a764-8f54b9f33e11) Feb 04 19:58:03 ip-10-31-9-63.us-east-1.aws.redhat.com NetworkManager[618]: [1707076683.1381] manager: (eth0): new Ethernet device (/org/freedesktop/NetworkManager/Devices/2) Feb 04 19:58:03 ip-10-31-9-63.us-east-1.aws.redhat.com NetworkManager[618]: [1707076683.1386] device (eth0): state change: unmanaged -> unavailable (reason 'managed', sys-iface-state: 'external') Feb 04 19:58:03 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: Started Network Manager. ░░ Subject: A start job for unit NetworkManager.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit NetworkManager.service has finished successfully. ░░ ░░ The job identifier is 185. Feb 04 19:58:03 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: Reached target Network. ░░ Subject: A start job for unit network.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit network.target has finished successfully. ░░ ░░ The job identifier is 188. Feb 04 19:58:03 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: Starting Network Manager Wait Online... ░░ Subject: A start job for unit NetworkManager-wait-online.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit NetworkManager-wait-online.service has begun execution. ░░ ░░ The job identifier is 184. Feb 04 19:58:03 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: Starting GSSAPI Proxy Daemon... ░░ Subject: A start job for unit gssproxy.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit gssproxy.service has begun execution. ░░ ░░ The job identifier is 231. Feb 04 19:58:03 ip-10-31-9-63.us-east-1.aws.redhat.com NetworkManager[618]: [1707076683.1443] bus-manager: acquired D-Bus service "org.freedesktop.NetworkManager" Feb 04 19:58:03 ip-10-31-9-63.us-east-1.aws.redhat.com NetworkManager[618]: [1707076683.1455] device (lo): state change: disconnected -> prepare (reason 'none', sys-iface-state: 'external') Feb 04 19:58:03 ip-10-31-9-63.us-east-1.aws.redhat.com NetworkManager[618]: [1707076683.1461] device (lo): state change: prepare -> config (reason 'none', sys-iface-state: 'external') Feb 04 19:58:03 ip-10-31-9-63.us-east-1.aws.redhat.com NetworkManager[618]: [1707076683.1466] device (lo): state change: config -> ip-config (reason 'none', sys-iface-state: 'external') Feb 04 19:58:03 ip-10-31-9-63.us-east-1.aws.redhat.com NetworkManager[618]: [1707076683.1470] device (eth0): carrier: link connected Feb 04 19:58:03 ip-10-31-9-63.us-east-1.aws.redhat.com NetworkManager[618]: [1707076683.1479] device (lo): state change: ip-config -> ip-check (reason 'none', sys-iface-state: 'external') Feb 04 19:58:03 ip-10-31-9-63.us-east-1.aws.redhat.com NetworkManager[618]: [1707076683.1486] device (eth0): state change: unavailable -> disconnected (reason 'carrier-changed', sys-iface-state: 'managed') Feb 04 19:58:03 ip-10-31-9-63.us-east-1.aws.redhat.com NetworkManager[618]: [1707076683.1497] policy: auto-activating connection 'System eth0' (5fb06bd0-0bb0-7ffb-45f1-d6edd65f3e03) Feb 04 19:58:03 ip-10-31-9-63.us-east-1.aws.redhat.com NetworkManager[618]: [1707076683.1501] device (eth0): Activation: starting connection 'System eth0' (5fb06bd0-0bb0-7ffb-45f1-d6edd65f3e03) Feb 04 19:58:03 ip-10-31-9-63.us-east-1.aws.redhat.com NetworkManager[618]: [1707076683.1504] device (eth0): state change: disconnected -> prepare (reason 'none', sys-iface-state: 'managed') Feb 04 19:58:03 ip-10-31-9-63.us-east-1.aws.redhat.com NetworkManager[618]: [1707076683.1506] manager: NetworkManager state is now CONNECTING Feb 04 19:58:03 ip-10-31-9-63.us-east-1.aws.redhat.com NetworkManager[618]: [1707076683.1508] device (eth0): state change: prepare -> config (reason 'none', sys-iface-state: 'managed') Feb 04 19:58:03 ip-10-31-9-63.us-east-1.aws.redhat.com NetworkManager[618]: [1707076683.1514] device (eth0): state change: config -> ip-config (reason 'none', sys-iface-state: 'managed') Feb 04 19:58:03 ip-10-31-9-63.us-east-1.aws.redhat.com NetworkManager[618]: [1707076683.1517] dhcp4 (eth0): activation: beginning transaction (timeout in 45 seconds) Feb 04 19:58:03 ip-10-31-9-63.us-east-1.aws.redhat.com NetworkManager[618]: [1707076683.1533] dhcp4 (eth0): state changed new lease, address=10.31.9.63 Feb 04 19:58:03 ip-10-31-9-63.us-east-1.aws.redhat.com NetworkManager[618]: [1707076683.1536] policy: set 'System eth0' (eth0) as default for IPv4 routing and DNS Feb 04 19:58:03 ip-10-31-9-63.us-east-1.aws.redhat.com NetworkManager[618]: [1707076683.1612] device (eth0): state change: ip-config -> ip-check (reason 'none', sys-iface-state: 'managed') Feb 04 19:58:03 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: Started Network Manager Script Dispatcher Service. ░░ Subject: A start job for unit NetworkManager-dispatcher.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit NetworkManager-dispatcher.service has finished successfully. ░░ ░░ The job identifier is 389. Feb 04 19:58:03 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: Started GSSAPI Proxy Daemon. ░░ Subject: A start job for unit gssproxy.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit gssproxy.service has finished successfully. ░░ ░░ The job identifier is 231. Feb 04 19:58:03 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: RPC security service for NFS client and server was skipped because of an unmet condition check (ConditionPathExists=/etc/krb5.keytab). ░░ Subject: A start job for unit rpc-gssd.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit rpc-gssd.service has finished successfully. ░░ ░░ The job identifier is 232. Feb 04 19:58:03 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: Reached target NFS client services. ░░ Subject: A start job for unit nfs-client.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit nfs-client.target has finished successfully. ░░ ░░ The job identifier is 227. Feb 04 19:58:03 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: Reached target Preparation for Remote File Systems. ░░ Subject: A start job for unit remote-fs-pre.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit remote-fs-pre.target has finished successfully. ░░ ░░ The job identifier is 228. Feb 04 19:58:03 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: Reached target Remote File Systems. ░░ Subject: A start job for unit remote-fs.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit remote-fs.target has finished successfully. ░░ ░░ The job identifier is 226. Feb 04 19:58:03 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: TPM2 PCR Barrier (User) was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/StubPcrKernelImage-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). ░░ Subject: A start job for unit systemd-pcrphase.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-pcrphase.service has finished successfully. ░░ ░░ The job identifier is 151. Feb 04 19:58:03 ip-10-31-9-63.us-east-1.aws.redhat.com NetworkManager[618]: [1707076683.4340] device (lo): state change: ip-check -> secondaries (reason 'none', sys-iface-state: 'external') Feb 04 19:58:03 ip-10-31-9-63.us-east-1.aws.redhat.com NetworkManager[618]: [1707076683.4342] device (eth0): state change: ip-check -> secondaries (reason 'none', sys-iface-state: 'managed') Feb 04 19:58:03 ip-10-31-9-63.us-east-1.aws.redhat.com NetworkManager[618]: [1707076683.4345] device (lo): state change: secondaries -> activated (reason 'none', sys-iface-state: 'external') Feb 04 19:58:03 ip-10-31-9-63.us-east-1.aws.redhat.com NetworkManager[618]: [1707076683.4351] device (lo): Activation: successful, device activated. Feb 04 19:58:03 ip-10-31-9-63.us-east-1.aws.redhat.com NetworkManager[618]: [1707076683.4357] device (eth0): state change: secondaries -> activated (reason 'none', sys-iface-state: 'managed') Feb 04 19:58:03 ip-10-31-9-63.us-east-1.aws.redhat.com NetworkManager[618]: [1707076683.4361] manager: NetworkManager state is now CONNECTED_SITE Feb 04 19:58:03 ip-10-31-9-63.us-east-1.aws.redhat.com NetworkManager[618]: [1707076683.4364] device (eth0): Activation: successful, device activated. Feb 04 19:58:03 ip-10-31-9-63.us-east-1.aws.redhat.com NetworkManager[618]: [1707076683.4369] manager: NetworkManager state is now CONNECTED_GLOBAL Feb 04 19:58:03 ip-10-31-9-63.us-east-1.aws.redhat.com NetworkManager[618]: [1707076683.4372] manager: startup complete Feb 04 19:58:03 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: Finished Network Manager Wait Online. ░░ Subject: A start job for unit NetworkManager-wait-online.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit NetworkManager-wait-online.service has finished successfully. ░░ ░░ The job identifier is 184. Feb 04 19:58:03 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: Starting Initial cloud-init job (metadata service crawler)... ░░ Subject: A start job for unit cloud-init.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit cloud-init.service has begun execution. ░░ ░░ The job identifier is 212. Feb 04 19:58:03 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: Cloud-init v. 23.4-2.el9 running 'init' at Sun, 04 Feb 2024 19:58:03 +0000. Up 26.61 seconds. Feb 04 19:58:03 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: ci-info: +++++++++++++++++++++++++++++++++++++++Net device info+++++++++++++++++++++++++++++++++++++++ Feb 04 19:58:03 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: ci-info: +--------+------+------------------------------+---------------+--------+-------------------+ Feb 04 19:58:03 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: ci-info: | Device | Up | Address | Mask | Scope | Hw-Address | Feb 04 19:58:03 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: ci-info: +--------+------+------------------------------+---------------+--------+-------------------+ Feb 04 19:58:03 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: ci-info: | eth0 | True | 10.31.9.63 | 255.255.252.0 | global | 12:cd:2e:aa:47:c5 | Feb 04 19:58:03 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: ci-info: | eth0 | True | fe80::10cd:2eff:feaa:47c5/64 | . | link | 12:cd:2e:aa:47:c5 | Feb 04 19:58:03 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: ci-info: | lo | True | 127.0.0.1 | 255.0.0.0 | host | . | Feb 04 19:58:03 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: ci-info: | lo | True | ::1/128 | . | host | . | Feb 04 19:58:03 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: ci-info: +--------+------+------------------------------+---------------+--------+-------------------+ Feb 04 19:58:03 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: ci-info: ++++++++++++++++++++++++++++Route IPv4 info++++++++++++++++++++++++++++ Feb 04 19:58:03 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: ci-info: +-------+-------------+-----------+---------------+-----------+-------+ Feb 04 19:58:03 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: ci-info: | Route | Destination | Gateway | Genmask | Interface | Flags | Feb 04 19:58:03 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: ci-info: +-------+-------------+-----------+---------------+-----------+-------+ Feb 04 19:58:03 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: ci-info: | 0 | 0.0.0.0 | 10.31.8.1 | 0.0.0.0 | eth0 | UG | Feb 04 19:58:03 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: ci-info: | 1 | 10.31.8.0 | 0.0.0.0 | 255.255.252.0 | eth0 | U | Feb 04 19:58:03 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: ci-info: +-------+-------------+-----------+---------------+-----------+-------+ Feb 04 19:58:03 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: ci-info: +++++++++++++++++++Route IPv6 info+++++++++++++++++++ Feb 04 19:58:03 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: ci-info: +-------+-------------+---------+-----------+-------+ Feb 04 19:58:03 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: ci-info: | Route | Destination | Gateway | Interface | Flags | Feb 04 19:58:03 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: ci-info: +-------+-------------+---------+-----------+-------+ Feb 04 19:58:03 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: ci-info: | 1 | fe80::/64 | :: | eth0 | U | Feb 04 19:58:03 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: ci-info: | 3 | multicast | :: | eth0 | U | Feb 04 19:58:03 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: ci-info: +-------+-------------+---------+-----------+-------+ Feb 04 19:58:03 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: 2024-02-04 19:58:03,859 - __init__.py[WARNING]: Unhandled non-multipart (text/x-not-multipart) userdata: 'b'~/.citool.d/post-install'...' Feb 04 19:58:05 ip-10-31-9-63.us-east-1.aws.redhat.com /usr/sbin/irqbalance[582]: Cannot change IRQ 0 affinity: Input/output error Feb 04 19:58:05 ip-10-31-9-63.us-east-1.aws.redhat.com /usr/sbin/irqbalance[582]: IRQ 0 affinity is now unmanaged Feb 04 19:58:05 ip-10-31-9-63.us-east-1.aws.redhat.com /usr/sbin/irqbalance[582]: Cannot change IRQ 48 affinity: Input/output error Feb 04 19:58:05 ip-10-31-9-63.us-east-1.aws.redhat.com /usr/sbin/irqbalance[582]: IRQ 48 affinity is now unmanaged Feb 04 19:58:05 ip-10-31-9-63.us-east-1.aws.redhat.com /usr/sbin/irqbalance[582]: Cannot change IRQ 49 affinity: Input/output error Feb 04 19:58:05 ip-10-31-9-63.us-east-1.aws.redhat.com /usr/sbin/irqbalance[582]: IRQ 49 affinity is now unmanaged Feb 04 19:58:05 ip-10-31-9-63.us-east-1.aws.redhat.com /usr/sbin/irqbalance[582]: Cannot change IRQ 50 affinity: Input/output error Feb 04 19:58:05 ip-10-31-9-63.us-east-1.aws.redhat.com /usr/sbin/irqbalance[582]: IRQ 50 affinity is now unmanaged Feb 04 19:58:05 ip-10-31-9-63.us-east-1.aws.redhat.com /usr/sbin/irqbalance[582]: Cannot change IRQ 51 affinity: Input/output error Feb 04 19:58:05 ip-10-31-9-63.us-east-1.aws.redhat.com /usr/sbin/irqbalance[582]: IRQ 51 affinity is now unmanaged Feb 04 19:58:05 ip-10-31-9-63.us-east-1.aws.redhat.com /usr/sbin/irqbalance[582]: Cannot change IRQ 52 affinity: Input/output error Feb 04 19:58:05 ip-10-31-9-63.us-east-1.aws.redhat.com /usr/sbin/irqbalance[582]: IRQ 52 affinity is now unmanaged Feb 04 19:58:05 ip-10-31-9-63.us-east-1.aws.redhat.com /usr/sbin/irqbalance[582]: Cannot change IRQ 53 affinity: Input/output error Feb 04 19:58:05 ip-10-31-9-63.us-east-1.aws.redhat.com /usr/sbin/irqbalance[582]: IRQ 53 affinity is now unmanaged Feb 04 19:58:05 ip-10-31-9-63.us-east-1.aws.redhat.com /usr/sbin/irqbalance[582]: Cannot change IRQ 54 affinity: Input/output error Feb 04 19:58:05 ip-10-31-9-63.us-east-1.aws.redhat.com /usr/sbin/irqbalance[582]: IRQ 54 affinity is now unmanaged Feb 04 19:58:05 ip-10-31-9-63.us-east-1.aws.redhat.com /usr/sbin/irqbalance[582]: Cannot change IRQ 55 affinity: Input/output error Feb 04 19:58:05 ip-10-31-9-63.us-east-1.aws.redhat.com /usr/sbin/irqbalance[582]: IRQ 55 affinity is now unmanaged Feb 04 19:58:05 ip-10-31-9-63.us-east-1.aws.redhat.com /usr/sbin/irqbalance[582]: Cannot change IRQ 56 affinity: Input/output error Feb 04 19:58:05 ip-10-31-9-63.us-east-1.aws.redhat.com /usr/sbin/irqbalance[582]: IRQ 56 affinity is now unmanaged Feb 04 19:58:05 ip-10-31-9-63.us-east-1.aws.redhat.com /usr/sbin/irqbalance[582]: Cannot change IRQ 57 affinity: Input/output error Feb 04 19:58:05 ip-10-31-9-63.us-east-1.aws.redhat.com /usr/sbin/irqbalance[582]: IRQ 57 affinity is now unmanaged Feb 04 19:58:05 ip-10-31-9-63.us-east-1.aws.redhat.com /usr/sbin/irqbalance[582]: Cannot change IRQ 58 affinity: Input/output error Feb 04 19:58:05 ip-10-31-9-63.us-east-1.aws.redhat.com /usr/sbin/irqbalance[582]: IRQ 58 affinity is now unmanaged Feb 04 19:58:05 ip-10-31-9-63.us-east-1.aws.redhat.com /usr/sbin/irqbalance[582]: Cannot change IRQ 59 affinity: Input/output error Feb 04 19:58:05 ip-10-31-9-63.us-east-1.aws.redhat.com /usr/sbin/irqbalance[582]: IRQ 59 affinity is now unmanaged Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: Generating public/private rsa key pair. Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: Your identification has been saved in /etc/ssh/ssh_host_rsa_key Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: Your public key has been saved in /etc/ssh/ssh_host_rsa_key.pub Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: The key fingerprint is: Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: SHA256:THTfmPftPCGISRLNGSnhV1FfhWARNSpihY6xYioxi44 root@ip-10-31-9-63.us-east-1.aws.redhat.com Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: The key's randomart image is: Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: +---[RSA 3072]----+ Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: | o=o*oB*o.+| Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: | oooB.o =o..| Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: | B=o. = o. | Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: |o o o+=.o... ..| Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: |.+ o . So . . .o| Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: |+ . .o.| Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: |o. .o| Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: |E. .| Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: | | Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: +----[SHA256]-----+ Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: Generating public/private dsa key pair. Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: Your identification has been saved in /etc/ssh/ssh_host_dsa_key Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: Your public key has been saved in /etc/ssh/ssh_host_dsa_key.pub Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: The key fingerprint is: Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: SHA256:Zettioovlf6hhuFMjaQtQnEe2IsNUbd1+LfvbY6q5CU root@ip-10-31-9-63.us-east-1.aws.redhat.com Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: The key's randomart image is: Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: +---[DSA 1024]----+ Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: | .=. . ... | Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: | + +. o.. | Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: | B o. . o | Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: | o +. + o | Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: |. + o .S o . | Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: |. o = + . o | Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: | . = = .E..+ | Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: | =.o.oooo .o. | Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: | .++oooo.oooo | Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: +----[SHA256]-----+ Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: Generating public/private ecdsa key pair. Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: Your identification has been saved in /etc/ssh/ssh_host_ecdsa_key Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: Your public key has been saved in /etc/ssh/ssh_host_ecdsa_key.pub Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: The key fingerprint is: Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: SHA256:7oxpGPZGRN9iveM54pKmrR7a2wOS+WIG7NPSc1vt1bM root@ip-10-31-9-63.us-east-1.aws.redhat.com Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: The key's randomart image is: Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: +---[ECDSA 256]---+ Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: | | Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: | . | Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: | . . o | Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: | . + o | Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: |. o . .S. . | Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: |..+ + ... o . | Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: |..o+.* o.o + o | Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: | ++*o=X=o = o | Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: | o+oOO*++o . E | Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: +----[SHA256]-----+ Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: Generating public/private ed25519 key pair. Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: Your identification has been saved in /etc/ssh/ssh_host_ed25519_key Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: Your public key has been saved in /etc/ssh/ssh_host_ed25519_key.pub Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: The key fingerprint is: Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: SHA256:YM5BNAg+07UNGL+ZdaEuufbdJOIyuYKRyyUsdnME9zs root@ip-10-31-9-63.us-east-1.aws.redhat.com Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: The key's randomart image is: Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: +--[ED25519 256]--+ Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: | ...=* . | Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: | . oo= = . . | Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: | + + B + . | Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: | o = @ . | Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: | . .. O S | Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: |..=o.. E | Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: |.o.*o o.o . . | Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: | + ..+o o + | Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: | ..+o . . | Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[708]: +----[SHA256]-----+ Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: Finished Initial cloud-init job (metadata service crawler). ░░ Subject: A start job for unit cloud-init.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit cloud-init.service has finished successfully. ░░ ░░ The job identifier is 212. Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: Reached target Cloud-config availability. ░░ Subject: A start job for unit cloud-config.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit cloud-config.target has finished successfully. ░░ ░░ The job identifier is 211. Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: Reached target Network is Online. ░░ Subject: A start job for unit network-online.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit network-online.target has finished successfully. ░░ ░░ The job identifier is 183. Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: Starting Apply the settings specified in cloud-config... ░░ Subject: A start job for unit cloud-config.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit cloud-config.service has begun execution. ░░ ░░ The job identifier is 210. Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: Starting Crash recovery kernel arming... ░░ Subject: A start job for unit kdump.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit kdump.service has begun execution. ░░ ░░ The job identifier is 199. Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: Starting The restraint harness.... ░░ Subject: A start job for unit restraintd.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit restraintd.service has begun execution. ░░ ░░ The job identifier is 198. Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: Starting Notify NFS peers of a restart... ░░ Subject: A start job for unit rpc-statd-notify.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit rpc-statd-notify.service has begun execution. ░░ ░░ The job identifier is 229. Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: Starting System Logging Service... ░░ Subject: A start job for unit rsyslog.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit rsyslog.service has begun execution. ░░ ░░ The job identifier is 236. Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: Starting OpenSSH server daemon... ░░ Subject: A start job for unit sshd.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sshd.service has begun execution. ░░ ░░ The job identifier is 213. Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com sm-notify[785]: Version 2.5.4 starting Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: Started Notify NFS peers of a restart. ░░ Subject: A start job for unit rpc-statd-notify.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit rpc-statd-notify.service has finished successfully. ░░ ░░ The job identifier is 229. Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com sshd[787]: Server listening on 0.0.0.0 port 22. Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com sshd[787]: Server listening on :: port 22. Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: Started OpenSSH server daemon. ░░ Subject: A start job for unit sshd.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sshd.service has finished successfully. ░░ ░░ The job identifier is 213. Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: Started The restraint harness.. ░░ Subject: A start job for unit restraintd.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit restraintd.service has finished successfully. ░░ ░░ The job identifier is 198. Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: Started System Logging Service. ░░ Subject: A start job for unit rsyslog.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit rsyslog.service has finished successfully. ░░ ░░ The job identifier is 236. Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com rsyslogd[786]: [origin software="rsyslogd" swVersion="8.2310.0-3.el9" x-pid="786" x-info="https://www.rsyslog.com"] start Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[824]: Cloud-init v. 23.4-2.el9 running 'modules:config' at Sun, 04 Feb 2024 19:58:07 +0000. Up 30.37 seconds. Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: Stopping OpenSSH server daemon... ░░ Subject: A stop job for unit sshd.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A stop job for unit sshd.service has begun execution. ░░ ░░ The job identifier is 473. Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com sshd[787]: Received signal 15; terminating. Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: sshd.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit sshd.service has successfully entered the 'dead' state. Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: Stopped OpenSSH server daemon. ░░ Subject: A stop job for unit sshd.service has finished ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A stop job for unit sshd.service has finished. ░░ ░░ The job identifier is 473 and the job result is done. Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: Stopped target sshd-keygen.target. ░░ Subject: A stop job for unit sshd-keygen.target has finished ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A stop job for unit sshd-keygen.target has finished. ░░ ░░ The job identifier is 540 and the job result is done. Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: Stopping sshd-keygen.target... ░░ Subject: A stop job for unit sshd-keygen.target has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A stop job for unit sshd-keygen.target has begun execution. ░░ ░░ The job identifier is 540. Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: OpenSSH ecdsa Server Key Generation was skipped because of an unmet condition check (ConditionPathExists=!/run/systemd/generator.early/multi-user.target.wants/cloud-init.target). ░░ Subject: A start job for unit sshd-keygen@ecdsa.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sshd-keygen@ecdsa.service has finished successfully. ░░ ░░ The job identifier is 538. Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: OpenSSH ed25519 Server Key Generation was skipped because of an unmet condition check (ConditionPathExists=!/run/systemd/generator.early/multi-user.target.wants/cloud-init.target). ░░ Subject: A start job for unit sshd-keygen@ed25519.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sshd-keygen@ed25519.service has finished successfully. ░░ ░░ The job identifier is 536. Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: OpenSSH rsa Server Key Generation was skipped because of an unmet condition check (ConditionPathExists=!/run/systemd/generator.early/multi-user.target.wants/cloud-init.target). ░░ Subject: A start job for unit sshd-keygen@rsa.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sshd-keygen@rsa.service has finished successfully. ░░ ░░ The job identifier is 539. Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: Reached target sshd-keygen.target. ░░ Subject: A start job for unit sshd-keygen.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sshd-keygen.target has finished successfully. ░░ ░░ The job identifier is 540. Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: Starting OpenSSH server daemon... ░░ Subject: A start job for unit sshd.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sshd.service has begun execution. ░░ ░░ The job identifier is 473. Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com rsyslogd[786]: imjournal: journal files changed, reloading... [v8.2310.0-3.el9 try https://www.rsyslog.com/e/0 ] Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com sshd[844]: Server listening on 0.0.0.0 port 22. Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com sshd[844]: Server listening on :: port 22. Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: Started OpenSSH server daemon. ░░ Subject: A start job for unit sshd.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sshd.service has finished successfully. ░░ ░░ The job identifier is 473. Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: Finished Apply the settings specified in cloud-config. ░░ Subject: A start job for unit cloud-config.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit cloud-config.service has finished successfully. ░░ ░░ The job identifier is 210. Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: Starting Execute cloud user/final scripts... ░░ Subject: A start job for unit cloud-final.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit cloud-final.service has begun execution. ░░ ░░ The job identifier is 209. Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: Starting Permit User Sessions... ░░ Subject: A start job for unit systemd-user-sessions.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-user-sessions.service has begun execution. ░░ ░░ The job identifier is 243. Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: Finished Permit User Sessions. ░░ Subject: A start job for unit systemd-user-sessions.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-user-sessions.service has finished successfully. ░░ ░░ The job identifier is 243. Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: Started Command Scheduler. ░░ Subject: A start job for unit crond.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit crond.service has finished successfully. ░░ ░░ The job identifier is 244. Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: Started Getty on tty1. ░░ Subject: A start job for unit getty@tty1.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit getty@tty1.service has finished successfully. ░░ ░░ The job identifier is 201. Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: Reached target Login Prompts. ░░ Subject: A start job for unit getty.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit getty.target has finished successfully. ░░ ░░ The job identifier is 200. Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: Reached target Multi-User System. ░░ Subject: A start job for unit multi-user.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit multi-user.target has finished successfully. ░░ ░░ The job identifier is 116. Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: Starting Record Runlevel Change in UTMP... ░░ Subject: A start job for unit systemd-update-utmp-runlevel.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-update-utmp-runlevel.service has begun execution. ░░ ░░ The job identifier is 222. Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: systemd-update-utmp-runlevel.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit systemd-update-utmp-runlevel.service has successfully entered the 'dead' state. Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: Finished Record Runlevel Change in UTMP. ░░ Subject: A start job for unit systemd-update-utmp-runlevel.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-update-utmp-runlevel.service has finished successfully. ░░ ░░ The job identifier is 222. Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com crond[847]: (CRON) STARTUP (1.5.7) Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com crond[847]: (CRON) INFO (Syslog will be used instead of sendmail.) Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com crond[847]: (CRON) INFO (RANDOM_DELAY will be scaled with factor 11% if used.) Feb 04 19:58:07 ip-10-31-9-63.us-east-1.aws.redhat.com crond[847]: (CRON) INFO (running with inotify support) Feb 04 19:58:08 ip-10-31-9-63.us-east-1.aws.redhat.com restraintd[814]: Listening on http://localhost:8081 Feb 04 19:58:08 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[874]: Cloud-init v. 23.4-2.el9 running 'modules:final' at Sun, 04 Feb 2024 19:58:07 +0000. Up 30.85 seconds. Feb 04 19:58:08 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[883]: ############################################################# Feb 04 19:58:08 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[884]: -----BEGIN SSH HOST KEY FINGERPRINTS----- Feb 04 19:58:08 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[886]: 1024 SHA256:Zettioovlf6hhuFMjaQtQnEe2IsNUbd1+LfvbY6q5CU root@ip-10-31-9-63.us-east-1.aws.redhat.com (DSA) Feb 04 19:58:08 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[888]: 256 SHA256:7oxpGPZGRN9iveM54pKmrR7a2wOS+WIG7NPSc1vt1bM root@ip-10-31-9-63.us-east-1.aws.redhat.com (ECDSA) Feb 04 19:58:08 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[890]: 256 SHA256:YM5BNAg+07UNGL+ZdaEuufbdJOIyuYKRyyUsdnME9zs root@ip-10-31-9-63.us-east-1.aws.redhat.com (ED25519) Feb 04 19:58:08 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[898]: 3072 SHA256:THTfmPftPCGISRLNGSnhV1FfhWARNSpihY6xYioxi44 root@ip-10-31-9-63.us-east-1.aws.redhat.com (RSA) Feb 04 19:58:08 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[903]: -----END SSH HOST KEY FINGERPRINTS----- Feb 04 19:58:08 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[904]: ############################################################# Feb 04 19:58:08 ip-10-31-9-63.us-east-1.aws.redhat.com cloud-init[874]: Cloud-init v. 23.4-2.el9 finished at Sun, 04 Feb 2024 19:58:08 +0000. Datasource DataSourceEc2Local. Up 31.03 seconds Feb 04 19:58:08 ip-10-31-9-63.us-east-1.aws.redhat.com kdumpctl[790]: kdump: Detected change(s) in the following file(s): /etc/fstab Feb 04 19:58:08 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: Finished Execute cloud user/final scripts. ░░ Subject: A start job for unit cloud-final.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit cloud-final.service has finished successfully. ░░ ░░ The job identifier is 209. Feb 04 19:58:08 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: Reached target Cloud-init target. ░░ Subject: A start job for unit cloud-init.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit cloud-init.target has finished successfully. ░░ ░░ The job identifier is 208. Feb 04 19:58:10 ip-10-31-9-63.us-east-1.aws.redhat.com chronyd[592]: Selected source 162.159.200.123 (2.centos.pool.ntp.org) Feb 04 19:58:10 ip-10-31-9-63.us-east-1.aws.redhat.com chronyd[592]: System clock TAI offset set to 37 seconds Feb 04 19:58:13 ip-10-31-9-63.us-east-1.aws.redhat.com kernel: block xvda: the capability attribute has been deprecated. Feb 04 19:58:13 ip-10-31-9-63.us-east-1.aws.redhat.com kdumpctl[790]: kdump: Rebuilding /boot/initramfs-5.14.0-410.el9.x86_64kdump.img Feb 04 19:58:13 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: NetworkManager-dispatcher.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit NetworkManager-dispatcher.service has successfully entered the 'dead' state. Feb 04 19:58:14 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1407]: dracut-057-53.git20240104.el9 Feb 04 19:58:15 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: Executing: /usr/bin/dracut --add kdumpbase --quiet --hostonly --hostonly-cmdline --hostonly-i18n --hostonly-mode strict --hostonly-nics -o "plymouth resume ifcfg earlykdump" --mount "/dev/disk/by-uuid/f72128f2-bcfd-46ab-b509-3a36632be876 /sysroot xfs rw,relatime,seclabel,attr2,inode64,logbufs=8,logbsize=32k,noquota" --squash-compressor zstd --no-hostonly-default-device -f /boot/initramfs-5.14.0-410.el9.x86_64kdump.img 5.14.0-410.el9.x86_64 Feb 04 19:58:16 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: dracut module 'systemd-networkd' will not be installed, because command 'networkctl' could not be found! Feb 04 19:58:16 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: dracut module 'systemd-networkd' will not be installed, because command '/usr/lib/systemd/systemd-networkd' could not be found! Feb 04 19:58:16 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: dracut module 'systemd-networkd' will not be installed, because command '/usr/lib/systemd/systemd-networkd-wait-online' could not be found! Feb 04 19:58:16 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: dracut module 'systemd-resolved' will not be installed, because command 'resolvectl' could not be found! Feb 04 19:58:16 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: dracut module 'systemd-resolved' will not be installed, because command '/usr/lib/systemd/systemd-resolved' could not be found! Feb 04 19:58:16 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: dracut module 'systemd-timesyncd' will not be installed, because command '/usr/lib/systemd/systemd-timesyncd' could not be found! Feb 04 19:58:16 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: dracut module 'systemd-timesyncd' will not be installed, because command '/usr/lib/systemd/systemd-time-wait-sync' could not be found! Feb 04 19:58:16 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: dracut module 'busybox' will not be installed, because command 'busybox' could not be found! Feb 04 19:58:16 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: dracut module 'dbus-daemon' will not be installed, because command 'dbus-daemon' could not be found! Feb 04 19:58:16 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: dracut module 'connman' will not be installed, because command 'connmand' could not be found! Feb 04 19:58:16 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: dracut module 'connman' will not be installed, because command 'connmanctl' could not be found! Feb 04 19:58:16 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: dracut module 'connman' will not be installed, because command 'connmand-wait-online' could not be found! Feb 04 19:58:16 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: dracut module 'network-wicked' will not be installed, because command 'wicked' could not be found! Feb 04 19:58:16 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: dracut module 'ifcfg' will not be installed, because it's in the list to be omitted! Feb 04 19:58:16 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: dracut module 'plymouth' will not be installed, because it's in the list to be omitted! Feb 04 19:58:16 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: 62bluetooth: Could not find any command of '/usr/lib/bluetooth/bluetoothd /usr/libexec/bluetooth/bluetoothd'! Feb 04 19:58:16 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: dracut module 'lvmmerge' will not be installed, because command 'lvm' could not be found! Feb 04 19:58:16 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: dracut module 'lvmthinpool-monitor' will not be installed, because command 'lvm' could not be found! Feb 04 19:58:16 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: dracut module 'btrfs' will not be installed, because command 'btrfs' could not be found! Feb 04 19:58:16 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: dracut module 'dmraid' will not be installed, because command 'dmraid' could not be found! Feb 04 19:58:16 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: dracut module 'lvm' will not be installed, because command 'lvm' could not be found! Feb 04 19:58:16 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: dracut module 'mdraid' will not be installed, because command 'mdadm' could not be found! Feb 04 19:58:17 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: dracut module 'pcsc' will not be installed, because command 'pcscd' could not be found! Feb 04 19:58:17 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: dracut module 'tpm2-tss' will not be installed, because command 'tpm2' could not be found! Feb 04 19:58:17 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: dracut module 'cifs' will not be installed, because command 'mount.cifs' could not be found! Feb 04 19:58:17 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: dracut module 'iscsi' will not be installed, because command 'iscsi-iname' could not be found! Feb 04 19:58:17 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: dracut module 'iscsi' will not be installed, because command 'iscsiadm' could not be found! Feb 04 19:58:17 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: dracut module 'iscsi' will not be installed, because command 'iscsid' could not be found! Feb 04 19:58:17 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: dracut module 'nvmf' will not be installed, because command 'nvme' could not be found! Feb 04 19:58:17 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: dracut module 'resume' will not be installed, because it's in the list to be omitted! Feb 04 19:58:17 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: dracut module 'biosdevname' will not be installed, because command 'biosdevname' could not be found! Feb 04 19:58:17 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: dracut module 'earlykdump' will not be installed, because it's in the list to be omitted! Feb 04 19:58:17 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: dracut module 'memstrack' will not be installed, because command 'memstrack' could not be found! Feb 04 19:58:17 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: memstrack is not available Feb 04 19:58:17 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: If you need to use rd.memdebug>=4, please install memstrack and procps-ng Feb 04 19:58:17 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: dracut module 'systemd-resolved' will not be installed, because command 'resolvectl' could not be found! Feb 04 19:58:17 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: dracut module 'systemd-resolved' will not be installed, because command '/usr/lib/systemd/systemd-resolved' could not be found! Feb 04 19:58:17 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: dracut module 'systemd-timesyncd' will not be installed, because command '/usr/lib/systemd/systemd-timesyncd' could not be found! Feb 04 19:58:17 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: dracut module 'systemd-timesyncd' will not be installed, because command '/usr/lib/systemd/systemd-time-wait-sync' could not be found! Feb 04 19:58:17 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: dracut module 'busybox' will not be installed, because command 'busybox' could not be found! Feb 04 19:58:17 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: dracut module 'dbus-daemon' will not be installed, because command 'dbus-daemon' could not be found! Feb 04 19:58:17 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: dracut module 'connman' will not be installed, because command 'connmand' could not be found! Feb 04 19:58:17 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: dracut module 'connman' will not be installed, because command 'connmanctl' could not be found! Feb 04 19:58:17 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: dracut module 'connman' will not be installed, because command 'connmand-wait-online' could not be found! Feb 04 19:58:17 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: dracut module 'network-wicked' will not be installed, because command 'wicked' could not be found! Feb 04 19:58:17 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: 62bluetooth: Could not find any command of '/usr/lib/bluetooth/bluetoothd /usr/libexec/bluetooth/bluetoothd'! Feb 04 19:58:17 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: dracut module 'lvmmerge' will not be installed, because command 'lvm' could not be found! Feb 04 19:58:17 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: dracut module 'lvmthinpool-monitor' will not be installed, because command 'lvm' could not be found! Feb 04 19:58:17 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: dracut module 'btrfs' will not be installed, because command 'btrfs' could not be found! Feb 04 19:58:17 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: dracut module 'dmraid' will not be installed, because command 'dmraid' could not be found! Feb 04 19:58:17 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: dracut module 'lvm' will not be installed, because command 'lvm' could not be found! Feb 04 19:58:17 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: dracut module 'mdraid' will not be installed, because command 'mdadm' could not be found! Feb 04 19:58:17 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: dracut module 'pcsc' will not be installed, because command 'pcscd' could not be found! Feb 04 19:58:17 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: dracut module 'tpm2-tss' will not be installed, because command 'tpm2' could not be found! Feb 04 19:58:17 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: dracut module 'cifs' will not be installed, because command 'mount.cifs' could not be found! Feb 04 19:58:17 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: dracut module 'iscsi' will not be installed, because command 'iscsi-iname' could not be found! Feb 04 19:58:17 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: dracut module 'iscsi' will not be installed, because command 'iscsiadm' could not be found! Feb 04 19:58:17 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: dracut module 'iscsi' will not be installed, because command 'iscsid' could not be found! Feb 04 19:58:17 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: dracut module 'nvmf' will not be installed, because command 'nvme' could not be found! Feb 04 19:58:17 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: dracut module 'memstrack' will not be installed, because command 'memstrack' could not be found! Feb 04 19:58:17 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: memstrack is not available Feb 04 19:58:17 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: If you need to use rd.memdebug>=4, please install memstrack and procps-ng Feb 04 19:58:17 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: *** Including module: systemd *** Feb 04 19:58:17 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: *** Including module: systemd-initrd *** Feb 04 19:58:17 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: *** Including module: nss-softokn *** Feb 04 19:58:17 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: *** Including module: rngd *** Feb 04 19:58:17 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: *** Including module: i18n *** Feb 04 19:58:18 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: *** Including module: drm *** Feb 04 19:58:18 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: *** Including module: prefixdevname *** Feb 04 19:58:18 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: *** Including module: kernel-modules *** Feb 04 19:58:19 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: *** Including module: kernel-modules-extra *** Feb 04 19:58:19 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: kernel-modules-extra: configuration source "/run/depmod.d" does not exist Feb 04 19:58:19 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: kernel-modules-extra: configuration source "/lib/depmod.d" does not exist Feb 04 19:58:19 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: kernel-modules-extra: parsing configuration file "/etc/depmod.d/dist.conf" Feb 04 19:58:19 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: kernel-modules-extra: /etc/depmod.d/dist.conf: added "updates extra built-in weak-updates" to the list of search directories Feb 04 19:58:19 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: *** Including module: fstab-sys *** Feb 04 19:58:19 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: *** Including module: rootfs-block *** Feb 04 19:58:19 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: *** Including module: terminfo *** Feb 04 19:58:19 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: *** Including module: udev-rules *** Feb 04 19:58:19 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: Skipping udev rule: 91-permissions.rules Feb 04 19:58:19 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: Skipping udev rule: 80-drivers-modprobe.rules Feb 04 19:58:19 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: *** Including module: dracut-systemd *** Feb 04 19:58:19 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: *** Including module: usrmount *** Feb 04 19:58:19 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: *** Including module: base *** Feb 04 19:58:20 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: *** Including module: fs-lib *** Feb 04 19:58:20 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: *** Including module: kdumpbase *** Feb 04 19:58:20 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: *** Including module: microcode_ctl-fw_dir_override *** Feb 04 19:58:20 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: microcode_ctl module: mangling fw_dir Feb 04 19:58:20 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: microcode_ctl: reset fw_dir to "/lib/firmware/updates /lib/firmware" Feb 04 19:58:20 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: microcode_ctl: processing data directory "/usr/share/microcode_ctl/ucode_with_caveats/intel"... Feb 04 19:58:20 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: microcode_ctl: intel: caveats check for kernel version "5.14.0-410.el9.x86_64" passed, adding "/usr/share/microcode_ctl/ucode_with_caveats/intel" to fw_dir variable Feb 04 19:58:20 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: microcode_ctl: processing data directory "/usr/share/microcode_ctl/ucode_with_caveats/intel-06-2d-07"... Feb 04 19:58:20 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: microcode_ctl: configuration "intel-06-2d-07" is ignored Feb 04 19:58:20 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: microcode_ctl: processing data directory "/usr/share/microcode_ctl/ucode_with_caveats/intel-06-4e-03"... Feb 04 19:58:20 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: microcode_ctl: configuration "intel-06-4e-03" is ignored Feb 04 19:58:20 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: microcode_ctl: processing data directory "/usr/share/microcode_ctl/ucode_with_caveats/intel-06-4f-01"... Feb 04 19:58:20 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: microcode_ctl: configuration "intel-06-4f-01" is ignored Feb 04 19:58:20 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: microcode_ctl: processing data directory "/usr/share/microcode_ctl/ucode_with_caveats/intel-06-55-04"... Feb 04 19:58:20 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: microcode_ctl: configuration "intel-06-55-04" is ignored Feb 04 19:58:20 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: microcode_ctl: processing data directory "/usr/share/microcode_ctl/ucode_with_caveats/intel-06-5e-03"... Feb 04 19:58:20 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: microcode_ctl: configuration "intel-06-5e-03" is ignored Feb 04 19:58:20 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: microcode_ctl: processing data directory "/usr/share/microcode_ctl/ucode_with_caveats/intel-06-8c-01"... Feb 04 19:58:20 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: microcode_ctl: configuration "intel-06-8c-01" is ignored Feb 04 19:58:20 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: microcode_ctl: processing data directory "/usr/share/microcode_ctl/ucode_with_caveats/intel-06-8e-9e-0x-0xca"... Feb 04 19:58:20 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: microcode_ctl: configuration "intel-06-8e-9e-0x-0xca" is ignored Feb 04 19:58:20 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: microcode_ctl: processing data directory "/usr/share/microcode_ctl/ucode_with_caveats/intel-06-8e-9e-0x-dell"... Feb 04 19:58:20 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: microcode_ctl: configuration "intel-06-8e-9e-0x-dell" is ignored Feb 04 19:58:20 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: microcode_ctl: final fw_dir: "/usr/share/microcode_ctl/ucode_with_caveats/intel /lib/firmware/updates /lib/firmware" Feb 04 19:58:20 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: *** Including module: shutdown *** Feb 04 19:58:20 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: *** Including module: squash *** Feb 04 19:58:20 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: *** Including modules done *** Feb 04 19:58:21 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: *** Installing kernel module dependencies *** Feb 04 19:58:21 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: *** Installing kernel module dependencies done *** Feb 04 19:58:21 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: *** Resolving executable dependencies *** Feb 04 19:58:23 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: *** Resolving executable dependencies done *** Feb 04 19:58:23 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: *** Hardlinking files *** Feb 04 19:58:23 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: Mode: real Feb 04 19:58:23 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: Files: 436 Feb 04 19:58:23 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: Linked: 1 files Feb 04 19:58:23 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: Compared: 0 xattrs Feb 04 19:58:23 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: Compared: 6 files Feb 04 19:58:23 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: Saved: 55.98 KiB Feb 04 19:58:23 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: Duration: 0.005397 seconds Feb 04 19:58:23 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: *** Hardlinking files done *** Feb 04 19:58:23 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: *** Generating early-microcode cpio image *** Feb 04 19:58:23 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: *** Constructing GenuineIntel.bin *** Feb 04 19:58:23 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: *** Constructing GenuineIntel.bin *** Feb 04 19:58:23 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: *** Store current command line parameters *** Feb 04 19:58:23 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: Stored kernel commandline: Feb 04 19:58:23 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: No dracut internal kernel commandline stored in the initramfs Feb 04 19:58:23 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: *** Install squash loader *** Feb 04 19:58:24 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: *** Stripping files *** Feb 04 19:58:25 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: *** Stripping files done *** Feb 04 19:58:25 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: *** Squashing the files inside the initramfs *** Feb 04 19:58:30 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: *** Squashing the files inside the initramfs done *** Feb 04 19:58:30 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: *** Creating image file '/boot/initramfs-5.14.0-410.el9.x86_64kdump.img' *** Feb 04 19:58:30 ip-10-31-9-63.us-east-1.aws.redhat.com dracut[1409]: *** Creating initramfs image file '/boot/initramfs-5.14.0-410.el9.x86_64kdump.img' done *** Feb 04 19:58:31 ip-10-31-9-63.us-east-1.aws.redhat.com kdumpctl[790]: kdump: kexec: loaded kdump kernel Feb 04 19:58:31 ip-10-31-9-63.us-east-1.aws.redhat.com kdumpctl[790]: kdump: Starting kdump: [OK] Feb 04 19:58:31 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: Finished Crash recovery kernel arming. ░░ Subject: A start job for unit kdump.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit kdump.service has finished successfully. ░░ ░░ The job identifier is 199. Feb 04 19:58:31 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: Startup finished in 909ms (kernel) + 6.490s (initrd) + 47.007s (userspace) = 54.407s. ░░ Subject: System start-up is now complete ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ All system services necessary queued for starting at boot have been ░░ started. Note that this does not mean that the machine is now idle as services ░░ might still be busy with completing start-up. ░░ ░░ Kernel start-up required 909461 microseconds. ░░ ░░ Initrd start-up required 6490320 microseconds. ░░ ░░ Userspace start-up required 47007995 microseconds. Feb 04 19:58:33 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: systemd-hostnamed.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit systemd-hostnamed.service has successfully entered the 'dead' state. Feb 04 19:59:16 ip-10-31-9-63.us-east-1.aws.redhat.com sshd[4033]: Accepted publickey for root from 10.30.33.101 port 41810 ssh2: RSA SHA256:W3cSdmPJK+d9RwU97ardijPXIZnxHswrpTHWW9oYtEU Feb 04 19:59:16 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: Created slice User Slice of UID 0. ░░ Subject: A start job for unit user-0.slice has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit user-0.slice has finished successfully. ░░ ░░ The job identifier is 542. Feb 04 19:59:16 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: Starting User Runtime Directory /run/user/0... ░░ Subject: A start job for unit user-runtime-dir@0.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit user-runtime-dir@0.service has begun execution. ░░ ░░ The job identifier is 605. Feb 04 19:59:16 ip-10-31-9-63.us-east-1.aws.redhat.com systemd-logind[584]: New session 1 of user root. ░░ Subject: A new session 1 has been created for user root ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ Documentation: sd-login(3) ░░ ░░ A new session with the ID 1 has been created for the user root. ░░ ░░ The leading process of the session is 4033. Feb 04 19:59:16 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: Finished User Runtime Directory /run/user/0. ░░ Subject: A start job for unit user-runtime-dir@0.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit user-runtime-dir@0.service has finished successfully. ░░ ░░ The job identifier is 605. Feb 04 19:59:16 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: Starting User Manager for UID 0... ░░ Subject: A start job for unit user@0.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit user@0.service has begun execution. ░░ ░░ The job identifier is 541. Feb 04 19:59:16 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[4037]: pam_unix(systemd-user:session): session opened for user root(uid=0) by (uid=0) Feb 04 19:59:16 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[4037]: Queued start job for default target Main User Target. Feb 04 19:59:16 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[4037]: Created slice User Application Slice. ░░ Subject: A start job for unit UNIT has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has finished successfully. ░░ ░░ The job identifier is 5. Feb 04 19:59:16 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[4037]: Mark boot as successful after the user session has run 2 minutes was skipped because of an unmet condition check (ConditionUser=!@system). ░░ Subject: A start job for unit UNIT has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has finished successfully. ░░ ░░ The job identifier is 11. Feb 04 19:59:16 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[4037]: Started Daily Cleanup of User's Temporary Directories. ░░ Subject: A start job for unit UNIT has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has finished successfully. ░░ ░░ The job identifier is 12. Feb 04 19:59:16 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[4037]: Reached target Paths. ░░ Subject: A start job for unit UNIT has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has finished successfully. ░░ ░░ The job identifier is 9. Feb 04 19:59:16 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[4037]: Reached target Timers. ░░ Subject: A start job for unit UNIT has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has finished successfully. ░░ ░░ The job identifier is 10. Feb 04 19:59:16 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[4037]: Starting D-Bus User Message Bus Socket... ░░ Subject: A start job for unit UNIT has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has begun execution. ░░ ░░ The job identifier is 4. Feb 04 19:59:16 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[4037]: Starting Create User's Volatile Files and Directories... ░░ Subject: A start job for unit UNIT has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has begun execution. ░░ ░░ The job identifier is 8. Feb 04 19:59:16 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[4037]: Finished Create User's Volatile Files and Directories. ░░ Subject: A start job for unit UNIT has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has finished successfully. ░░ ░░ The job identifier is 8. Feb 04 19:59:16 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[4037]: Listening on D-Bus User Message Bus Socket. ░░ Subject: A start job for unit UNIT has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has finished successfully. ░░ ░░ The job identifier is 4. Feb 04 19:59:16 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[4037]: Reached target Sockets. ░░ Subject: A start job for unit UNIT has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has finished successfully. ░░ ░░ The job identifier is 3. Feb 04 19:59:16 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[4037]: Reached target Basic System. ░░ Subject: A start job for unit UNIT has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has finished successfully. ░░ ░░ The job identifier is 2. Feb 04 19:59:16 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[4037]: Reached target Main User Target. ░░ Subject: A start job for unit UNIT has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has finished successfully. ░░ ░░ The job identifier is 1. Feb 04 19:59:16 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[4037]: Startup finished in 75ms. ░░ Subject: User manager start-up is now complete ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The user manager instance for user 0 has been started. All services queued ░░ for starting have been started. Note that other services might still be starting ░░ up or be started at any later time. ░░ ░░ Startup of the manager took 75909 microseconds. Feb 04 19:59:16 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: Started User Manager for UID 0. ░░ Subject: A start job for unit user@0.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit user@0.service has finished successfully. ░░ ░░ The job identifier is 541. Feb 04 19:59:16 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: Started Session 1 of User root. ░░ Subject: A start job for unit session-1.scope has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit session-1.scope has finished successfully. ░░ ░░ The job identifier is 606. Feb 04 19:59:16 ip-10-31-9-63.us-east-1.aws.redhat.com sshd[4033]: pam_unix(sshd:session): session opened for user root(uid=0) by (uid=0) Feb 04 19:59:16 ip-10-31-9-63.us-east-1.aws.redhat.com sshd[4046]: Received disconnect from 10.30.33.101 port 41810:11: disconnected by user Feb 04 19:59:16 ip-10-31-9-63.us-east-1.aws.redhat.com sshd[4046]: Disconnected from user root 10.30.33.101 port 41810 Feb 04 19:59:16 ip-10-31-9-63.us-east-1.aws.redhat.com sshd[4033]: pam_unix(sshd:session): session closed for user root Feb 04 19:59:16 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: session-1.scope: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit session-1.scope has successfully entered the 'dead' state. Feb 04 19:59:16 ip-10-31-9-63.us-east-1.aws.redhat.com systemd-logind[584]: Session 1 logged out. Waiting for processes to exit. Feb 04 19:59:16 ip-10-31-9-63.us-east-1.aws.redhat.com systemd-logind[584]: Removed session 1. ░░ Subject: Session 1 has been terminated ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ Documentation: sd-login(3) ░░ ░░ A session with the ID 1 has been terminated. Feb 04 19:59:18 ip-10-31-9-63.us-east-1.aws.redhat.com sshd[4075]: error: kex_exchange_identification: Connection closed by remote host Feb 04 19:59:18 ip-10-31-9-63.us-east-1.aws.redhat.com sshd[4075]: Connection closed by 10.29.163.167 port 49310 Feb 04 19:59:18 ip-10-31-9-63.us-east-1.aws.redhat.com sshd[4076]: Accepted publickey for root from 10.29.163.167 port 49320 ssh2: RSA SHA256:W3cSdmPJK+d9RwU97ardijPXIZnxHswrpTHWW9oYtEU Feb 04 19:59:18 ip-10-31-9-63.us-east-1.aws.redhat.com systemd-logind[584]: New session 3 of user root. ░░ Subject: A new session 3 has been created for user root ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ Documentation: sd-login(3) ░░ ░░ A new session with the ID 3 has been created for the user root. ░░ ░░ The leading process of the session is 4076. Feb 04 19:59:18 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: Started Session 3 of User root. ░░ Subject: A start job for unit session-3.scope has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit session-3.scope has finished successfully. ░░ ░░ The job identifier is 672. Feb 04 19:59:18 ip-10-31-9-63.us-east-1.aws.redhat.com sshd[4076]: pam_unix(sshd:session): session opened for user root(uid=0) by (uid=0) Feb 04 19:59:18 ip-10-31-9-63.us-east-1.aws.redhat.com sshd[4079]: Received disconnect from 10.29.163.167 port 49320:11: disconnected by user Feb 04 19:59:18 ip-10-31-9-63.us-east-1.aws.redhat.com sshd[4079]: Disconnected from user root 10.29.163.167 port 49320 Feb 04 19:59:18 ip-10-31-9-63.us-east-1.aws.redhat.com sshd[4076]: pam_unix(sshd:session): session closed for user root Feb 04 19:59:18 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: session-3.scope: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit session-3.scope has successfully entered the 'dead' state. Feb 04 19:59:18 ip-10-31-9-63.us-east-1.aws.redhat.com systemd-logind[584]: Session 3 logged out. Waiting for processes to exit. Feb 04 19:59:18 ip-10-31-9-63.us-east-1.aws.redhat.com systemd-logind[584]: Removed session 3. ░░ Subject: Session 3 has been terminated ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ Documentation: sd-login(3) ░░ ░░ A session with the ID 3 has been terminated. Feb 04 19:59:18 ip-10-31-9-63.us-east-1.aws.redhat.com sshd[4096]: Accepted publickey for root from 10.29.163.167 port 49328 ssh2: RSA SHA256:W3cSdmPJK+d9RwU97ardijPXIZnxHswrpTHWW9oYtEU Feb 04 19:59:18 ip-10-31-9-63.us-east-1.aws.redhat.com systemd-logind[584]: New session 4 of user root. ░░ Subject: A new session 4 has been created for user root ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ Documentation: sd-login(3) ░░ ░░ A new session with the ID 4 has been created for the user root. ░░ ░░ The leading process of the session is 4096. Feb 04 19:59:18 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: Started Session 4 of User root. ░░ Subject: A start job for unit session-4.scope has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit session-4.scope has finished successfully. ░░ ░░ The job identifier is 738. Feb 04 19:59:18 ip-10-31-9-63.us-east-1.aws.redhat.com sshd[4096]: pam_unix(sshd:session): session opened for user root(uid=0) by (uid=0) Feb 04 19:59:18 ip-10-31-9-63.us-east-1.aws.redhat.com sshd[4099]: Received disconnect from 10.29.163.167 port 49328:11: disconnected by user Feb 04 19:59:18 ip-10-31-9-63.us-east-1.aws.redhat.com sshd[4099]: Disconnected from user root 10.29.163.167 port 49328 Feb 04 19:59:18 ip-10-31-9-63.us-east-1.aws.redhat.com sshd[4096]: pam_unix(sshd:session): session closed for user root Feb 04 19:59:18 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: session-4.scope: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit session-4.scope has successfully entered the 'dead' state. Feb 04 19:59:18 ip-10-31-9-63.us-east-1.aws.redhat.com systemd-logind[584]: Session 4 logged out. Waiting for processes to exit. Feb 04 19:59:18 ip-10-31-9-63.us-east-1.aws.redhat.com systemd-logind[584]: Removed session 4. ░░ Subject: Session 4 has been terminated ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ Documentation: sd-login(3) ░░ ░░ A session with the ID 4 has been terminated. Feb 04 19:59:18 ip-10-31-9-63.us-east-1.aws.redhat.com sshd[4116]: Accepted publickey for root from 10.29.163.167 port 49336 ssh2: RSA SHA256:W3cSdmPJK+d9RwU97ardijPXIZnxHswrpTHWW9oYtEU Feb 04 19:59:18 ip-10-31-9-63.us-east-1.aws.redhat.com systemd-logind[584]: New session 5 of user root. ░░ Subject: A new session 5 has been created for user root ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ Documentation: sd-login(3) ░░ ░░ A new session with the ID 5 has been created for the user root. ░░ ░░ The leading process of the session is 4116. Feb 04 19:59:18 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: Started Session 5 of User root. ░░ Subject: A start job for unit session-5.scope has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit session-5.scope has finished successfully. ░░ ░░ The job identifier is 804. Feb 04 19:59:18 ip-10-31-9-63.us-east-1.aws.redhat.com sshd[4116]: pam_unix(sshd:session): session opened for user root(uid=0) by (uid=0) Feb 04 19:59:18 ip-10-31-9-63.us-east-1.aws.redhat.com sshd[4119]: Received disconnect from 10.29.163.167 port 49336:11: disconnected by user Feb 04 19:59:18 ip-10-31-9-63.us-east-1.aws.redhat.com sshd[4119]: Disconnected from user root 10.29.163.167 port 49336 Feb 04 19:59:18 ip-10-31-9-63.us-east-1.aws.redhat.com sshd[4116]: pam_unix(sshd:session): session closed for user root Feb 04 19:59:18 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: session-5.scope: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit session-5.scope has successfully entered the 'dead' state. Feb 04 19:59:18 ip-10-31-9-63.us-east-1.aws.redhat.com systemd-logind[584]: Session 5 logged out. Waiting for processes to exit. Feb 04 19:59:18 ip-10-31-9-63.us-east-1.aws.redhat.com systemd-logind[584]: Removed session 5. ░░ Subject: Session 5 has been terminated ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ Documentation: sd-login(3) ░░ ░░ A session with the ID 5 has been terminated. Feb 04 19:59:19 ip-10-31-9-63.us-east-1.aws.redhat.com sshd[4136]: Accepted publickey for root from 10.29.163.167 port 49350 ssh2: RSA SHA256:W3cSdmPJK+d9RwU97ardijPXIZnxHswrpTHWW9oYtEU Feb 04 19:59:19 ip-10-31-9-63.us-east-1.aws.redhat.com systemd-logind[584]: New session 6 of user root. ░░ Subject: A new session 6 has been created for user root ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ Documentation: sd-login(3) ░░ ░░ A new session with the ID 6 has been created for the user root. ░░ ░░ The leading process of the session is 4136. Feb 04 19:59:19 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: Started Session 6 of User root. ░░ Subject: A start job for unit session-6.scope has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit session-6.scope has finished successfully. ░░ ░░ The job identifier is 870. Feb 04 19:59:19 ip-10-31-9-63.us-east-1.aws.redhat.com sshd[4136]: pam_unix(sshd:session): session opened for user root(uid=0) by (uid=0) Feb 04 19:59:49 ip-10-31-9-63.us-east-1.aws.redhat.com python3[4233]: ansible-ansible.legacy.setup Invoked with gather_subset=['all'] gather_timeout=10 filter=[] fact_path=/etc/ansible/facts.d Feb 04 19:59:50 ip-10-31-9-63.us-east-1.aws.redhat.com python3[4274]: ansible-service_facts Invoked Feb 04 19:59:52 ip-10-31-9-63.us-east-1.aws.redhat.com python3[4372]: ansible-ansible.legacy.command Invoked with _raw_params=rpm -qa _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 04 19:59:53 ip-10-31-9-63.us-east-1.aws.redhat.com python3[4390]: ansible-stat Invoked with path=/etc/yum.repos.d/qa-tools.repo follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Feb 04 19:59:54 ip-10-31-9-63.us-east-1.aws.redhat.com python3[4409]: ansible-ansible.legacy.command Invoked with _raw_params=curl -sLI --connect-timeout 5 -w '%{response_code}' http://liver2.lab.eng.brq.redhat.com/repo/ | grep ^200 _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 04 19:59:54 ip-10-31-9-63.us-east-1.aws.redhat.com python3[4430]: ansible-ansible.legacy.dnf Invoked with name=['ca-certificates'] state=latest allow_downgrade=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 allowerasing=False nobest=False conf_file=None disable_excludes=None download_dir=None list=None releasever=None Feb 04 19:59:55 ip-10-31-9-63.us-east-1.aws.redhat.com python3[4448]: ansible-ansible.legacy.dnf Invoked with name=['curl'] state=latest allow_downgrade=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 allowerasing=False nobest=False conf_file=None disable_excludes=None download_dir=None list=None releasever=None Feb 04 19:59:56 ip-10-31-9-63.us-east-1.aws.redhat.com python3[4466]: ansible-ansible.legacy.command Invoked with _raw_params=curl -skL -o /etc/pki/ca-trust/source/anchors/Current-IT-Root-CAs.pem https://certs.corp.redhat.com/certs/Current-IT-Root-CAs.pem _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 04 19:59:56 ip-10-31-9-63.us-east-1.aws.redhat.com python3[4485]: ansible-ansible.legacy.command Invoked with _raw_params=bash -c 'update-ca-trust' _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 04 19:59:58 ip-10-31-9-63.us-east-1.aws.redhat.com python3[4509]: ansible-ansible.legacy.command Invoked with _raw_params=bash -c 'update-ca-trust extract' _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 04 19:59:59 ip-10-31-9-63.us-east-1.aws.redhat.com python3[4533]: ansible-ansible.legacy.command Invoked with _raw_params=curl -v https://gitlab.cee.redhat.com _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 04 19:59:59 ip-10-31-9-63.us-east-1.aws.redhat.com python3[4552]: ansible-ansible.legacy.command Invoked with _raw_params=curl -v https://beaker.engineering.redhat.com/ _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 04 20:00:00 ip-10-31-9-63.us-east-1.aws.redhat.com python3[4571]: ansible-ansible.legacy.command Invoked with _raw_params=if [ $(find /etc/yum.repos.d/ -name 'centos-addons.repo' | wc -l ) -gt 0 ]; then sed -i "/^\[nfv-source\]/,/^$/d" /etc/yum.repos.d/centos-addons.repo; fi _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 04 20:00:00 ip-10-31-9-63.us-east-1.aws.redhat.com python3[4593]: ansible-file Invoked with path=/etc/yum.repos.d/CentOS-Media.repo state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Feb 04 20:00:00 ip-10-31-9-63.us-east-1.aws.redhat.com python3[4610]: ansible-file Invoked with path=/etc/yum.repos.d/CentOS-Linux-Media.repo state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Feb 04 20:00:01 ip-10-31-9-63.us-east-1.aws.redhat.com python3[4627]: ansible-file Invoked with path=/etc/yum.repos.d/CentOS-Stream-Media.repo state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Feb 04 20:00:01 ip-10-31-9-63.us-east-1.aws.redhat.com python3[4644]: ansible-file Invoked with path=/etc/yum.repos.d/beaker-client-testing.repo state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Feb 04 20:00:01 ip-10-31-9-63.us-east-1.aws.redhat.com python3[4661]: ansible-stat Invoked with path=/etc/yum.repos.d/beaker-client.repo follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Feb 04 20:00:01 ip-10-31-9-63.us-east-1.aws.redhat.com python3[4680]: ansible-replace Invoked with path=/etc/yum.repos.d/beaker-client.repo regexp=7|\$releasever replace=8 backup=False encoding=utf-8 unsafe_writes=False after=None before=None validate=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Feb 04 20:00:02 ip-10-31-9-63.us-east-1.aws.redhat.com python3[4697]: ansible-stat Invoked with path=/etc/yum.repos.d/beaker-client.repo follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Feb 04 20:00:02 ip-10-31-9-63.us-east-1.aws.redhat.com python3[4716]: ansible-community.general.ini_file Invoked with path=/etc/yum.repos.d/beaker-client.repo section=beaker-client option=skip_if_unavailable value=1 backup=False state=present exclusive=True no_extra_spaces=False ignore_spaces=False allow_no_value=False modify_inactive_option=True create=True follow=False unsafe_writes=False values=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Feb 04 20:00:02 ip-10-31-9-63.us-east-1.aws.redhat.com python3[4733]: ansible-ansible.legacy.command Invoked with _raw_params=for repofile in $(ls -1 /etc/yum.repos.d/); do echo "# vvvvv $repofile vvvvv vvvvv"; cat "/etc/yum.repos.d/$repofile"; echo "# ^^^^^ ^^^^^ ^^^^^ ^^^^^"; echo ""; done _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 04 20:00:02 ip-10-31-9-63.us-east-1.aws.redhat.com python3[4764]: ansible-ansible.legacy.command Invoked with _raw_params=yum -q repolist _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 04 20:00:03 ip-10-31-9-63.us-east-1.aws.redhat.com python3[4782]: ansible-stat Invoked with path=/etc/yum.repos.d/baseos-ci.repo follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Feb 04 20:00:03 ip-10-31-9-63.us-east-1.aws.redhat.com python3[4799]: ansible-ansible.legacy.command Invoked with _raw_params=for repofile in $(ls -1 /etc/yum.repos.d/); do echo "# vvvvv $repofile vvvvv vvvvv"; cat "/etc/yum.repos.d/$repofile"; echo "# ^^^^^ ^^^^^ ^^^^^ ^^^^^"; echo ""; done _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 04 20:00:04 ip-10-31-9-63.us-east-1.aws.redhat.com python3[4830]: ansible-ansible.legacy.command Invoked with _raw_params=yum -q repolist _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 04 20:00:04 ip-10-31-9-63.us-east-1.aws.redhat.com python3[4848]: ansible-ansible.legacy.command Invoked with _raw_params=for repofile in $(ls -1 /etc/yum.repos.d/); do echo "# vvvvv $repofile vvvvv vvvvv"; cat "/etc/yum.repos.d/$repofile"; echo "# ^^^^^ ^^^^^ ^^^^^ ^^^^^"; echo ""; done _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 04 20:00:05 ip-10-31-9-63.us-east-1.aws.redhat.com python3[4879]: ansible-ansible.legacy.command Invoked with _raw_params=yum -q repolist _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 04 20:00:05 ip-10-31-9-63.us-east-1.aws.redhat.com python3[4897]: ansible-ansible.legacy.command Invoked with _raw_params=for repofile in $(ls -1 /etc/yum.repos.d/); do echo "# vvvvv $repofile vvvvv vvvvv"; cat "/etc/yum.repos.d/$repofile"; echo "# ^^^^^ ^^^^^ ^^^^^ ^^^^^"; echo ""; done _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 04 20:00:05 ip-10-31-9-63.us-east-1.aws.redhat.com python3[4928]: ansible-ansible.legacy.command Invoked with _raw_params=yum -q repolist _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 04 20:00:06 ip-10-31-9-63.us-east-1.aws.redhat.com python3[4946]: ansible-ansible.legacy.command Invoked with _raw_params=for repofile in $(ls -1 /etc/yum.repos.d/); do echo "# vvvvv $repofile vvvvv vvvvv"; cat "/etc/yum.repos.d/$repofile"; echo "# ^^^^^ ^^^^^ ^^^^^ ^^^^^"; echo ""; done _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 04 20:00:06 ip-10-31-9-63.us-east-1.aws.redhat.com python3[4977]: ansible-ansible.legacy.command Invoked with _raw_params=yum -q repolist _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 04 20:00:06 ip-10-31-9-63.us-east-1.aws.redhat.com python3[4995]: ansible-ansible.legacy.command Invoked with _raw_params=dnf clean all --enablerepo=* _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 04 20:00:07 ip-10-31-9-63.us-east-1.aws.redhat.com python3[5013]: ansible-ansible.legacy.command Invoked with _raw_params=dnf makecache --enablerepo=* _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 04 20:01:01 ip-10-31-9-63.us-east-1.aws.redhat.com CROND[5195]: (root) CMD (run-parts /etc/cron.hourly) Feb 04 20:01:01 ip-10-31-9-63.us-east-1.aws.redhat.com run-parts[5198]: (/etc/cron.hourly) starting 0anacron Feb 04 20:01:01 ip-10-31-9-63.us-east-1.aws.redhat.com anacron[5206]: Anacron started on 2024-02-04 Feb 04 20:01:01 ip-10-31-9-63.us-east-1.aws.redhat.com run-parts[5208]: (/etc/cron.hourly) finished 0anacron Feb 04 20:01:01 ip-10-31-9-63.us-east-1.aws.redhat.com anacron[5206]: Will run job `cron.daily' in 18 min. Feb 04 20:01:01 ip-10-31-9-63.us-east-1.aws.redhat.com anacron[5206]: Will run job `cron.weekly' in 38 min. Feb 04 20:01:01 ip-10-31-9-63.us-east-1.aws.redhat.com anacron[5206]: Will run job `cron.monthly' in 58 min. Feb 04 20:01:01 ip-10-31-9-63.us-east-1.aws.redhat.com anacron[5206]: Jobs will be executed sequentially Feb 04 20:01:01 ip-10-31-9-63.us-east-1.aws.redhat.com CROND[5194]: (root) CMDEND (run-parts /etc/cron.hourly) Feb 04 20:01:08 ip-10-31-9-63.us-east-1.aws.redhat.com python3[5225]: ansible-ansible.legacy.dnf Invoked with name=['createrepo'] state=latest allow_downgrade=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 allowerasing=False nobest=False conf_file=None disable_excludes=None download_dir=None list=None releasever=None Feb 04 20:01:16 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: Started /usr/bin/systemctl start man-db-cache-update. ░░ Subject: A start job for unit run-r7aca08224459473da35f6ae53d9757a4.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit run-r7aca08224459473da35f6ae53d9757a4.service has finished successfully. ░░ ░░ The job identifier is 937. Feb 04 20:01:16 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: Starting man-db-cache-update.service... ░░ Subject: A start job for unit man-db-cache-update.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit man-db-cache-update.service has begun execution. ░░ ░░ The job identifier is 999. Feb 04 20:01:18 ip-10-31-9-63.us-east-1.aws.redhat.com python3[5255]: ansible-ansible.legacy.dnf Invoked with name=['make'] state=latest allow_downgrade=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 allowerasing=False nobest=False conf_file=None disable_excludes=None download_dir=None list=None releasever=None Feb 04 20:01:19 ip-10-31-9-63.us-east-1.aws.redhat.com python3[5344]: ansible-ansible.legacy.dnf Invoked with name=['parted'] state=latest allow_downgrade=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 allowerasing=False nobest=False conf_file=None disable_excludes=None download_dir=None list=None releasever=None Feb 04 20:01:19 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: man-db-cache-update.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit man-db-cache-update.service has successfully entered the 'dead' state. Feb 04 20:01:19 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: Finished man-db-cache-update.service. ░░ Subject: A start job for unit man-db-cache-update.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit man-db-cache-update.service has finished successfully. ░░ ░░ The job identifier is 999. Feb 04 20:01:19 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: run-r7aca08224459473da35f6ae53d9757a4.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit run-r7aca08224459473da35f6ae53d9757a4.service has successfully entered the 'dead' state. Feb 04 20:01:20 ip-10-31-9-63.us-east-1.aws.redhat.com python3[5363]: ansible-ansible.legacy.dnf Invoked with name=['psmisc'] state=latest allow_downgrade=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 allowerasing=False nobest=False conf_file=None disable_excludes=None download_dir=None list=None releasever=None Feb 04 20:01:21 ip-10-31-9-63.us-east-1.aws.redhat.com python3[5381]: ansible-ansible.legacy.dnf Invoked with name=['chrony'] state=latest allow_downgrade=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 allowerasing=False nobest=False conf_file=None disable_excludes=None download_dir=None list=None releasever=None Feb 04 20:01:23 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: Started /usr/bin/systemctl start man-db-cache-update. ░░ Subject: A start job for unit run-r9aba2d573e3a4f31b0702422a50dcb51.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit run-r9aba2d573e3a4f31b0702422a50dcb51.service has finished successfully. ░░ ░░ The job identifier is 1061. Feb 04 20:01:23 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: Starting man-db-cache-update.service... ░░ Subject: A start job for unit man-db-cache-update.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit man-db-cache-update.service has begun execution. ░░ ░░ The job identifier is 1123. Feb 04 20:01:23 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: Reloading. Feb 04 20:01:23 ip-10-31-9-63.us-east-1.aws.redhat.com systemd-rc-local-generator[5419]: /etc/rc.d/rc.local is not marked executable, skipping. Feb 04 20:01:23 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: Queuing reload/restart jobs for marked units… Feb 04 20:01:23 ip-10-31-9-63.us-east-1.aws.redhat.com chronyd[592]: chronyd exiting Feb 04 20:01:23 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: Stopping NTP client/server... ░░ Subject: A stop job for unit chronyd.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A stop job for unit chronyd.service has begun execution. ░░ ░░ The job identifier is 1185. Feb 04 20:01:23 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: chronyd.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit chronyd.service has successfully entered the 'dead' state. Feb 04 20:01:23 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: Stopped NTP client/server. ░░ Subject: A stop job for unit chronyd.service has finished ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A stop job for unit chronyd.service has finished. ░░ ░░ The job identifier is 1185 and the job result is done. Feb 04 20:01:23 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: Starting NTP client/server... ░░ Subject: A start job for unit chronyd.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit chronyd.service has begun execution. ░░ ░░ The job identifier is 1185. Feb 04 20:01:23 ip-10-31-9-63.us-east-1.aws.redhat.com chronyd[5462]: chronyd version 4.5 starting (+CMDMON +NTP +REFCLOCK +RTC +PRIVDROP +SCFILTER +SIGND +ASYNCDNS +NTS +SECHASH +IPV6 +DEBUG) Feb 04 20:01:23 ip-10-31-9-63.us-east-1.aws.redhat.com chronyd[5462]: Loaded 0 symmetric keys Feb 04 20:01:23 ip-10-31-9-63.us-east-1.aws.redhat.com chronyd[5462]: Frequency 3.514 +/- 6.968 ppm read from /var/lib/chrony/drift Feb 04 20:01:23 ip-10-31-9-63.us-east-1.aws.redhat.com chronyd[5462]: Using right/UTC timezone to obtain leap second data Feb 04 20:01:23 ip-10-31-9-63.us-east-1.aws.redhat.com chronyd[5462]: Loaded seccomp filter (level 2) Feb 04 20:01:23 ip-10-31-9-63.us-east-1.aws.redhat.com chronyd[5462]: Added source 10.11.160.238 Feb 04 20:01:23 ip-10-31-9-63.us-east-1.aws.redhat.com chronyd[5462]: Added source 10.18.100.10 Feb 04 20:01:23 ip-10-31-9-63.us-east-1.aws.redhat.com chronyd[5462]: Added source 10.2.32.37 Feb 04 20:01:23 ip-10-31-9-63.us-east-1.aws.redhat.com chronyd[5462]: Added source 10.2.32.38 Feb 04 20:01:23 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: Started NTP client/server. ░░ Subject: A start job for unit chronyd.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit chronyd.service has finished successfully. ░░ ░░ The job identifier is 1185. Feb 04 20:01:23 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: Started /usr/bin/systemctl start man-db-cache-update. ░░ Subject: A start job for unit run-r1c18577214164bbc9081e4c1f515084b.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit run-r1c18577214164bbc9081e4c1f515084b.service has finished successfully. ░░ ░░ The job identifier is 1249. Feb 04 20:01:23 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: man-db-cache-update.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit man-db-cache-update.service has successfully entered the 'dead' state. Feb 04 20:01:23 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: Finished man-db-cache-update.service. ░░ Subject: A start job for unit man-db-cache-update.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit man-db-cache-update.service has finished successfully. ░░ ░░ The job identifier is 1123. Feb 04 20:01:23 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: run-r9aba2d573e3a4f31b0702422a50dcb51.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit run-r9aba2d573e3a4f31b0702422a50dcb51.service has successfully entered the 'dead' state. Feb 04 20:01:23 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: run-r1c18577214164bbc9081e4c1f515084b.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit run-r1c18577214164bbc9081e4c1f515084b.service has successfully entered the 'dead' state. Feb 04 20:01:24 ip-10-31-9-63.us-east-1.aws.redhat.com python3[5634]: ansible-ansible.legacy.command Invoked with _raw_params=bash -c "chronyc sources" _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 04 20:01:24 ip-10-31-9-63.us-east-1.aws.redhat.com python3[5652]: ansible-ansible.legacy.command Invoked with _raw_params=bash -c "chronyc sources | grep redhat | grep -v util.phx2.redhat || chronyc add server clock.corp.redhat.com iburst" _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 04 20:01:25 ip-10-31-9-63.us-east-1.aws.redhat.com python3[5673]: ansible-ansible.legacy.command Invoked with _raw_params=bash -c "chronyc sources" _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 04 20:01:29 ip-10-31-9-63.us-east-1.aws.redhat.com chronyd[5462]: Selected source 96.126.122.39 (2.centos.pool.ntp.org) Feb 04 20:01:29 ip-10-31-9-63.us-east-1.aws.redhat.com chronyd[5462]: System clock TAI offset set to 37 seconds Feb 04 20:01:55 ip-10-31-9-63.us-east-1.aws.redhat.com python3[5691]: ansible-ansible.legacy.command Invoked with _raw_params=bash -c "chronyc sources" _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 04 20:01:56 ip-10-31-9-63.us-east-1.aws.redhat.com python3[5709]: ansible-ansible.legacy.command Invoked with _raw_params=chronyc waitsync 5 _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 04 20:01:56 ip-10-31-9-63.us-east-1.aws.redhat.com python3[5727]: ansible-ansible.legacy.command Invoked with _raw_params=setenforce 0; timedatectl set-timezone UTC; setenforce 1 _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 04 20:01:56 ip-10-31-9-63.us-east-1.aws.redhat.com dbus-broker-launch[578]: avc: op=setenforce lsm=selinux enforcing=0 res=1 Feb 04 20:01:56 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: Starting Time & Date Service... ░░ Subject: A start job for unit systemd-timedated.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-timedated.service has begun execution. ░░ ░░ The job identifier is 1373. Feb 04 20:01:56 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: Started Time & Date Service. ░░ Subject: A start job for unit systemd-timedated.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-timedated.service has finished successfully. ░░ ░░ The job identifier is 1373. Feb 04 20:01:56 ip-10-31-9-63.us-east-1.aws.redhat.com systemd-timedated[5731]: Changed time zone to 'UTC' (UTC). ░░ Subject: Time zone change to UTC ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The system timezone has been changed to UTC. Feb 04 20:01:57 ip-10-31-9-63.us-east-1.aws.redhat.com python3[5748]: ansible-yum_repository Invoked with state=present name=beaker-tasks description=Beaker tasks baseurl=['http://beaker.engineering.redhat.com/rpms/'] gpgcheck=False enabled=True reposdir=/etc/yum.repos.d unsafe_writes=False bandwidth=None cost=None deltarpm_metadata_percentage=None deltarpm_percentage=None enablegroups=None exclude=None failovermethod=None file=None gpgcakey=None gpgkey=None module_hotfixes=None http_caching=None include=None includepkgs=None ip_resolve=None keepalive=None keepcache=None metadata_expire=None metadata_expire_filter=None metalink=None mirrorlist=None mirrorlist_expire=None params=None password=NOT_LOGGING_PARAMETER priority=None protect=None proxy=None proxy_password=NOT_LOGGING_PARAMETER proxy_username=None repo_gpgcheck=None retries=None s3_enabled=None skip_if_unavailable=None sslcacert=None ssl_check_cert_permissions=None sslclientcert=None sslclientkey=None sslverify=None throttle=None timeout=None ui_repoid_vars=None username=None async=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Feb 04 20:01:57 ip-10-31-9-63.us-east-1.aws.redhat.com python3[5765]: ansible-user Invoked with name=root update_password=always password=NOT_LOGGING_PARAMETER state=present non_unique=False force=False remove=False create_home=True system=False move_home=False append=False ssh_key_bits=0 ssh_key_type=rsa ssh_key_comment=ansible-generated on ip-10-31-9-63.us-east-1.aws.redhat.com uid=None group=None groups=None comment=None home=None shell=None login_class=None password_expire_max=None password_expire_min=None hidden=None seuser=None skeleton=None generate_ssh_key=None ssh_key_file=None ssh_key_passphrase=NOT_LOGGING_PARAMETER expires=None password_lock=None local=None profile=None authorization=None role=None umask=None Feb 04 20:01:57 ip-10-31-9-63.us-east-1.aws.redhat.com usermod[5768]: change user 'root' password Feb 04 20:01:58 ip-10-31-9-63.us-east-1.aws.redhat.com python3[5791]: ansible-lineinfile Invoked with dest=/etc/ssh/sshd_config regexp=#?PasswordAuthentication (?:yes|no) line=PasswordAuthentication yes state=present path=/etc/ssh/sshd_config backrefs=False create=False backup=False firstmatch=False unsafe_writes=False search_string=None insertafter=None insertbefore=None validate=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Feb 04 20:01:58 ip-10-31-9-63.us-east-1.aws.redhat.com python3[5808]: ansible-lineinfile Invoked with dest=/etc/ssh/sshd_config line=PermitRootLogin yes state=present path=/etc/ssh/sshd_config backrefs=False create=False backup=False firstmatch=False unsafe_writes=False regexp=None search_string=None insertafter=None insertbefore=None validate=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Feb 04 20:01:58 ip-10-31-9-63.us-east-1.aws.redhat.com python3[5825]: ansible-file Invoked with path=/var/lib/tft state=directory recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Feb 04 20:01:58 ip-10-31-9-63.us-east-1.aws.redhat.com python3[5842]: ansible-stat Invoked with path=/var/lib/tft/lib.sh follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Feb 04 20:01:59 ip-10-31-9-63.us-east-1.aws.redhat.com python3[5897]: ansible-ansible.legacy.stat Invoked with path=/var/lib/tft/lib.sh follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Feb 04 20:01:59 ip-10-31-9-63.us-east-1.aws.redhat.com python3[5947]: ansible-ansible.legacy.copy Invoked with src=/root/.ansible/tmp/ansible-tmp-1707076919.0212371-7340-75232704328444/source dest=/var/lib/tft/lib.sh owner=root group=root mode=u=rx,go= follow=False _original_basename=lib.sh.j2 checksum=cdcb4e3b7fb4fbc298e05439468daaeb6289f07c backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Feb 04 20:01:59 ip-10-31-9-63.us-east-1.aws.redhat.com python3[5981]: ansible-stat Invoked with path=/usr/local/bin/ci-extendtesttime.sh follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Feb 04 20:02:00 ip-10-31-9-63.us-east-1.aws.redhat.com python3[6036]: ansible-ansible.legacy.stat Invoked with path=/usr/local/bin/ci-extendtesttime.sh follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Feb 04 20:02:00 ip-10-31-9-63.us-east-1.aws.redhat.com python3[6086]: ansible-ansible.legacy.copy Invoked with src=/root/.ansible/tmp/ansible-tmp-1707076919.8875058-7353-105796243772402/source dest=/usr/local/bin/ci-extendtesttime.sh owner=root group=root mode=u=rx,go= follow=False _original_basename=ci-extendtesttime.sh.j2 checksum=2d0f6c296873c17e8b22c9490b000973b2a8a350 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Feb 04 20:02:00 ip-10-31-9-63.us-east-1.aws.redhat.com python3[6120]: ansible-stat Invoked with path=/usr/bin/extendtesttime.sh follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Feb 04 20:02:00 ip-10-31-9-63.us-east-1.aws.redhat.com python3[6137]: ansible-stat Invoked with path=/usr/local/bin/ci-return2pool.sh follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Feb 04 20:02:01 ip-10-31-9-63.us-east-1.aws.redhat.com python3[6192]: ansible-ansible.legacy.stat Invoked with path=/usr/local/bin/ci-return2pool.sh follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Feb 04 20:02:01 ip-10-31-9-63.us-east-1.aws.redhat.com python3[6242]: ansible-ansible.legacy.copy Invoked with src=/root/.ansible/tmp/ansible-tmp-1707076920.861413-7380-4085268333930/source dest=/usr/local/bin/ci-return2pool.sh owner=root group=root mode=u=rx,go= follow=False _original_basename=ci-return2pool.sh.j2 checksum=f77cae3b2a729c60bcb0c01c82bf347f13a9b2c5 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Feb 04 20:02:01 ip-10-31-9-63.us-east-1.aws.redhat.com python3[6276]: ansible-stat Invoked with path=/etc/motd follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Feb 04 20:02:01 ip-10-31-9-63.us-east-1.aws.redhat.com python3[6333]: ansible-ansible.legacy.stat Invoked with path=/root/.guest-metadata.json follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Feb 04 20:02:02 ip-10-31-9-63.us-east-1.aws.redhat.com python3[6383]: ansible-ansible.legacy.copy Invoked with dest=/root/.guest-metadata.json owner=root group=root mode=u=rw,go= src=/root/.ansible/tmp/ansible-tmp-1707076921.6423726-7398-79811369140814/source _original_basename=tmps6_l2hyq follow=False checksum=bf21a9e8fbc5a3846fb05b4fa0859e0917b2202f backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Feb 04 20:02:02 ip-10-31-9-63.us-east-1.aws.redhat.com python3[6417]: ansible-ansible.legacy.command Invoked with _raw_params=curl -sLI --connect-timeout 5 -w '%{response_code}' http://169.254.169.254/latest/meta-data/instance-id | grep ^200 _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 04 20:02:02 ip-10-31-9-63.us-east-1.aws.redhat.com python3[6437]: ansible-hostname Invoked with name=0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm use=None Feb 04 20:02:02 ip-10-31-9-63.us-east-1.aws.redhat.com dbus-broker-launch[578]: avc: op=setenforce lsm=selinux enforcing=1 res=1 Feb 04 20:02:02 ip-10-31-9-63.us-east-1.aws.redhat.com systemd-journald[484]: Field hash table of /run/log/journal/75ecf5059a924499888af4d7df65fc87/system.journal has a fill level at 75.1 (250 of 333 items), suggesting rotation. Feb 04 20:02:02 ip-10-31-9-63.us-east-1.aws.redhat.com systemd-journald[484]: /run/log/journal/75ecf5059a924499888af4d7df65fc87/system.journal: Journal header limits reached or header out-of-date, rotating. Feb 04 20:02:02 ip-10-31-9-63.us-east-1.aws.redhat.com rsyslogd[786]: imjournal: journal files changed, reloading... [v8.2310.0-3.el9 try https://www.rsyslog.com/e/0 ] Feb 04 20:02:02 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: Starting Hostname Service... ░░ Subject: A start job for unit systemd-hostnamed.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-hostnamed.service has begun execution. ░░ ░░ The job identifier is 1436. Feb 04 20:02:02 ip-10-31-9-63.us-east-1.aws.redhat.com systemd[1]: Started Hostname Service. ░░ Subject: A start job for unit systemd-hostnamed.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-hostnamed.service has finished successfully. ░░ ░░ The job identifier is 1436. Feb 04 20:02:02 ip-10-31-9-63.us-east-1.aws.redhat.com systemd-hostnamed[6442]: Changed pretty hostname to '0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm' Feb 04 20:02:02 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm systemd-hostnamed[6442]: Hostname set to <0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm> (static) Feb 04 20:02:02 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm NetworkManager[618]: [1707076922.9779] hostname: static hostname changed from "ip-10-31-9-63.us-east-1.aws.redhat.com" to "0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm" Feb 04 20:02:02 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm systemd[1]: Starting Network Manager Script Dispatcher Service... ░░ Subject: A start job for unit NetworkManager-dispatcher.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit NetworkManager-dispatcher.service has begun execution. ░░ ░░ The job identifier is 1499. Feb 04 20:02:02 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm systemd[1]: Started Network Manager Script Dispatcher Service. ░░ Subject: A start job for unit NetworkManager-dispatcher.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit NetworkManager-dispatcher.service has finished successfully. ░░ ░░ The job identifier is 1499. Feb 04 20:02:03 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[6471]: ansible-ansible.legacy.command Invoked with _raw_params=ping -4 -c1 -W10 $(hostname) _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 04 20:02:03 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[6490]: ansible-file Invoked with path=/var/log/messages state=touch recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Feb 04 20:02:04 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[6507]: ansible-ansible.legacy.command Invoked with _raw_params=setsebool nis_enabled on _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 04 20:02:04 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm setsebool[6508]: The nis_enabled policy boolean was changed to on by root Feb 04 20:02:04 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[6525]: ansible-stat Invoked with path=/usr/bin/rstrnt-package follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Feb 04 20:02:04 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[6544]: ansible-ansible.legacy.command Invoked with _raw_params=sed -e 's/rpm -q $package/rpm -q --whatprovides $package/' -i /usr/bin/rstrnt-package _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 04 20:02:04 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[6562]: ansible-stat Invoked with path=/var/lib/restraint follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Feb 04 20:02:04 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[6581]: ansible-file Invoked with path=/var/lib/restraint/avc_since state=touch recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Feb 04 20:02:05 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[6598]: ansible-stat Invoked with path=/usr/share/beakerlib/beakerlib.sh follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Feb 04 20:02:05 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[6617]: ansible-file Invoked with dest=/usr/lib/beakerlib state=directory path=/usr/lib/beakerlib recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Feb 04 20:02:05 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[6634]: ansible-file Invoked with dest=/usr/share/rhts-library state=directory path=/usr/share/rhts-library recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Feb 04 20:02:05 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[6651]: ansible-file Invoked with src=/usr/share/beakerlib/beakerlib.sh dest=/usr/lib/beakerlib/beakerlib.sh state=link path=/usr/lib/beakerlib/beakerlib.sh recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Feb 04 20:02:06 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[6668]: ansible-file Invoked with src=/usr/share/beakerlib/beakerlib.sh dest=/usr/share/rhts-library/rhtslib.sh state=link path=/usr/share/rhts-library/rhtslib.sh recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Feb 04 20:02:06 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[6685]: ansible-ansible.legacy.command Invoked with _raw_params=mv /var/log/audit/audit.log /var/log/audit/audit.log.bak _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 04 20:02:06 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[6703]: ansible-ansible.legacy.command Invoked with _raw_params=/usr/sbin/service auditd restart _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 04 20:02:06 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm auditd[539]: The audit daemon is exiting. Feb 04 20:02:06 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm kernel: audit: type=1305 audit(1707076926.595:724): op=set audit_pid=0 old=539 auid=4294967295 ses=4294967295 subj=system_u:system_r:auditd_t:s0 res=1 Feb 04 20:02:06 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm systemd[1]: auditd.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit auditd.service has successfully entered the 'dead' state. Feb 04 20:02:06 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm kernel: audit: type=1131 audit(1707076926.596:725): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:init_t:s0 msg='unit=auditd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 04 20:02:06 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm dbus-broker-launch[578]: avc: op=load_policy lsm=selinux seqno=2 res=1 Feb 04 20:02:08 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm systemd[1]: Starting Security Auditing Service... ░░ Subject: A start job for unit auditd.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit auditd.service has begun execution. ░░ ░░ The job identifier is 1562. Feb 04 20:02:08 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm auditd[6719]: No plugins found, not dispatching events Feb 04 20:02:08 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm auditd[6719]: Init complete, auditd 3.1.2 listening for events (startup state enable) Feb 04 20:02:08 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm kernel: audit: type=1305 audit(1707076928.595:726): op=set audit_enabled=1 old=1 auid=4294967295 ses=4294967295 subj=system_u:system_r:auditd_t:s0 res=1 Feb 04 20:02:08 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm kernel: audit: type=1300 audit(1707076928.595:726): arch=c000003e syscall=44 success=yes exit=60 a0=3 a1=7ffc9859cde0 a2=3c a3=0 items=0 ppid=6718 pid=6719 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="auditd" exe="/usr/sbin/auditd" subj=system_u:system_r:auditd_t:s0 key=(null) Feb 04 20:02:08 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm kernel: audit: type=1327 audit(1707076928.595:726): proctitle="/sbin/auditd" Feb 04 20:02:08 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm augenrules[6722]: /sbin/augenrules: No change Feb 04 20:02:08 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm augenrules[6732]: No rules Feb 04 20:02:08 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm augenrules[6732]: enabled 1 Feb 04 20:02:08 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm augenrules[6732]: failure 1 Feb 04 20:02:08 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm augenrules[6732]: pid 6719 Feb 04 20:02:08 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm augenrules[6732]: rate_limit 0 Feb 04 20:02:08 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm augenrules[6732]: backlog_limit 8192 Feb 04 20:02:08 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm augenrules[6732]: lost 0 Feb 04 20:02:08 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm augenrules[6732]: backlog 0 Feb 04 20:02:08 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm augenrules[6732]: backlog_wait_time 60000 Feb 04 20:02:08 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm augenrules[6732]: backlog_wait_time_actual 0 Feb 04 20:02:08 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm augenrules[6732]: enabled 1 Feb 04 20:02:08 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm augenrules[6732]: failure 1 Feb 04 20:02:08 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm augenrules[6732]: pid 6719 Feb 04 20:02:08 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm augenrules[6732]: rate_limit 0 Feb 04 20:02:08 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm augenrules[6732]: backlog_limit 8192 Feb 04 20:02:08 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm augenrules[6732]: lost 0 Feb 04 20:02:08 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm augenrules[6732]: backlog 3 Feb 04 20:02:08 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm augenrules[6732]: backlog_wait_time 60000 Feb 04 20:02:08 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm augenrules[6732]: backlog_wait_time_actual 0 Feb 04 20:02:08 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm augenrules[6732]: enabled 1 Feb 04 20:02:08 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm augenrules[6732]: failure 1 Feb 04 20:02:08 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm augenrules[6732]: pid 6719 Feb 04 20:02:08 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm augenrules[6732]: rate_limit 0 Feb 04 20:02:08 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm augenrules[6732]: backlog_limit 8192 Feb 04 20:02:08 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm augenrules[6732]: lost 0 Feb 04 20:02:08 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm augenrules[6732]: backlog 4 Feb 04 20:02:08 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm augenrules[6732]: backlog_wait_time 60000 Feb 04 20:02:08 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm augenrules[6732]: backlog_wait_time_actual 0 Feb 04 20:02:08 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm systemd[1]: Started Security Auditing Service. ░░ Subject: A start job for unit auditd.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit auditd.service has finished successfully. ░░ ░░ The job identifier is 1562. Feb 04 20:02:08 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[6755]: ansible-stat Invoked with path=/etc/NetworkManager/conf.d follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Feb 04 20:02:09 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[6812]: ansible-ansible.legacy.stat Invoked with path=/etc/NetworkManager/conf.d/99-zzz-tft-workaround-dns-default.conf follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Feb 04 20:02:09 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[6862]: ansible-ansible.legacy.copy Invoked with dest=/etc/NetworkManager/conf.d/99-zzz-tft-workaround-dns-default.conf src=/root/.ansible/tmp/ansible-tmp-1707076928.8947954-7529-10325819842836/source _original_basename=tmp80zkbq3j follow=False checksum=0145f3ae57eef5aa08bbb678fedbb3edd001cd2d backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Feb 04 20:02:09 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[6896]: ansible-ansible.legacy.command Invoked with _raw_params=for repofile in $(ls -1 /etc/yum.repos.d/); do echo "# vvvvv $repofile vvvvv vvvvv"; cat "/etc/yum.repos.d/$repofile"; echo "# ^^^^^ ^^^^^ ^^^^^ ^^^^^"; echo ""; done _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 04 20:02:10 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[6927]: ansible-ansible.legacy.command Invoked with _raw_params=yum -q repolist _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 04 20:02:10 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[6945]: ansible-ansible.legacy.command Invoked with _raw_params=dnf config-manager --set-enabled highavailability _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 04 20:02:11 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[6963]: ansible-ansible.legacy.command Invoked with _raw_params=dnf clean all --enablerepo=* _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 04 20:02:11 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[6981]: ansible-ansible.legacy.command Invoked with _raw_params=dnf makecache --enablerepo=* _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 04 20:02:13 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm systemd[1]: NetworkManager-dispatcher.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit NetworkManager-dispatcher.service has successfully entered the 'dead' state. Feb 04 20:02:26 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm systemd[1]: systemd-timedated.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit systemd-timedated.service has successfully entered the 'dead' state. Feb 04 20:02:32 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm systemd[1]: systemd-hostnamed.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit systemd-hostnamed.service has successfully entered the 'dead' state. Feb 04 20:02:36 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm chronyd[5462]: Selected source 10.2.32.38 Feb 04 20:03:15 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[7200]: ansible-ansible.legacy.setup Invoked with gather_subset=['all'] gather_timeout=10 filter=[] fact_path=/etc/ansible/facts.d Feb 04 20:03:16 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[7241]: ansible-service_facts Invoked Feb 04 20:03:17 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[7340]: ansible-ansible.legacy.command Invoked with _raw_params=rpm -qa _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 04 20:03:18 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[7358]: ansible-ansible.legacy.command Invoked with _raw_params=if [ $(find /etc/yum.repos.d/ -name 'epel.repo' | wc -l ) -gt 0 ]; then dnf config-manager --set-disabled epel; fi _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 04 20:03:19 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[7380]: ansible-ansible.legacy.command Invoked with _raw_params=dnf clean all --enablerepo=* _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 04 20:03:20 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[7398]: ansible-ansible.legacy.command Invoked with _raw_params=dnf makecache --enablerepo=* _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 04 20:04:24 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[7613]: ansible-ansible.legacy.setup Invoked with gather_subset=['all'] gather_timeout=10 filter=[] fact_path=/etc/ansible/facts.d Feb 04 20:04:24 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[7654]: ansible-service_facts Invoked Feb 04 20:04:26 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[7753]: ansible-ansible.legacy.command Invoked with _raw_params=rpm -qa _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 04 20:04:29 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[7789]: ansible-ansible.legacy.setup Invoked with gather_subset=['all'] gather_timeout=10 filter=[] fact_path=/etc/ansible/facts.d Feb 04 20:04:30 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[7830]: ansible-ansible.legacy.dnf Invoked with name=['ca-certificates'] state=latest allow_downgrade=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 allowerasing=False nobest=False conf_file=None disable_excludes=None download_dir=None list=None releasever=None Feb 04 20:04:31 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[7848]: ansible-ansible.legacy.dnf Invoked with name=['curl'] state=latest allow_downgrade=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 allowerasing=False nobest=False conf_file=None disable_excludes=None download_dir=None list=None releasever=None Feb 04 20:04:32 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[7866]: ansible-ansible.legacy.command Invoked with _raw_params=curl -skL -o /etc/pki/ca-trust/source/anchors/Current-IT-Root-CAs.pem https://certs.corp.redhat.com/certs/Current-IT-Root-CAs.pem _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 04 20:04:33 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[7885]: ansible-ansible.legacy.command Invoked with _raw_params=bash -c 'update-ca-trust' _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 04 20:04:34 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[7909]: ansible-ansible.legacy.command Invoked with _raw_params=bash -c 'update-ca-trust extract' _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 04 20:04:35 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[7933]: ansible-ansible.legacy.command Invoked with _raw_params=curl -v https://gitlab.cee.redhat.com _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 04 20:04:35 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[7952]: ansible-ansible.legacy.command Invoked with _raw_params=curl -v https://beaker.engineering.redhat.com/ _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 04 20:04:36 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[7971]: ansible-service_facts Invoked Feb 04 20:04:38 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[8070]: ansible-ansible.legacy.command Invoked with _raw_params=rpm -qa _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 04 20:04:39 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[8088]: ansible-service_facts Invoked Feb 04 20:04:40 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[8187]: ansible-ansible.legacy.command Invoked with _raw_params=rpm -qa _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 04 20:04:41 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[8205]: ansible-ansible.legacy.command Invoked with _raw_params=for repofile in $(ls -1 /etc/yum.repos.d/); do echo "# vvvvv $repofile vvvvv vvvvv"; cat "/etc/yum.repos.d/$repofile"; echo "# ^^^^^ ^^^^^ ^^^^^ ^^^^^"; echo ""; done _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 04 20:04:41 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[8236]: ansible-ansible.legacy.command Invoked with _raw_params=yum -q repolist _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 04 20:04:42 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[8254]: ansible-ansible.legacy.command Invoked with _raw_params=dnf clean all --enablerepo=* _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 04 20:04:43 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[8272]: ansible-ansible.legacy.command Invoked with _raw_params=dnf makecache --enablerepo=* _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 04 20:04:44 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm systemd[4037]: Created slice User Background Tasks Slice. ░░ Subject: A start job for unit UNIT has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has finished successfully. ░░ ░░ The job identifier is 14. Feb 04 20:04:44 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm systemd[4037]: Starting Cleanup of User's Temporary Files and Directories... ░░ Subject: A start job for unit UNIT has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has begun execution. ░░ ░░ The job identifier is 13. Feb 04 20:04:44 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm systemd[4037]: Finished Cleanup of User's Temporary Files and Directories. ░░ Subject: A start job for unit UNIT has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has finished successfully. ░░ ░░ The job identifier is 13. Feb 04 20:05:46 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[8489]: ansible-setup Invoked with gather_subset=['all'] gather_timeout=10 filter=* fact_path=/etc/ansible/facts.d Feb 04 20:05:46 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[8529]: ansible-stat Invoked with path=/run/ostree-booted follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Feb 04 20:05:47 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[8546]: ansible-dnf Invoked with name=['python3-pyasn1', 'python3-cryptography', 'python3-dbus'] state=present allow_downgrade=False autoremove=False bugfix=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True lock_timeout=30 conf_file=None disable_excludes=None download_dir=None list=None releasever=None Feb 04 20:05:52 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[8575]: ansible-dnf Invoked with name=['certmonger', 'python3-packaging'] state=present allow_downgrade=False autoremove=False bugfix=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True lock_timeout=30 conf_file=None disable_excludes=None download_dir=None list=None releasever=None Feb 04 20:05:54 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm dbus-broker-launch[577]: Noticed file-system modification, trigger reload. ░░ Subject: A configuration directory was written to ░░ Defined-By: dbus-broker ░░ Support: https://groups.google.com/forum/#!forum/bus1-devel ░░ ░░ A write was detected to one of the directories containing D-Bus configuration ░░ files, triggering a configuration reload. ░░ ░░ This functionality exists for backwards compatibility to pick up changes to ░░ D-Bus configuration without an explicit reolad request. Typically when ░░ installing or removing third-party software causes D-Bus configuration files ░░ to be added or removed. ░░ ░░ It is worth noting that this may cause partial configuration to be loaded in ░░ case dispatching this notification races with the writing of the configuration ░░ files. However, a future notification will then cause the configuration to be ░░ reladed again. Feb 04 20:05:54 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm dbus-broker-launch[577]: Noticed file-system modification, trigger reload. ░░ Subject: A configuration directory was written to ░░ Defined-By: dbus-broker ░░ Support: https://groups.google.com/forum/#!forum/bus1-devel ░░ ░░ A write was detected to one of the directories containing D-Bus configuration ░░ files, triggering a configuration reload. ░░ ░░ This functionality exists for backwards compatibility to pick up changes to ░░ D-Bus configuration without an explicit reolad request. Typically when ░░ installing or removing third-party software causes D-Bus configuration files ░░ to be added or removed. ░░ ░░ It is worth noting that this may cause partial configuration to be loaded in ░░ case dispatching this notification races with the writing of the configuration ░░ files. However, a future notification will then cause the configuration to be ░░ reladed again. Feb 04 20:05:54 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm dbus-broker-launch[577]: Noticed file-system modification, trigger reload. ░░ Subject: A configuration directory was written to ░░ Defined-By: dbus-broker ░░ Support: https://groups.google.com/forum/#!forum/bus1-devel ░░ ░░ A write was detected to one of the directories containing D-Bus configuration ░░ files, triggering a configuration reload. ░░ ░░ This functionality exists for backwards compatibility to pick up changes to ░░ D-Bus configuration without an explicit reolad request. Typically when ░░ installing or removing third-party software causes D-Bus configuration files ░░ to be added or removed. ░░ ░░ It is worth noting that this may cause partial configuration to be loaded in ░░ case dispatching this notification races with the writing of the configuration ░░ files. However, a future notification will then cause the configuration to be ░░ reladed again. Feb 04 20:05:54 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm systemd[1]: Reloading. Feb 04 20:05:54 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm systemd-rc-local-generator[8602]: /etc/rc.d/rc.local is not marked executable, skipping. Feb 04 20:05:54 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm systemd[1]: Started /usr/bin/systemctl start man-db-cache-update. ░░ Subject: A start job for unit run-r6895349a70fe46078d2c60e0cfb3f134.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit run-r6895349a70fe46078d2c60e0cfb3f134.service has finished successfully. ░░ ░░ The job identifier is 1570. Feb 04 20:05:54 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm systemd[1]: Starting man-db-cache-update.service... ░░ Subject: A start job for unit man-db-cache-update.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit man-db-cache-update.service has begun execution. ░░ ░░ The job identifier is 1632. Feb 04 20:05:54 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm systemd[1]: Reloading. Feb 04 20:05:54 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm systemd-rc-local-generator[8639]: /etc/rc.d/rc.local is not marked executable, skipping. Feb 04 20:05:54 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm systemd[1]: Queuing reload/restart jobs for marked units… Feb 04 20:05:55 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm systemd[1]: man-db-cache-update.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit man-db-cache-update.service has successfully entered the 'dead' state. Feb 04 20:05:55 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm systemd[1]: Finished man-db-cache-update.service. ░░ Subject: A start job for unit man-db-cache-update.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit man-db-cache-update.service has finished successfully. ░░ ░░ The job identifier is 1632. Feb 04 20:05:55 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm systemd[1]: run-r6895349a70fe46078d2c60e0cfb3f134.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit run-r6895349a70fe46078d2c60e0cfb3f134.service has successfully entered the 'dead' state. Feb 04 20:05:55 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[8979]: ansible-file Invoked with name=/etc/certmonger//pre-scripts owner=root group=root mode=0700 state=directory path=/etc/certmonger//pre-scripts recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Feb 04 20:05:56 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[8996]: ansible-file Invoked with name=/etc/certmonger//post-scripts owner=root group=root mode=0700 state=directory path=/etc/certmonger//post-scripts recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Feb 04 20:05:56 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[9013]: ansible-systemd Invoked with name=certmonger state=started enabled=True daemon_reload=False daemon_reexec=False no_block=False force=None masked=None user=None scope=None Feb 04 20:05:56 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm systemd[1]: Reloading. Feb 04 20:05:56 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm systemd-rc-local-generator[9032]: /etc/rc.d/rc.local is not marked executable, skipping. Feb 04 20:05:56 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm systemd[1]: Starting Certificate monitoring and PKI enrollment... ░░ Subject: A start job for unit certmonger.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit certmonger.service has begun execution. ░░ ░░ The job identifier is 1694. Feb 04 20:05:56 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm systemd[1]: Started Certificate monitoring and PKI enrollment. ░░ Subject: A start job for unit certmonger.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit certmonger.service has finished successfully. ░░ ░░ The job identifier is 1694. Feb 04 20:05:57 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[9086]: ansible-fedora.linux_system_roles.certificate_request Invoked with name=quadlet_demo dns=['localhost'] directory=/etc/pki/tls wait=True ca=self-sign __header=# # Ansible managed # # system_role:certificate provider_config_directory=/etc/certmonger provider=certmonger key_usage=['digitalSignature', 'keyEncipherment'] extended_key_usage=['id-kp-serverAuth', 'id-kp-clientAuth'] auto_renew=True ip=None email=None common_name=None country=None state=None locality=None organization=None organizational_unit=None contact_email=None key_size=None owner=None group=None mode=None principal=None run_before=None run_after=None Feb 04 20:05:57 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm certmonger[9045]: 2024-02-04 20:05:57 [9045] Wrote to /var/lib/certmonger/requests/20240204200557 Feb 04 20:05:57 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm certmonger[9045]: 2024-02-04 20:05:57 [9045] Wrote to /var/lib/certmonger/requests/20240204200557 Feb 04 20:05:57 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm certmonger[9045]: 2024-02-04 20:05:57 [9045] Wrote to /var/lib/certmonger/requests/20240204200557 Feb 04 20:05:57 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm certmonger[9045]: 2024-02-04 20:05:57 [9045] Wrote to /var/lib/certmonger/requests/20240204200557 Feb 04 20:05:57 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm certmonger[9045]: 2024-02-04 20:05:57 [9045] Wrote to /var/lib/certmonger/requests/20240204200557 Feb 04 20:05:57 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm certmonger[9045]: 2024-02-04 20:05:57 [9045] Wrote to /var/lib/certmonger/requests/20240204200557 Feb 04 20:05:57 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm certmonger[9045]: 2024-02-04 20:05:57 [9045] Wrote to /var/lib/certmonger/requests/20240204200557 Feb 04 20:05:57 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm certmonger[9045]: 2024-02-04 20:05:57 [9045] Wrote to /var/lib/certmonger/requests/20240204200557 Feb 04 20:05:57 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm certmonger[9045]: 2024-02-04 20:05:57 [9045] Wrote to /var/lib/certmonger/requests/20240204200557 Feb 04 20:05:57 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm certmonger[9045]: 2024-02-04 20:05:57 [9045] Wrote to /var/lib/certmonger/requests/20240204200557 Feb 04 20:05:57 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm certmonger[9045]: 2024-02-04 20:05:57 [9045] Wrote to /var/lib/certmonger/requests/20240204200557 Feb 04 20:05:57 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm certmonger[9045]: 2024-02-04 20:05:57 [9045] Wrote to /var/lib/certmonger/requests/20240204200557 Feb 04 20:05:57 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm certmonger[9045]: 2024-02-04 20:05:57 [9045] Wrote to /var/lib/certmonger/requests/20240204200557 Feb 04 20:05:57 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm certmonger[9045]: 2024-02-04 20:05:57 [9045] Wrote to /var/lib/certmonger/requests/20240204200557 Feb 04 20:05:57 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm certmonger[9045]: 2024-02-04 20:05:57 [9045] Wrote to /var/lib/certmonger/requests/20240204200557 Feb 04 20:05:57 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm certmonger[9045]: 2024-02-04 20:05:57 [9045] Wrote to /var/lib/certmonger/requests/20240204200557 Feb 04 20:05:57 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm certmonger[9045]: 2024-02-04 20:05:57 [9045] Wrote to /var/lib/certmonger/requests/20240204200557 Feb 04 20:05:57 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm certmonger[9045]: 2024-02-04 20:05:57 [9045] Wrote to /var/lib/certmonger/requests/20240204200557 Feb 04 20:05:57 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm certmonger[9045]: 2024-02-04 20:05:57 [9045] Wrote to /var/lib/certmonger/requests/20240204200557 Feb 04 20:05:57 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm certmonger[9045]: 2024-02-04 20:05:57 [9045] Wrote to /var/lib/certmonger/requests/20240204200557 Feb 04 20:05:57 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm certmonger[9045]: 2024-02-04 20:05:57 [9045] Wrote to /var/lib/certmonger/requests/20240204200557 Feb 04 20:05:57 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm certmonger[9045]: 2024-02-04 20:05:57 [9045] Wrote to /var/lib/certmonger/requests/20240204200557 Feb 04 20:05:57 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm certmonger[9045]: 2024-02-04 20:05:57 [9045] Wrote to /var/lib/certmonger/requests/20240204200557 Feb 04 20:05:57 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm certmonger[9045]: 2024-02-04 20:05:57 [9045] Wrote to /var/lib/certmonger/requests/20240204200557 Feb 04 20:05:57 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm certmonger[9045]: 2024-02-04 20:05:57 [9045] Wrote to /var/lib/certmonger/requests/20240204200557 Feb 04 20:05:57 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm certmonger[9045]: 2024-02-04 20:05:57 [9045] Wrote to /var/lib/certmonger/requests/20240204200557 Feb 04 20:05:57 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm certmonger[9045]: 2024-02-04 20:05:57 [9045] Wrote to /var/lib/certmonger/requests/20240204200557 Feb 04 20:05:57 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm certmonger[9045]: 2024-02-04 20:05:57 [9045] Wrote to /var/lib/certmonger/requests/20240204200557 Feb 04 20:05:57 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm certmonger[9045]: 2024-02-04 20:05:57 [9045] Wrote to /var/lib/certmonger/requests/20240204200557 Feb 04 20:05:57 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm certmonger[9104]: Certificate in file "/etc/pki/tls/certs/quadlet_demo.crt" issued by CA and saved. Feb 04 20:05:57 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm certmonger[9045]: 2024-02-04 20:05:57 [9045] Wrote to /var/lib/certmonger/requests/20240204200557 Feb 04 20:05:57 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[9121]: ansible-slurp Invoked with path=/etc/pki/tls/certs/quadlet_demo.crt src=/etc/pki/tls/certs/quadlet_demo.crt Feb 04 20:05:58 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[9138]: ansible-slurp Invoked with path=/etc/pki/tls/private/quadlet_demo.key src=/etc/pki/tls/private/quadlet_demo.key Feb 04 20:05:58 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[9155]: ansible-slurp Invoked with path=/etc/pki/tls/certs/quadlet_demo.crt src=/etc/pki/tls/certs/quadlet_demo.crt Feb 04 20:05:58 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[9172]: ansible-command Invoked with _raw_params=getcert stop-tracking -f /etc/pki/tls/certs/quadlet_demo.crt warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 04 20:05:58 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm certmonger[9045]: 2024-02-04 20:05:58 [9045] Wrote to /var/lib/certmonger/requests/20240204200557 Feb 04 20:05:58 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[9190]: ansible-file Invoked with path=/etc/pki/tls/certs/quadlet_demo.crt state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Feb 04 20:05:59 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[9207]: ansible-file Invoked with path=/etc/pki/tls/private/quadlet_demo.key state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Feb 04 20:05:59 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[9224]: ansible-file Invoked with path=/etc/pki/tls/certs/quadlet_demo.crt state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Feb 04 20:05:59 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[9241]: ansible-stat Invoked with path=/run/ostree-booted follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Feb 04 20:06:01 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[9275]: ansible-dnf Invoked with name=['podman', 'shadow-utils-subid'] state=present allow_downgrade=False autoremove=False bugfix=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True lock_timeout=30 conf_file=None disable_excludes=None download_dir=None list=None releasever=None Feb 04 20:06:08 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm setsebool[9304]: The virt_use_nfs policy boolean was changed to 1 by root Feb 04 20:06:08 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm setsebool[9304]: The virt_sandbox_use_all_caps policy boolean was changed to 1 by root Feb 04 20:06:19 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm kernel: SELinux: Converting 398 SID table entries... Feb 04 20:06:19 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm kernel: SELinux: policy capability network_peer_controls=1 Feb 04 20:06:19 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm kernel: SELinux: policy capability open_perms=1 Feb 04 20:06:19 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm kernel: SELinux: policy capability extended_socket_class=1 Feb 04 20:06:19 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm kernel: SELinux: policy capability always_check_network=0 Feb 04 20:06:19 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm kernel: SELinux: policy capability cgroup_seclabel=1 Feb 04 20:06:19 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm kernel: SELinux: policy capability nnp_nosuid_transition=1 Feb 04 20:06:19 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm kernel: SELinux: policy capability genfs_seclabel_symlinks=1 Feb 04 20:06:48 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm dbus-broker-launch[578]: avc: op=load_policy lsm=selinux seqno=4 res=1 Feb 04 20:06:48 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm systemd[1]: Started /usr/bin/systemctl start man-db-cache-update. ░░ Subject: A start job for unit run-r6662d0aa72104e8c84b7c8a8a62436a0.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit run-r6662d0aa72104e8c84b7c8a8a62436a0.service has finished successfully. ░░ ░░ The job identifier is 1757. Feb 04 20:06:48 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm systemd[1]: Starting man-db-cache-update.service... ░░ Subject: A start job for unit man-db-cache-update.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit man-db-cache-update.service has begun execution. ░░ ░░ The job identifier is 1819. Feb 04 20:06:48 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm systemd[1]: Reloading. Feb 04 20:06:48 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm systemd-rc-local-generator[10022]: /etc/rc.d/rc.local is not marked executable, skipping. Feb 04 20:06:48 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm systemd[1]: Queuing reload/restart jobs for marked units… Feb 04 20:06:50 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm systemd[1]: man-db-cache-update.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit man-db-cache-update.service has successfully entered the 'dead' state. Feb 04 20:06:50 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm systemd[1]: Finished man-db-cache-update.service. ░░ Subject: A start job for unit man-db-cache-update.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit man-db-cache-update.service has finished successfully. ░░ ░░ The job identifier is 1819. Feb 04 20:06:50 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm systemd[1]: run-r6662d0aa72104e8c84b7c8a8a62436a0.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit run-r6662d0aa72104e8c84b7c8a8a62436a0.service has successfully entered the 'dead' state. Feb 04 20:06:50 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[11106]: ansible-command Invoked with _raw_params=podman --version warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 04 20:06:51 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[11129]: ansible-getent Invoked with database=passwd key=root fail_key=False service=None split=None Feb 04 20:06:51 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[11147]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None Feb 04 20:06:51 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[11165]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Feb 04 20:06:52 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[11184]: ansible-stat Invoked with path=/run/ostree-booted follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Feb 04 20:06:52 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[11201]: ansible-dnf Invoked with name=['firewalld'] state=present allow_downgrade=False autoremove=False bugfix=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True lock_timeout=30 conf_file=None disable_excludes=None download_dir=None list=None releasever=None Feb 04 20:06:53 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[11218]: ansible-systemd Invoked with name=firewalld masked=False daemon_reload=False daemon_reexec=False no_block=False state=None enabled=None force=None user=None scope=None Feb 04 20:06:54 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[11237]: ansible-systemd Invoked with name=firewalld state=started enabled=True daemon_reload=False daemon_reexec=False no_block=False force=None masked=None user=None scope=None Feb 04 20:06:54 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm systemd[1]: Reloading. Feb 04 20:06:54 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm systemd-rc-local-generator[11257]: /etc/rc.d/rc.local is not marked executable, skipping. Feb 04 20:06:54 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm systemd[1]: Starting firewalld - dynamic firewall daemon... ░░ Subject: A start job for unit firewalld.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit firewalld.service has begun execution. ░░ ░░ The job identifier is 1881. Feb 04 20:06:55 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm systemd[1]: Started firewalld - dynamic firewall daemon. ░░ Subject: A start job for unit firewalld.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit firewalld.service has finished successfully. ░░ ░░ The job identifier is 1881. Feb 04 20:06:55 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[11298]: ansible-fedora.linux_system_roles.firewall_lib Invoked with port=['8000/tcp'] permanent=True runtime=True state=enabled __report_changed=True service=[] source_port=[] forward_port=[] rich_rule=[] source=[] interface=[] interface_pci_id=[] icmp_block=[] timeout=0 ipset_entries=[] protocol=[] helper_module=[] destination=[] firewalld_conf=None masquerade=None icmp_block_inversion=None target=None zone=None set_default_zone=None ipset=None ipset_type=None description=None short=None Feb 04 20:06:56 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm kernel: Warning: Unmaintained driver is detected: ip_set Feb 04 20:06:56 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm kernel: Warning: Deprecated Driver is detected: ip_set will not be maintained in a future major release and may be disabled Feb 04 20:06:57 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[11353]: ansible-fedora.linux_system_roles.firewall_lib Invoked with port=['9000/tcp'] permanent=True runtime=True state=enabled __report_changed=True service=[] source_port=[] forward_port=[] rich_rule=[] source=[] interface=[] interface_pci_id=[] icmp_block=[] timeout=0 ipset_entries=[] protocol=[] helper_module=[] destination=[] firewalld_conf=None masquerade=None icmp_block_inversion=None target=None zone=None set_default_zone=None ipset=None ipset_type=None description=None short=None Feb 04 20:06:59 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[11439]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None Feb 04 20:06:59 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[11457]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Feb 04 20:07:00 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[11476]: ansible-file Invoked with path=/etc/containers/systemd state=directory owner=root group=0 mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Feb 04 20:07:00 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[11531]: ansible-stat Invoked with path=/etc/containers/systemd/quadlet-demo.network follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Feb 04 20:07:01 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[11581]: ansible-copy Invoked with src=/root/.ansible/tmp/ansible-tmp-1707077220.554798-10119-150119709159900/source dest=/etc/containers/systemd/quadlet-demo.network owner=root group=0 mode=0644 _original_basename=quadlet-demo.network follow=False checksum=e57c08d49aff4bae8daab138d913aeddaa8682a0 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None regexp=None delimiter=None Feb 04 20:07:01 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[11615]: ansible-systemd Invoked with daemon_reload=True scope=system daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None user=None Feb 04 20:07:01 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm systemd[1]: Reloading. Feb 04 20:07:01 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm systemd-rc-local-generator[11634]: /etc/rc.d/rc.local is not marked executable, skipping. Feb 04 20:07:02 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[11665]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None Feb 04 20:07:02 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[11683]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Feb 04 20:07:03 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[11702]: ansible-file Invoked with path=/etc/containers/systemd state=directory owner=root group=0 mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Feb 04 20:07:03 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[11757]: ansible-stat Invoked with path=/etc/containers/systemd/quadlet-demo-mysql.volume follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Feb 04 20:07:03 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[11807]: ansible-copy Invoked with src=/root/.ansible/tmp/ansible-tmp-1707077223.2156909-10203-183059669130975/source dest=/etc/containers/systemd/quadlet-demo-mysql.volume owner=root group=0 mode=0644 _original_basename=quadlet-demo-mysql.volume follow=False checksum=585f8cbdf0ec73000f9227dcffbef71e9552ea4a backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None regexp=None delimiter=None Feb 04 20:07:04 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[11841]: ansible-systemd Invoked with daemon_reload=True scope=system daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None user=None Feb 04 20:07:04 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm systemd[1]: Reloading. Feb 04 20:07:04 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm systemd-rc-local-generator[11859]: /etc/rc.d/rc.local is not marked executable, skipping. Feb 04 20:07:04 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[11891]: ansible-systemd Invoked with name=quadlet-demo-mysql-volume.service scope=system state=started daemon_reload=False daemon_reexec=False no_block=False enabled=None force=None masked=None user=None Feb 04 20:07:04 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm systemd[1]: Starting quadlet-demo-mysql-volume.service... ░░ Subject: A start job for unit quadlet-demo-mysql-volume.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit quadlet-demo-mysql-volume.service has begun execution. ░░ ░░ The job identifier is 1950. Feb 04 20:07:04 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm systemd[1]: var-lib-containers-storage-overlay-compat3945589589-merged.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit var-lib-containers-storage-overlay-compat3945589589-merged.mount has successfully entered the 'dead' state. Feb 04 20:07:04 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm systemd[1]: var-lib-containers-storage-overlay-metacopy\x2dcheck4126965686-merged.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit var-lib-containers-storage-overlay-metacopy\x2dcheck4126965686-merged.mount has successfully entered the 'dead' state. Feb 04 20:07:04 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm podman[11894]: 2024-02-04 20:07:04.653932839 +0000 UTC m=+0.096173957 system refresh Feb 04 20:07:04 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm podman[11894]: 2024-02-04 20:07:04.66061953 +0000 UTC m=+0.102860599 volume create systemd-quadlet-demo-mysql Feb 04 20:07:04 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm quadlet-demo-mysql-volume[11894]: systemd-quadlet-demo-mysql Feb 04 20:07:04 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm systemd[1]: Finished quadlet-demo-mysql-volume.service. ░░ Subject: A start job for unit quadlet-demo-mysql-volume.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit quadlet-demo-mysql-volume.service has finished successfully. ░░ ░░ The job identifier is 1950. Feb 04 20:07:05 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[11919]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None Feb 04 20:07:05 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[11937]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Feb 04 20:07:05 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Feb 04 20:07:06 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[11956]: ansible-file Invoked with path=/tmp/quadlet_demo state=directory owner=root group=root mode=0777 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Feb 04 20:07:06 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[11973]: ansible-containers.podman.podman_image Invoked with name=quay.io/linux-system-roles/mysql:5.6 force=True tag=latest pull=True push=False state=present executable=podman build={'force_rm': False, 'format': 'oci', 'cache': True, 'rm': True, 'annotation': None, 'file': None, 'volume': None, 'extra_args': None, 'target': None} push_args={'compress': None, 'format': None, 'remove_signatures': None, 'sign_by': None, 'dest': None, 'transport': None} arch=None path=None validate_certs=None auth_file=None username=None password=NOT_LOGGING_PARAMETER ca_cert_dir=None Feb 04 20:07:06 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[11973]: ansible-containers.podman.podman_image PODMAN-IMAGE-DEBUG: /usr/bin/podman image ls quay.io/linux-system-roles/mysql:5.6 --format json Feb 04 20:07:06 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[11973]: ansible-containers.podman.podman_image PODMAN-IMAGE-DEBUG: /usr/bin/podman image exists quay.io/linux-system-roles/mysql:5.6 Feb 04 20:07:06 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Feb 04 20:07:06 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[11973]: ansible-containers.podman.podman_image PODMAN-IMAGE-DEBUG: /usr/bin/podman pull quay.io/linux-system-roles/mysql:5.6 -q Feb 04 20:07:08 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm kernel: overlayfs: idmapped layers are currently not supported Feb 04 20:07:14 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm podman[11987]: 2024-02-04 20:07:14.422642938 +0000 UTC m=+7.886178984 image pull dd3b2a5dcb48ff61113592ed5ddd762581be4387c7bc552375a2159422aa6bf5 quay.io/linux-system-roles/mysql:5.6 Feb 04 20:07:14 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Feb 04 20:07:14 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[11973]: ansible-containers.podman.podman_image PODMAN-IMAGE-DEBUG: /usr/bin/podman inspect dd3b2a5dcb48ff61113592ed5ddd762581be4387c7bc552375a2159422aa6bf5 --format json Feb 04 20:07:14 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Feb 04 20:07:14 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[12088]: ansible-file Invoked with path=/etc/containers/systemd state=directory owner=root group=0 mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Feb 04 20:07:14 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[12143]: ansible-stat Invoked with path=/etc/containers/systemd/quadlet-demo-mysql.container follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Feb 04 20:07:15 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[12193]: ansible-copy Invoked with dest=/etc/containers/systemd/quadlet-demo-mysql.container owner=root group=0 mode=0644 src=/root/.ansible/tmp/ansible-tmp-1707077234.8014777-10292-186010283769801/source _original_basename=tmpoqixqlv6 follow=False checksum=7bf9312d57274d4dabbd222b01717e2cc0cfa0ed backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None regexp=None delimiter=None Feb 04 20:07:15 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[12227]: ansible-systemd Invoked with daemon_reload=True scope=system daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None user=None Feb 04 20:07:15 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm systemd[1]: Reloading. Feb 04 20:07:15 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm systemd-rc-local-generator[12245]: /etc/rc.d/rc.local is not marked executable, skipping. Feb 04 20:07:16 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[12277]: ansible-systemd Invoked with name=quadlet-demo-mysql.service scope=system state=started daemon_reload=False daemon_reexec=False no_block=False enabled=None force=None masked=None user=None Feb 04 20:07:16 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm systemd[1]: Starting quadlet-demo-network.service... ░░ Subject: A start job for unit quadlet-demo-network.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit quadlet-demo-network.service has begun execution. ░░ ░░ The job identifier is 2075. Feb 04 20:07:16 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm quadlet-demo-network[12280]: systemd-quadlet-demo Feb 04 20:07:16 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Feb 04 20:07:16 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm systemd[1]: Finished quadlet-demo-network.service. ░░ Subject: A start job for unit quadlet-demo-network.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit quadlet-demo-network.service has finished successfully. ░░ ░░ The job identifier is 2075. Feb 04 20:07:16 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm systemd[1]: Starting quadlet-demo-mysql.service... ░░ Subject: A start job for unit quadlet-demo-mysql.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit quadlet-demo-mysql.service has begun execution. ░░ ░░ The job identifier is 2013. Feb 04 20:07:16 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm quadlet-demo-mysql[12287]: Error: running container create option: no secret with name or id "mysql-root-password-container": no such secret Feb 04 20:07:16 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm podman[12287]: 2024-02-04 20:07:16.236224632 +0000 UTC m=+0.030806069 image pull dd3b2a5dcb48ff61113592ed5ddd762581be4387c7bc552375a2159422aa6bf5 quay.io/linux-system-roles/mysql:5.6 Feb 04 20:07:16 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Feb 04 20:07:16 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm systemd[1]: quadlet-demo-mysql.service: Main process exited, code=exited, status=125/n/a ░░ Subject: Unit process exited ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ An ExecStart= process belonging to unit quadlet-demo-mysql.service has exited. ░░ ░░ The process' exit code is 'exited' and its exit status is 125. Feb 04 20:07:16 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm systemd[1]: quadlet-demo-mysql.service: Failed with result 'exit-code'. ░░ Subject: Unit failed ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit quadlet-demo-mysql.service has entered the 'failed' state with result 'exit-code'. Feb 04 20:07:16 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm systemd[1]: Failed to start quadlet-demo-mysql.service. ░░ Subject: A start job for unit quadlet-demo-mysql.service has failed ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit quadlet-demo-mysql.service has finished with a failure. ░░ ░░ The job identifier is 2013 and the job result is failed. Feb 04 20:07:16 0ab5199a-dd60-4ea1-a2e9-5b4a7892c1dc.testing-farm python3[12318]: ansible-command Invoked with _raw_params=journalctl -ex warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None TASK [Check] ******************************************************************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/tests_quadlet_demo.yml:136 Sunday 04 February 2024 20:07:16 +0000 (0:00:00.250) 0:01:31.069 ******* ok: [sut] => { "changed": false, "cmd": [ "podman", "ps", "-a" ], "delta": "0:00:00.037132", "end": "2024-02-04 20:07:16.775936", "rc": 0, "start": "2024-02-04 20:07:16.738804" } STDOUT: CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES TASK [Check pods] ************************************************************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/tests_quadlet_demo.yml:140 Sunday 04 February 2024 20:07:16 +0000 (0:00:00.229) 0:01:31.299 ******* ok: [sut] => { "changed": false, "cmd": [ "podman", "pod", "ps", "--ctr-ids", "--ctr-names", "--ctr-status" ], "delta": "0:00:00.032171", "end": "2024-02-04 20:07:16.999364", "failed_when_result": false, "rc": 0, "start": "2024-02-04 20:07:16.967193" } STDOUT: POD ID NAME STATUS CREATED INFRA ID IDS NAMES STATUS TASK [Check systemd] *********************************************************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/tests_quadlet_demo.yml:145 Sunday 04 February 2024 20:07:17 +0000 (0:00:00.224) 0:01:31.523 ******* ok: [sut] => { "changed": false, "cmd": "set -euo pipefail; systemctl list-units --all --with-dependencies | grep quadlet", "delta": "0:00:00.012887", "end": "2024-02-04 20:07:17.206394", "failed_when_result": false, "rc": 0, "start": "2024-02-04 20:07:17.193507" } STDOUT: quadlet-demo-mysql-volume.service loaded active exited quadlet-demo-mysql-volume.service ● quadlet-demo-mysql.service loaded failed failed quadlet-demo-mysql.service quadlet-demo-network.service loaded active exited quadlet-demo-network.service TASK [LS] ********************************************************************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/tests_quadlet_demo.yml:153 Sunday 04 February 2024 20:07:17 +0000 (0:00:00.206) 0:01:31.730 ******* ok: [sut] => { "changed": false, "cmd": [ "ls", "-alrtF", "/etc/systemd/system" ], "delta": "0:00:00.004124", "end": "2024-02-04 20:07:17.404027", "failed_when_result": false, "rc": 0, "start": "2024-02-04 20:07:17.399903" } STDOUT: total 8 drwxr-xr-x. 2 root root 32 Jan 30 09:28 getty.target.wants/ lrwxrwxrwx. 1 root root 37 Jan 30 09:28 ctrl-alt-del.target -> /usr/lib/systemd/system/reboot.target lrwxrwxrwx. 1 root root 43 Jan 30 09:28 dbus.service -> /usr/lib/systemd/system/dbus-broker.service drwxr-xr-x. 4 root root 166 Jan 30 09:29 ../ drwxr-xr-x. 2 root root 152 Jan 30 09:29 sysinit.target.wants/ lrwxrwxrwx. 1 root root 57 Jan 30 09:29 dbus-org.freedesktop.nm-dispatcher.service -> /usr/lib/systemd/system/NetworkManager-dispatcher.service drwxr-xr-x. 2 root root 48 Jan 30 09:29 network-online.target.wants/ drwxr-xr-x. 2 root root 56 Jan 30 09:29 timers.target.wants/ drwxr-xr-x. 2 root root 38 Jan 30 09:29 dev-virtio\x2dports-org.qemu.guest_agent.0.device.wants/ drwxr-xr-x. 2 root root 31 Jan 30 09:29 basic.target.wants/ lrwxrwxrwx. 1 root root 41 Jan 30 09:31 default.target -> /usr/lib/systemd/system/multi-user.target drwxr-xr-x. 2 root root 70 Jan 30 09:48 sockets.target.wants/ drwxr-xr-x. 2 root root 31 Jan 30 09:48 remote-fs.target.wants/ drwxr-xr-x. 2 root root 59 Jan 30 09:50 sshd-keygen@.service.d/ drwxr-xr-x. 2 root root 119 Jan 30 09:50 cloud-init.target.wants/ drwxr-xr-x. 2 root root 4096 Feb 4 20:06 multi-user.target.wants/ lrwxrwxrwx. 1 root root 41 Feb 4 20:06 dbus-org.fedoraproject.FirewallD1.service -> /usr/lib/systemd/system/firewalld.service drwxr-xr-x. 13 root root 4096 Feb 4 20:06 ./ TASK [Cleanup] ***************************************************************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/tests_quadlet_demo.yml:158 Sunday 04 February 2024 20:07:17 +0000 (0:00:00.197) 0:01:31.927 ******* TASK [linux-system-roles.podman : Set platform/version specific variables] ***** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/main.yml:3 Sunday 04 February 2024 20:07:17 +0000 (0:00:00.061) 0:01:31.989 ******* included: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/set_vars.yml for sut TASK [linux-system-roles.podman : Ensure ansible_facts used by role] *********** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/set_vars.yml:3 Sunday 04 February 2024 20:07:17 +0000 (0:00:00.039) 0:01:32.028 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Check if system is ostree] ******************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/set_vars.yml:12 Sunday 04 February 2024 20:07:17 +0000 (0:00:00.028) 0:01:32.057 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Set flag to indicate system is ostree] ******* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/set_vars.yml:17 Sunday 04 February 2024 20:07:17 +0000 (0:00:00.022) 0:01:32.079 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Set platform/version specific variables] ***** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/set_vars.yml:21 Sunday 04 February 2024 20:07:17 +0000 (0:00:00.021) 0:01:32.101 ******* ok: [sut] => (item=RedHat.yml) => { "ansible_facts": { "__podman_packages": [ "podman", "shadow-utils-subid" ] }, "ansible_included_var_files": [ "/WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/vars/RedHat.yml" ], "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml" } skipping: [sut] => (item=CentOS.yml) => { "ansible_loop_var": "item", "changed": false, "item": "CentOS.yml", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=CentOS_9.yml) => { "ansible_loop_var": "item", "changed": false, "item": "CentOS_9.yml", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=CentOS_9.yml) => { "ansible_loop_var": "item", "changed": false, "item": "CentOS_9.yml", "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Gather the package facts] ******************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/main.yml:6 Sunday 04 February 2024 20:07:17 +0000 (0:00:00.055) 0:01:32.156 ******* ok: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [linux-system-roles.podman : Enable copr if requested] ******************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/main.yml:10 Sunday 04 February 2024 20:07:18 +0000 (0:00:00.948) 0:01:33.105 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Ensure required packages are installed] ****** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/main.yml:14 Sunday 04 February 2024 20:07:18 +0000 (0:00:00.023) 0:01:33.129 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Get podman version] ************************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/main.yml:23 Sunday 04 February 2024 20:07:18 +0000 (0:00:00.054) 0:01:33.184 ******* ok: [sut] => { "changed": false, "cmd": [ "podman", "--version" ], "delta": "0:00:00.030396", "end": "2024-02-04 20:07:18.885548", "rc": 0, "start": "2024-02-04 20:07:18.855152" } STDOUT: podman version 4.8.1 TASK [linux-system-roles.podman : Set podman version] ************************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/main.yml:28 Sunday 04 February 2024 20:07:18 +0000 (0:00:00.225) 0:01:33.409 ******* ok: [sut] => { "ansible_facts": { "podman_version": "4.8.1" }, "changed": false } TASK [linux-system-roles.podman : Podman package version must be 4.2 or later] *** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/main.yml:32 Sunday 04 February 2024 20:07:18 +0000 (0:00:00.023) 0:01:33.433 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Podman package version must be 4.4 or later for quadlet, secrets] *** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/main.yml:39 Sunday 04 February 2024 20:07:18 +0000 (0:00:00.022) 0:01:33.455 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } META: end_host conditional evaluated to false, continuing execution for sut TASK [linux-system-roles.podman : Check user and group information] ************ task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/main.yml:56 Sunday 04 February 2024 20:07:19 +0000 (0:00:00.042) 0:01:33.497 ******* included: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml for sut TASK [linux-system-roles.podman : Get user information] ************************ task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:2 Sunday 04 February 2024 20:07:19 +0000 (0:00:00.051) 0:01:33.549 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Fail if user does not exist] ***************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:10 Sunday 04 February 2024 20:07:19 +0000 (0:00:00.029) 0:01:33.579 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Set group for podman user] ******************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:17 Sunday 04 February 2024 20:07:19 +0000 (0:00:00.027) 0:01:33.607 ******* ok: [sut] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [linux-system-roles.podman : Get group information] *********************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:28 Sunday 04 February 2024 20:07:19 +0000 (0:00:00.031) 0:01:33.639 ******* ok: [sut] => { "ansible_facts": { "getent_group": { "root": [ "x", "0", "" ] } }, "changed": false } TASK [linux-system-roles.podman : Set group name] ****************************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:36 Sunday 04 February 2024 20:07:19 +0000 (0:00:00.208) 0:01:33.847 ******* ok: [sut] => { "ansible_facts": { "__podman_group_name": "root" }, "changed": false } TASK [linux-system-roles.podman : See if getsubids exists] ********************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:41 Sunday 04 February 2024 20:07:19 +0000 (0:00:00.029) 0:01:33.877 ******* ok: [sut] => { "changed": false, "stat": { "atime": 1707077211.5674798, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "076f89863b474504550ab1b5c11216a77ed7123c", "ctime": 1707077179.8497021, "dev": 51713, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 4380701, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-pie-executable", "mode": "0755", "mtime": 1689166618.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 15496, "uid": 0, "version": "2573075764", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [linux-system-roles.podman : Check user with getsubids] ******************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:52 Sunday 04 February 2024 20:07:19 +0000 (0:00:00.209) 0:01:34.087 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Check group with getsubids] ****************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:56 Sunday 04 February 2024 20:07:19 +0000 (0:00:00.025) 0:01:34.112 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Check if user is in subuid file] ************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:66 Sunday 04 February 2024 20:07:19 +0000 (0:00:00.023) 0:01:34.135 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Fail if user not in subuid file] ************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:74 Sunday 04 February 2024 20:07:19 +0000 (0:00:00.021) 0:01:34.157 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Check if group is in subgid file] ************ task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:81 Sunday 04 February 2024 20:07:19 +0000 (0:00:00.050) 0:01:34.208 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Fail if group not in subgid file] ************ task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:89 Sunday 04 February 2024 20:07:19 +0000 (0:00:00.023) 0:01:34.231 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Set config file paths] *********************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/main.yml:62 Sunday 04 February 2024 20:07:19 +0000 (0:00:00.024) 0:01:34.255 ******* ok: [sut] => { "ansible_facts": { "__podman_container_conf_file": "/etc/containers/containers.conf.d/50-systemroles.conf", "__podman_policy_json_file": "/etc/containers/policy.json", "__podman_registries_conf_file": "/etc/containers/registries.conf.d/50-systemroles.conf", "__podman_storage_conf_file": "/etc/containers/storage.conf" }, "changed": false } TASK [linux-system-roles.podman : Handle container.conf.d] ********************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/main.yml:80 Sunday 04 February 2024 20:07:19 +0000 (0:00:00.030) 0:01:34.286 ******* included: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_container_conf_d.yml for sut TASK [linux-system-roles.podman : Ensure containers.d exists] ****************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_container_conf_d.yml:5 Sunday 04 February 2024 20:07:19 +0000 (0:00:00.043) 0:01:34.330 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Update container config file] **************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_container_conf_d.yml:14 Sunday 04 February 2024 20:07:19 +0000 (0:00:00.022) 0:01:34.352 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Handle registries.conf.d] ******************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/main.yml:83 Sunday 04 February 2024 20:07:19 +0000 (0:00:00.022) 0:01:34.375 ******* included: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_registries_conf_d.yml for sut TASK [linux-system-roles.podman : Ensure registries.d exists] ****************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_registries_conf_d.yml:5 Sunday 04 February 2024 20:07:19 +0000 (0:00:00.043) 0:01:34.419 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Update registries config file] *************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_registries_conf_d.yml:14 Sunday 04 February 2024 20:07:19 +0000 (0:00:00.022) 0:01:34.442 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Handle storage.conf] ************************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/main.yml:86 Sunday 04 February 2024 20:07:19 +0000 (0:00:00.021) 0:01:34.464 ******* included: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_storage_conf.yml for sut TASK [linux-system-roles.podman : Ensure storage.conf parent dir exists] ******* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_storage_conf.yml:5 Sunday 04 February 2024 20:07:20 +0000 (0:00:00.043) 0:01:34.508 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Update storage config file] ****************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_storage_conf.yml:14 Sunday 04 February 2024 20:07:20 +0000 (0:00:00.022) 0:01:34.530 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Handle policy.json] ************************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/main.yml:89 Sunday 04 February 2024 20:07:20 +0000 (0:00:00.022) 0:01:34.553 ******* included: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_policy_json.yml for sut TASK [linux-system-roles.podman : Ensure policy.json parent dir exists] ******** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_policy_json.yml:6 Sunday 04 February 2024 20:07:20 +0000 (0:00:00.044) 0:01:34.597 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Stat the policy.json file] ******************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_policy_json.yml:15 Sunday 04 February 2024 20:07:20 +0000 (0:00:00.021) 0:01:34.618 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Get the existing policy.json] **************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_policy_json.yml:20 Sunday 04 February 2024 20:07:20 +0000 (0:00:00.047) 0:01:34.666 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Write new policy.json file] ****************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_policy_json.yml:26 Sunday 04 February 2024 20:07:20 +0000 (0:00:00.023) 0:01:34.690 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Manage firewall for specified ports] ************************************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/main.yml:95 Sunday 04 February 2024 20:07:20 +0000 (0:00:00.021) 0:01:34.711 ******* TASK [fedora.linux_system_roles.firewall : Setup firewalld] ******************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:2 Sunday 04 February 2024 20:07:20 +0000 (0:00:00.066) 0:01:34.777 ******* included: /WORKDIR/git-fix-linger-handling7nz_rqha/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml for sut TASK [fedora.linux_system_roles.firewall : Ensure ansible_facts used by role] *** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:2 Sunday 04 February 2024 20:07:20 +0000 (0:00:00.041) 0:01:34.818 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Check if system is ostree] ********** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:10 Sunday 04 February 2024 20:07:20 +0000 (0:00:00.027) 0:01:34.846 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Set flag to indicate system is ostree] *** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:15 Sunday 04 February 2024 20:07:20 +0000 (0:00:00.025) 0:01:34.872 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Install firewalld] ****************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:19 Sunday 04 February 2024 20:07:20 +0000 (0:00:00.022) 0:01:34.895 ******* ok: [sut] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [fedora.linux_system_roles.firewall : Collect service facts] ************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:5 Sunday 04 February 2024 20:07:21 +0000 (0:00:00.888) 0:01:35.783 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Attempt to stop and disable conflicting services] *** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:9 Sunday 04 February 2024 20:07:21 +0000 (0:00:00.022) 0:01:35.806 ******* skipping: [sut] => (item=nftables) => { "ansible_loop_var": "item", "changed": false, "item": "nftables", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=iptables) => { "ansible_loop_var": "item", "changed": false, "item": "iptables", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=ufw) => { "ansible_loop_var": "item", "changed": false, "item": "ufw", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Unmask firewalld service] *********** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:22 Sunday 04 February 2024 20:07:21 +0000 (0:00:00.029) 0:01:35.835 ******* ok: [sut] => { "changed": false, "name": "firewalld", "status": { "AccessSELinuxContext": "system_u:object_r:firewalld_unit_file_t:s0", "ActiveEnterTimestamp": "Sun 2024-02-04 20:06:55 UTC", "ActiveEnterTimestampMonotonic": "558374451", "ActiveExitTimestampMonotonic": "0", "ActiveState": "active", "After": "basic.target polkit.service dbus.socket sysinit.target system.slice dbus-broker.service", "AllowIsolate": "no", "AssertResult": "yes", "AssertTimestamp": "Sun 2024-02-04 20:06:54 UTC", "AssertTimestampMonotonic": "557246826", "Before": "multi-user.target shutdown.target network-pre.target", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "BusName": "org.fedoraproject.FirewallD1", "CPUAccounting": "yes", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "674213000", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "yes", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore", "CleanResult": "success", "CollectMode": "inactive", "ConditionResult": "yes", "ConditionTimestamp": "Sun 2024-02-04 20:06:54 UTC", "ConditionTimestampMonotonic": "557246822", "ConfigurationDirectoryMode": "0755", "Conflicts": "ipset.service nftables.service ip6tables.service shutdown.target ebtables.service iptables.service", "ControlGroup": "/system.slice/firewalld.service", "ControlGroupId": "4544", "ControlPID": "0", "CoredumpFilter": "0x33", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "Delegate": "no", "Description": "firewalld - dynamic firewall daemon", "DevicePolicy": "auto", "Documentation": "\"man:firewalld(1)\"", "DynamicUser": "no", "EffectiveCPUs": "0-1", "EffectiveMemoryNodes": "0", "EnvironmentFiles": "/etc/sysconfig/firewalld (ignore_errors=yes)", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "11274", "ExecMainStartTimestamp": "Sun 2024-02-04 20:06:54 UTC", "ExecMainStartTimestampMonotonic": "557257820", "ExecMainStatus": "0", "ExecReload": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecReloadEx": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStart": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStartEx": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExitType": "main", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FinalKillSignal": "9", "FragmentPath": "/usr/lib/systemd/system/firewalld.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOReadBytes": "18446744073709551615", "IOReadOperations": "18446744073709551615", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "IOWeight": "[not set]", "IOWriteBytes": "18446744073709551615", "IOWriteOperations": "18446744073709551615", "IPAccounting": "no", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "Id": "firewalld.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestamp": "Sun 2024-02-04 20:06:54 UTC", "InactiveExitTimestampMonotonic": "557258199", "InvocationID": "47140a4e00a3426ca0a87c8895c10488", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "mixed", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "0", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "8388608", "LimitMEMLOCKSoft": "8388608", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "524288", "LimitNOFILESoft": "1024", "LimitNPROC": "13974", "LimitNPROCSoft": "13974", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "13974", "LimitSIGPENDINGSoft": "13974", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "11274", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "ManagedOOMSwap": "auto", "MemoryAccounting": "yes", "MemoryAvailable": "infinity", "MemoryCurrent": "30572544", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemorySwapMax": "infinity", "MountAPIVFS": "no", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAPolicy": "n/a", "Names": "firewalld.service dbus-org.fedoraproject.FirewallD1.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMPolicy": "stop", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "OnSuccessJobMode": "fail", "Perpetual": "no", "PrivateDevices": "no", "PrivateIPC": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProcSubset": "all", "ProtectClock": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectHostname": "no", "ProtectKernelLogs": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectProc": "default", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "ReloadResult": "success", "ReloadSignal": "1", "RemainAfterExit": "no", "RemoveIPC": "no", "Requires": "system.slice sysinit.target dbus.socket", "Restart": "no", "RestartKillSignal": "15", "RestartUSec": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "Slice": "system.slice", "StandardError": "null", "StandardInput": "null", "StandardOutput": "null", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StateChangeTimestamp": "Sun 2024-02-04 20:07:15 UTC", "StateChangeTimestampMonotonic": "578659841", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "running", "SuccessAction": "none", "SyslogFacility": "3", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "2147483646", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "2", "TasksMax": "22359", "TimeoutAbortUSec": "1min 30s", "TimeoutCleanUSec": "infinity", "TimeoutStartFailureMode": "terminate", "TimeoutStartUSec": "1min 30s", "TimeoutStopFailureMode": "terminate", "TimeoutStopUSec": "1min 30s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "dbus", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "enabled", "UnitFileState": "enabled", "UtmpMode": "init", "WantedBy": "multi-user.target", "Wants": "network-pre.target", "WatchdogSignal": "6", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "0" } } TASK [fedora.linux_system_roles.firewall : Enable and start firewalld service] *** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:28 Sunday 04 February 2024 20:07:21 +0000 (0:00:00.368) 0:01:36.204 ******* ok: [sut] => { "changed": false, "enabled": true, "name": "firewalld", "state": "started", "status": { "AccessSELinuxContext": "system_u:object_r:firewalld_unit_file_t:s0", "ActiveEnterTimestamp": "Sun 2024-02-04 20:06:55 UTC", "ActiveEnterTimestampMonotonic": "558374451", "ActiveExitTimestampMonotonic": "0", "ActiveState": "active", "After": "basic.target polkit.service dbus.socket sysinit.target system.slice dbus-broker.service", "AllowIsolate": "no", "AssertResult": "yes", "AssertTimestamp": "Sun 2024-02-04 20:06:54 UTC", "AssertTimestampMonotonic": "557246826", "Before": "multi-user.target shutdown.target network-pre.target", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "BusName": "org.fedoraproject.FirewallD1", "CPUAccounting": "yes", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "674213000", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "yes", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore", "CleanResult": "success", "CollectMode": "inactive", "ConditionResult": "yes", "ConditionTimestamp": "Sun 2024-02-04 20:06:54 UTC", "ConditionTimestampMonotonic": "557246822", "ConfigurationDirectoryMode": "0755", "Conflicts": "ipset.service nftables.service ip6tables.service shutdown.target ebtables.service iptables.service", "ControlGroup": "/system.slice/firewalld.service", "ControlGroupId": "4544", "ControlPID": "0", "CoredumpFilter": "0x33", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "Delegate": "no", "Description": "firewalld - dynamic firewall daemon", "DevicePolicy": "auto", "Documentation": "\"man:firewalld(1)\"", "DynamicUser": "no", "EffectiveCPUs": "0-1", "EffectiveMemoryNodes": "0", "EnvironmentFiles": "/etc/sysconfig/firewalld (ignore_errors=yes)", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "11274", "ExecMainStartTimestamp": "Sun 2024-02-04 20:06:54 UTC", "ExecMainStartTimestampMonotonic": "557257820", "ExecMainStatus": "0", "ExecReload": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecReloadEx": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStart": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStartEx": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExitType": "main", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FinalKillSignal": "9", "FragmentPath": "/usr/lib/systemd/system/firewalld.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOReadBytes": "18446744073709551615", "IOReadOperations": "18446744073709551615", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "IOWeight": "[not set]", "IOWriteBytes": "18446744073709551615", "IOWriteOperations": "18446744073709551615", "IPAccounting": "no", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "Id": "firewalld.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestamp": "Sun 2024-02-04 20:06:54 UTC", "InactiveExitTimestampMonotonic": "557258199", "InvocationID": "47140a4e00a3426ca0a87c8895c10488", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "mixed", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "0", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "8388608", "LimitMEMLOCKSoft": "8388608", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "524288", "LimitNOFILESoft": "1024", "LimitNPROC": "13974", "LimitNPROCSoft": "13974", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "13974", "LimitSIGPENDINGSoft": "13974", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "11274", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "ManagedOOMSwap": "auto", "MemoryAccounting": "yes", "MemoryAvailable": "infinity", "MemoryCurrent": "30572544", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemorySwapMax": "infinity", "MountAPIVFS": "no", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAPolicy": "n/a", "Names": "firewalld.service dbus-org.fedoraproject.FirewallD1.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMPolicy": "stop", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "OnSuccessJobMode": "fail", "Perpetual": "no", "PrivateDevices": "no", "PrivateIPC": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProcSubset": "all", "ProtectClock": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectHostname": "no", "ProtectKernelLogs": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectProc": "default", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "ReloadResult": "success", "ReloadSignal": "1", "RemainAfterExit": "no", "RemoveIPC": "no", "Requires": "system.slice sysinit.target dbus.socket", "Restart": "no", "RestartKillSignal": "15", "RestartUSec": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "Slice": "system.slice", "StandardError": "null", "StandardInput": "null", "StandardOutput": "null", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StateChangeTimestamp": "Sun 2024-02-04 20:07:15 UTC", "StateChangeTimestampMonotonic": "578659841", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "running", "SuccessAction": "none", "SyslogFacility": "3", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "2147483646", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "2", "TasksMax": "22359", "TimeoutAbortUSec": "1min 30s", "TimeoutCleanUSec": "infinity", "TimeoutStartFailureMode": "terminate", "TimeoutStartUSec": "1min 30s", "TimeoutStopFailureMode": "terminate", "TimeoutStopUSec": "1min 30s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "dbus", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "enabled", "UnitFileState": "enabled", "UtmpMode": "init", "WantedBy": "multi-user.target", "Wants": "network-pre.target", "WatchdogSignal": "6", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "0" } } TASK [fedora.linux_system_roles.firewall : Check if previous replaced is defined] *** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:34 Sunday 04 February 2024 20:07:22 +0000 (0:00:00.340) 0:01:36.544 ******* ok: [sut] => { "ansible_facts": { "__firewall_previous_replaced": false, "__firewall_python_cmd": "/usr/bin/python3", "__firewall_report_changed": true }, "changed": false } TASK [fedora.linux_system_roles.firewall : Get config files, checksums before and remove] *** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:43 Sunday 04 February 2024 20:07:22 +0000 (0:00:00.028) 0:01:36.573 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Tell firewall module it is able to report changed] *** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:55 Sunday 04 February 2024 20:07:22 +0000 (0:00:00.023) 0:01:36.597 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Configure firewall] ***************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:71 Sunday 04 February 2024 20:07:22 +0000 (0:00:00.024) 0:01:36.622 ******* ok: [sut] => (item={'port': '8000/tcp', 'state': 'enabled'}) => { "__firewall_changed": false, "ansible_loop_var": "item", "changed": false, "item": { "port": "8000/tcp", "state": "enabled" } } ok: [sut] => (item={'port': '9000/tcp', 'state': 'enabled'}) => { "__firewall_changed": false, "ansible_loop_var": "item", "changed": false, "item": { "port": "9000/tcp", "state": "enabled" } } TASK [fedora.linux_system_roles.firewall : Gather firewall config information] *** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:120 Sunday 04 February 2024 20:07:22 +0000 (0:00:00.720) 0:01:37.343 ******* skipping: [sut] => (item={'port': '8000/tcp', 'state': 'enabled'}) => { "ansible_loop_var": "item", "changed": false, "item": { "port": "8000/tcp", "state": "enabled" }, "skip_reason": "Conditional result was False" } skipping: [sut] => (item={'port': '9000/tcp', 'state': 'enabled'}) => { "ansible_loop_var": "item", "changed": false, "item": { "port": "9000/tcp", "state": "enabled" }, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Update firewalld_config fact] ******* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:130 Sunday 04 February 2024 20:07:22 +0000 (0:00:00.039) 0:01:37.382 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Gather firewall config if no arguments] *** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:139 Sunday 04 February 2024 20:07:22 +0000 (0:00:00.026) 0:01:37.409 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Update firewalld_config fact] ******* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:144 Sunday 04 February 2024 20:07:22 +0000 (0:00:00.023) 0:01:37.432 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Get config files, checksums after] *** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:153 Sunday 04 February 2024 20:07:22 +0000 (0:00:00.024) 0:01:37.457 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Calculate what has changed] ********* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:163 Sunday 04 February 2024 20:07:22 +0000 (0:00:00.023) 0:01:37.480 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Show diffs] ************************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:169 Sunday 04 February 2024 20:07:23 +0000 (0:00:00.022) 0:01:37.503 ******* skipping: [sut] => {} TASK [Manage selinux for specified ports] ************************************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/main.yml:102 Sunday 04 February 2024 20:07:23 +0000 (0:00:00.022) 0:01:37.526 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Keep track of users that need to cancel linger] *** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/main.yml:109 Sunday 04 February 2024 20:07:23 +0000 (0:00:00.024) 0:01:37.550 ******* ok: [sut] => { "ansible_facts": { "__podman_cancel_user_linger": [] }, "changed": false } TASK [linux-system-roles.podman : Handle secrets] ****************************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/main.yml:113 Sunday 04 February 2024 20:07:23 +0000 (0:00:00.021) 0:01:37.572 ******* included: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_secret.yml for sut included: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_secret.yml for sut included: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_secret.yml for sut TASK [linux-system-roles.podman : Set variables part 0] ************************ task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_secret.yml:3 Sunday 04 February 2024 20:07:23 +0000 (0:00:00.095) 0:01:37.667 ******* ok: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [linux-system-roles.podman : Set variables part 1] ************************ task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_secret.yml:12 Sunday 04 February 2024 20:07:23 +0000 (0:00:00.027) 0:01:37.694 ******* ok: [sut] => { "ansible_facts": { "__podman_user": "root" }, "changed": false } TASK [linux-system-roles.podman : Set variables part 2] ************************ task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_secret.yml:17 Sunday 04 February 2024 20:07:23 +0000 (0:00:00.023) 0:01:37.718 ******* ok: [sut] => { "ansible_facts": { "__podman_rootless": false, "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [linux-system-roles.podman : Manage linger] ******************************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_secret.yml:23 Sunday 04 February 2024 20:07:23 +0000 (0:00:00.030) 0:01:37.749 ******* included: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/manage_linger.yml for sut TASK [linux-system-roles.podman : Enable linger if needed] ********************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/manage_linger.yml:12 Sunday 04 February 2024 20:07:23 +0000 (0:00:00.047) 0:01:37.796 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Mark user as not yet needing to cancel linger] *** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/manage_linger.yml:18 Sunday 04 February 2024 20:07:23 +0000 (0:00:00.025) 0:01:37.822 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Mark user for possible linger cancel] ******** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/manage_linger.yml:22 Sunday 04 February 2024 20:07:23 +0000 (0:00:00.035) 0:01:37.858 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Manage each secret] ************************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_secret.yml:28 Sunday 04 February 2024 20:07:23 +0000 (0:00:00.027) 0:01:37.886 ******* ok: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [linux-system-roles.podman : Set variables part 0] ************************ task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_secret.yml:3 Sunday 04 February 2024 20:07:23 +0000 (0:00:00.246) 0:01:38.132 ******* ok: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [linux-system-roles.podman : Set variables part 1] ************************ task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_secret.yml:12 Sunday 04 February 2024 20:07:23 +0000 (0:00:00.027) 0:01:38.160 ******* ok: [sut] => { "ansible_facts": { "__podman_user": "root" }, "changed": false } TASK [linux-system-roles.podman : Set variables part 2] ************************ task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_secret.yml:17 Sunday 04 February 2024 20:07:23 +0000 (0:00:00.027) 0:01:38.187 ******* ok: [sut] => { "ansible_facts": { "__podman_rootless": false, "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [linux-system-roles.podman : Manage linger] ******************************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_secret.yml:23 Sunday 04 February 2024 20:07:23 +0000 (0:00:00.030) 0:01:38.218 ******* included: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/manage_linger.yml for sut TASK [linux-system-roles.podman : Enable linger if needed] ********************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/manage_linger.yml:12 Sunday 04 February 2024 20:07:23 +0000 (0:00:00.041) 0:01:38.259 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Mark user as not yet needing to cancel linger] *** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/manage_linger.yml:18 Sunday 04 February 2024 20:07:23 +0000 (0:00:00.024) 0:01:38.284 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Mark user for possible linger cancel] ******** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/manage_linger.yml:22 Sunday 04 February 2024 20:07:23 +0000 (0:00:00.024) 0:01:38.309 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Manage each secret] ************************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_secret.yml:28 Sunday 04 February 2024 20:07:23 +0000 (0:00:00.069) 0:01:38.378 ******* ok: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [linux-system-roles.podman : Set variables part 0] ************************ task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_secret.yml:3 Sunday 04 February 2024 20:07:24 +0000 (0:00:00.249) 0:01:38.627 ******* ok: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [linux-system-roles.podman : Set variables part 1] ************************ task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_secret.yml:12 Sunday 04 February 2024 20:07:24 +0000 (0:00:00.030) 0:01:38.657 ******* ok: [sut] => { "ansible_facts": { "__podman_user": "root" }, "changed": false } TASK [linux-system-roles.podman : Set variables part 2] ************************ task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_secret.yml:17 Sunday 04 February 2024 20:07:24 +0000 (0:00:00.038) 0:01:38.696 ******* ok: [sut] => { "ansible_facts": { "__podman_rootless": false, "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [linux-system-roles.podman : Manage linger] ******************************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_secret.yml:23 Sunday 04 February 2024 20:07:24 +0000 (0:00:00.037) 0:01:38.733 ******* included: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/manage_linger.yml for sut TASK [linux-system-roles.podman : Enable linger if needed] ********************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/manage_linger.yml:12 Sunday 04 February 2024 20:07:24 +0000 (0:00:00.042) 0:01:38.776 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Mark user as not yet needing to cancel linger] *** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/manage_linger.yml:18 Sunday 04 February 2024 20:07:24 +0000 (0:00:00.025) 0:01:38.802 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Mark user for possible linger cancel] ******** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/manage_linger.yml:22 Sunday 04 February 2024 20:07:24 +0000 (0:00:00.025) 0:01:38.828 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Manage each secret] ************************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_secret.yml:28 Sunday 04 February 2024 20:07:24 +0000 (0:00:00.023) 0:01:38.851 ******* ok: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [linux-system-roles.podman : Handle Kubernetes specifications] ************ task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/main.yml:120 Sunday 04 February 2024 20:07:24 +0000 (0:00:00.243) 0:01:39.095 ******* TASK [linux-system-roles.podman : Handle Quadlet specifications] *************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/main.yml:126 Sunday 04 February 2024 20:07:24 +0000 (0:00:00.021) 0:01:39.116 ******* included: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml for sut included: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml for sut included: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml for sut included: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml for sut included: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml for sut included: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml for sut TASK [linux-system-roles.podman : Set per-container variables part 0] ********** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:14 Sunday 04 February 2024 20:07:24 +0000 (0:00:00.096) 0:01:39.213 ******* ok: [sut] => { "ansible_facts": { "__podman_quadlet_file_src": "quadlet-demo-mysql.volume", "__podman_quadlet_spec": {}, "__podman_quadlet_str": "[Volume]", "__podman_quadlet_template_src": "" }, "changed": false } TASK [linux-system-roles.podman : Set per-container variables part 1] ********** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:40 Sunday 04 February 2024 20:07:24 +0000 (0:00:00.033) 0:01:39.246 ******* ok: [sut] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_pull_image": true, "__podman_state": "absent", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [linux-system-roles.podman : Fail if no quadlet spec is given] ************ task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:57 Sunday 04 February 2024 20:07:24 +0000 (0:00:00.028) 0:01:39.275 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Set per-container variables part 2] ********** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:70 Sunday 04 February 2024 20:07:24 +0000 (0:00:00.056) 0:01:39.332 ******* ok: [sut] => { "ansible_facts": { "__podman_quadlet_name": "quadlet-demo-mysql", "__podman_quadlet_type": "volume", "__podman_rootless": false }, "changed": false } TASK [linux-system-roles.podman : Check user and group information] ************ task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:97 Sunday 04 February 2024 20:07:24 +0000 (0:00:00.034) 0:01:39.366 ******* included: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml for sut TASK [linux-system-roles.podman : Get user information] ************************ task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:2 Sunday 04 February 2024 20:07:24 +0000 (0:00:00.046) 0:01:39.413 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Fail if user does not exist] ***************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:10 Sunday 04 February 2024 20:07:24 +0000 (0:00:00.030) 0:01:39.444 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Set group for podman user] ******************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:17 Sunday 04 February 2024 20:07:24 +0000 (0:00:00.031) 0:01:39.475 ******* ok: [sut] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [linux-system-roles.podman : Get group information] *********************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:28 Sunday 04 February 2024 20:07:25 +0000 (0:00:00.032) 0:01:39.508 ******* ok: [sut] => { "ansible_facts": { "getent_group": { "root": [ "x", "0", "" ] } }, "changed": false } TASK [linux-system-roles.podman : Set group name] ****************************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:36 Sunday 04 February 2024 20:07:25 +0000 (0:00:00.205) 0:01:39.714 ******* ok: [sut] => { "ansible_facts": { "__podman_group_name": "root" }, "changed": false } TASK [linux-system-roles.podman : See if getsubids exists] ********************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:41 Sunday 04 February 2024 20:07:25 +0000 (0:00:00.030) 0:01:39.745 ******* ok: [sut] => { "changed": false, "stat": { "atime": 1707077211.5674798, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "076f89863b474504550ab1b5c11216a77ed7123c", "ctime": 1707077179.8497021, "dev": 51713, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 4380701, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-pie-executable", "mode": "0755", "mtime": 1689166618.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 15496, "uid": 0, "version": "2573075764", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [linux-system-roles.podman : Check user with getsubids] ******************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:52 Sunday 04 February 2024 20:07:25 +0000 (0:00:00.196) 0:01:39.942 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Check group with getsubids] ****************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:56 Sunday 04 February 2024 20:07:25 +0000 (0:00:00.023) 0:01:39.965 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Check if user is in subuid file] ************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:66 Sunday 04 February 2024 20:07:25 +0000 (0:00:00.022) 0:01:39.988 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Fail if user not in subuid file] ************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:74 Sunday 04 February 2024 20:07:25 +0000 (0:00:00.022) 0:01:40.011 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Check if group is in subgid file] ************ task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:81 Sunday 04 February 2024 20:07:25 +0000 (0:00:00.021) 0:01:40.033 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Fail if group not in subgid file] ************ task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:89 Sunday 04 February 2024 20:07:25 +0000 (0:00:00.023) 0:01:40.056 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Set per-container variables part 3] ********** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:102 Sunday 04 February 2024 20:07:25 +0000 (0:00:00.021) 0:01:40.078 ******* ok: [sut] => { "ansible_facts": { "__podman_activate_systemd_unit": true, "__podman_images_found": [], "__podman_kube_yamls_raw": "", "__podman_service_name": "quadlet-demo-mysql-volume.service", "__podman_systemd_scope": "system", "__podman_user_home_dir": "/root", "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [linux-system-roles.podman : Set per-container variables part 4] ********** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:134 Sunday 04 February 2024 20:07:25 +0000 (0:00:00.042) 0:01:40.120 ******* ok: [sut] => { "ansible_facts": { "__podman_quadlet_path": "/etc/containers/systemd" }, "changed": false } TASK [linux-system-roles.podman : Get kube yaml contents] ********************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:140 Sunday 04 February 2024 20:07:25 +0000 (0:00:00.026) 0:01:40.147 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Set per-container variables part 5] ********** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:151 Sunday 04 February 2024 20:07:25 +0000 (0:00:00.050) 0:01:40.198 ******* ok: [sut] => { "ansible_facts": { "__podman_images": [], "__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo-mysql.volume", "__podman_volumes": [] }, "changed": false } TASK [linux-system-roles.podman : Cleanup quadlets] **************************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:205 Sunday 04 February 2024 20:07:25 +0000 (0:00:00.059) 0:01:40.257 ******* included: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/cleanup_quadlet_spec.yml for sut TASK [linux-system-roles.podman : Stop and disable service] ******************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/cleanup_quadlet_spec.yml:4 Sunday 04 February 2024 20:07:25 +0000 (0:00:00.047) 0:01:40.305 ******* changed: [sut] => { "changed": true, "enabled": false, "failed_when_result": false, "name": "quadlet-demo-mysql-volume.service", "state": "stopped", "status": { "AccessSELinuxContext": "system_u:object_r:systemd_unit_file_t:s0", "ActiveEnterTimestamp": "Sun 2024-02-04 20:07:04 UTC", "ActiveEnterTimestampMonotonic": "567528050", "ActiveExitTimestampMonotonic": "0", "ActiveState": "active", "After": "basic.target system.slice sysinit.target -.mount systemd-journald.socket", "AllowIsolate": "no", "AssertResult": "yes", "AssertTimestamp": "Sun 2024-02-04 20:07:04 UTC", "AssertTimestampMonotonic": "567407254", "Before": "shutdown.target quadlet-demo-mysql.service", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "CPUAccounting": "yes", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "68359000", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "no", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore", "CleanResult": "success", "CollectMode": "inactive", "ConditionResult": "yes", "ConditionTimestamp": "Sun 2024-02-04 20:07:04 UTC", "ConditionTimestampMonotonic": "567407250", "ConfigurationDirectoryMode": "0755", "Conflicts": "shutdown.target", "ControlGroupId": "0", "ControlPID": "0", "CoredumpFilter": "0x33", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "Delegate": "no", "Description": "quadlet-demo-mysql-volume.service", "DevicePolicy": "auto", "DynamicUser": "no", "ExecMainCode": "1", "ExecMainExitTimestamp": "Sun 2024-02-04 20:07:04 UTC", "ExecMainExitTimestampMonotonic": "567527917", "ExecMainPID": "11894", "ExecMainStartTimestamp": "Sun 2024-02-04 20:07:04 UTC", "ExecMainStartTimestampMonotonic": "567413512", "ExecMainStatus": "0", "ExecStart": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman volume create --ignore systemd-quadlet-demo-mysql ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStartEx": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman volume create --ignore systemd-quadlet-demo-mysql ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExitType": "main", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FinalKillSignal": "9", "FragmentPath": "/run/systemd/generator/quadlet-demo-mysql-volume.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOReadBytes": "18446744073709551615", "IOReadOperations": "18446744073709551615", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "IOWeight": "[not set]", "IOWriteBytes": "18446744073709551615", "IOWriteOperations": "18446744073709551615", "IPAccounting": "no", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "Id": "quadlet-demo-mysql-volume.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestamp": "Sun 2024-02-04 20:07:04 UTC", "InactiveExitTimestampMonotonic": "567413858", "InvocationID": "7a7ac339768040f7842968c2c83bea0f", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "control-group", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "0", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "8388608", "LimitMEMLOCKSoft": "8388608", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "524288", "LimitNOFILESoft": "1024", "LimitNPROC": "13974", "LimitNPROCSoft": "13974", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "13974", "LimitSIGPENDINGSoft": "13974", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "0", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "ManagedOOMSwap": "auto", "MemoryAccounting": "yes", "MemoryAvailable": "infinity", "MemoryCurrent": "[not set]", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemorySwapMax": "infinity", "MountAPIVFS": "no", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAPolicy": "n/a", "Names": "quadlet-demo-mysql-volume.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMPolicy": "stop", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "OnSuccessJobMode": "fail", "Perpetual": "no", "PrivateDevices": "no", "PrivateIPC": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProcSubset": "all", "ProtectClock": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectHostname": "no", "ProtectKernelLogs": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectProc": "default", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "ReloadResult": "success", "ReloadSignal": "1", "RemainAfterExit": "yes", "RemoveIPC": "no", "RequiredBy": "quadlet-demo-mysql.service", "Requires": "system.slice -.mount sysinit.target", "RequiresMountsFor": "/run/containers", "Restart": "no", "RestartKillSignal": "15", "RestartUSec": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "Slice": "system.slice", "StandardError": "inherit", "StandardInput": "null", "StandardOutput": "journal", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StateChangeTimestamp": "Sun 2024-02-04 20:07:04 UTC", "StateChangeTimestampMonotonic": "567528050", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "exited", "SuccessAction": "none", "SyslogFacility": "3", "SyslogIdentifier": "quadlet-demo-mysql-volume", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "2147483646", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "[not set]", "TasksMax": "22359", "TimeoutAbortUSec": "1min 30s", "TimeoutCleanUSec": "infinity", "TimeoutStartFailureMode": "terminate", "TimeoutStartUSec": "infinity", "TimeoutStopFailureMode": "terminate", "TimeoutStopUSec": "1min 30s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "oneshot", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "disabled", "UnitFileState": "generated", "UtmpMode": "init", "WatchdogSignal": "6", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "0" } } TASK [linux-system-roles.podman : Remove quadlet file] ************************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/cleanup_quadlet_spec.yml:22 Sunday 04 February 2024 20:07:26 +0000 (0:00:00.551) 0:01:40.856 ******* changed: [sut] => { "changed": true, "path": "/etc/containers/systemd/quadlet-demo-mysql.volume", "state": "absent" } TASK [linux-system-roles.podman : Manage linger] ******************************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/cleanup_quadlet_spec.yml:28 Sunday 04 February 2024 20:07:26 +0000 (0:00:00.198) 0:01:41.055 ******* included: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/manage_linger.yml for sut TASK [linux-system-roles.podman : Enable linger if needed] ********************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/manage_linger.yml:12 Sunday 04 February 2024 20:07:26 +0000 (0:00:00.039) 0:01:41.094 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Mark user as not yet needing to cancel linger] *** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/manage_linger.yml:18 Sunday 04 February 2024 20:07:26 +0000 (0:00:00.024) 0:01:41.118 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Mark user for possible linger cancel] ******** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/manage_linger.yml:22 Sunday 04 February 2024 20:07:26 +0000 (0:00:00.023) 0:01:41.142 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Reload systemctl] **************************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/cleanup_quadlet_spec.yml:36 Sunday 04 February 2024 20:07:26 +0000 (0:00:00.023) 0:01:41.166 ******* ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.podman : Prune images no longer in use] *************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/cleanup_quadlet_spec.yml:45 Sunday 04 February 2024 20:07:27 +0000 (0:00:00.508) 0:01:41.674 ******* changed: [sut] => { "changed": true, "cmd": [ "podman", "image", "prune", "-f" ], "delta": "0:00:00.040979", "end": "2024-02-04 20:07:27.391514", "rc": 0, "start": "2024-02-04 20:07:27.350535" } TASK [linux-system-roles.podman : Remove network] ****************************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/cleanup_quadlet_spec.yml:53 Sunday 04 February 2024 20:07:27 +0000 (0:00:00.273) 0:01:41.947 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Create and update quadlets] ****************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:209 Sunday 04 February 2024 20:07:27 +0000 (0:00:00.027) 0:01:41.975 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Set per-container variables part 0] ********** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:14 Sunday 04 February 2024 20:07:27 +0000 (0:00:00.028) 0:01:42.003 ******* ok: [sut] => { "ansible_facts": { "__podman_quadlet_file_src": "", "__podman_quadlet_spec": {}, "__podman_quadlet_str": "[Install]\nWantedBy=default.target\n\n[Container]\nImage=quay.io/linux-system-roles/mysql:5.6\nContainerName=quadlet-demo-mysql\nVolume=quadlet-demo-mysql.volume:/var/lib/mysql\nVolume=/tmp/quadlet_demo:/var/lib/quadlet_demo:Z\nNetwork=quadlet-demo.network\nSecret=mysql-root-password-container,type=env,target=MYSQL_ROOT_PASSWORD\nHealthCmd=/usr/bin/true\nHealthOnFailure=kill\n", "__podman_quadlet_template_src": "quadlet-demo-mysql.container.j2" }, "changed": false } TASK [linux-system-roles.podman : Set per-container variables part 1] ********** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:40 Sunday 04 February 2024 20:07:27 +0000 (0:00:00.080) 0:01:42.083 ******* ok: [sut] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_pull_image": true, "__podman_state": "absent", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [linux-system-roles.podman : Fail if no quadlet spec is given] ************ task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:57 Sunday 04 February 2024 20:07:27 +0000 (0:00:00.030) 0:01:42.114 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Set per-container variables part 2] ********** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:70 Sunday 04 February 2024 20:07:27 +0000 (0:00:00.024) 0:01:42.138 ******* ok: [sut] => { "ansible_facts": { "__podman_quadlet_name": "quadlet-demo-mysql", "__podman_quadlet_type": "container", "__podman_rootless": false }, "changed": false } TASK [linux-system-roles.podman : Check user and group information] ************ task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:97 Sunday 04 February 2024 20:07:27 +0000 (0:00:00.033) 0:01:42.172 ******* included: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml for sut TASK [linux-system-roles.podman : Get user information] ************************ task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:2 Sunday 04 February 2024 20:07:27 +0000 (0:00:00.045) 0:01:42.217 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Fail if user does not exist] ***************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:10 Sunday 04 February 2024 20:07:27 +0000 (0:00:00.029) 0:01:42.247 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Set group for podman user] ******************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:17 Sunday 04 February 2024 20:07:27 +0000 (0:00:00.029) 0:01:42.276 ******* ok: [sut] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [linux-system-roles.podman : Get group information] *********************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:28 Sunday 04 February 2024 20:07:27 +0000 (0:00:00.032) 0:01:42.309 ******* ok: [sut] => { "ansible_facts": { "getent_group": { "root": [ "x", "0", "" ] } }, "changed": false } TASK [linux-system-roles.podman : Set group name] ****************************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:36 Sunday 04 February 2024 20:07:28 +0000 (0:00:00.204) 0:01:42.514 ******* ok: [sut] => { "ansible_facts": { "__podman_group_name": "root" }, "changed": false } TASK [linux-system-roles.podman : See if getsubids exists] ********************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:41 Sunday 04 February 2024 20:07:28 +0000 (0:00:00.030) 0:01:42.545 ******* ok: [sut] => { "changed": false, "stat": { "atime": 1707077211.5674798, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "076f89863b474504550ab1b5c11216a77ed7123c", "ctime": 1707077179.8497021, "dev": 51713, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 4380701, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-pie-executable", "mode": "0755", "mtime": 1689166618.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 15496, "uid": 0, "version": "2573075764", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [linux-system-roles.podman : Check user with getsubids] ******************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:52 Sunday 04 February 2024 20:07:28 +0000 (0:00:00.200) 0:01:42.745 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Check group with getsubids] ****************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:56 Sunday 04 February 2024 20:07:28 +0000 (0:00:00.023) 0:01:42.768 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Check if user is in subuid file] ************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:66 Sunday 04 February 2024 20:07:28 +0000 (0:00:00.025) 0:01:42.794 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Fail if user not in subuid file] ************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:74 Sunday 04 February 2024 20:07:28 +0000 (0:00:00.022) 0:01:42.817 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Check if group is in subgid file] ************ task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:81 Sunday 04 February 2024 20:07:28 +0000 (0:00:00.053) 0:01:42.870 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Fail if group not in subgid file] ************ task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:89 Sunday 04 February 2024 20:07:28 +0000 (0:00:00.025) 0:01:42.896 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Set per-container variables part 3] ********** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:102 Sunday 04 February 2024 20:07:28 +0000 (0:00:00.023) 0:01:42.919 ******* ok: [sut] => { "ansible_facts": { "__podman_activate_systemd_unit": true, "__podman_images_found": [ "quay.io/linux-system-roles/mysql:5.6" ], "__podman_kube_yamls_raw": "", "__podman_service_name": "quadlet-demo-mysql.service", "__podman_systemd_scope": "system", "__podman_user_home_dir": "/root", "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [linux-system-roles.podman : Set per-container variables part 4] ********** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:134 Sunday 04 February 2024 20:07:28 +0000 (0:00:00.041) 0:01:42.961 ******* ok: [sut] => { "ansible_facts": { "__podman_quadlet_path": "/etc/containers/systemd" }, "changed": false } TASK [linux-system-roles.podman : Get kube yaml contents] ********************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:140 Sunday 04 February 2024 20:07:28 +0000 (0:00:00.027) 0:01:42.989 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Set per-container variables part 5] ********** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:151 Sunday 04 February 2024 20:07:28 +0000 (0:00:00.025) 0:01:43.014 ******* ok: [sut] => { "ansible_facts": { "__podman_images": [ "quay.io/linux-system-roles/mysql:5.6" ], "__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo-mysql.container", "__podman_volumes": [ "/tmp/quadlet_demo" ] }, "changed": false } TASK [linux-system-roles.podman : Cleanup quadlets] **************************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:205 Sunday 04 February 2024 20:07:28 +0000 (0:00:00.060) 0:01:43.075 ******* included: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/cleanup_quadlet_spec.yml for sut TASK [linux-system-roles.podman : Stop and disable service] ******************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/cleanup_quadlet_spec.yml:4 Sunday 04 February 2024 20:07:28 +0000 (0:00:00.049) 0:01:43.124 ******* changed: [sut] => { "changed": true, "enabled": false, "failed_when_result": false, "name": "quadlet-demo-mysql.service", "state": "stopped", "status": { "AccessSELinuxContext": "system_u:object_r:systemd_unit_file_t:s0", "ActiveEnterTimestampMonotonic": "0", "ActiveExitTimestampMonotonic": "0", "ActiveState": "failed", "After": "sysinit.target system.slice tmp.mount -.mount quadlet-demo-mysql-volume.service basic.target systemd-journald.socket quadlet-demo-network.service", "AllowIsolate": "no", "AssertResult": "yes", "AssertTimestamp": "Sun 2024-02-04 20:07:16 UTC", "AssertTimestampMonotonic": "579054785", "Before": "multi-user.target shutdown.target", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "CPUAccounting": "yes", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "89320000", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "no", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore", "CleanResult": "success", "CollectMode": "inactive", "ConditionResult": "yes", "ConditionTimestamp": "Sun 2024-02-04 20:07:16 UTC", "ConditionTimestampMonotonic": "579054781", "ConfigurationDirectoryMode": "0755", "Conflicts": "shutdown.target", "ControlGroupId": "0", "ControlPID": "0", "CoredumpFilter": "0x33", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "Delegate": "yes", "DelegateControllers": "cpu cpuacct cpuset io blkio memory devices pids bpf-firewall bpf-devices bpf-foreign bpf-socket-bind bpf-restrict-network-interfaces", "Description": "quadlet-demo-mysql.service", "DevicePolicy": "auto", "DynamicUser": "no", "Environment": "PODMAN_SYSTEMD_UNIT=quadlet-demo-mysql.service", "ExecMainCode": "1", "ExecMainExitTimestamp": "Sun 2024-02-04 20:07:16 UTC", "ExecMainExitTimestampMonotonic": "579112864", "ExecMainPID": "12287", "ExecMainStartTimestamp": "Sun 2024-02-04 20:07:16 UTC", "ExecMainStartTimestampMonotonic": "579062545", "ExecMainStatus": "125", "ExecStart": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman run --name=quadlet-demo-mysql --cidfile=/run/quadlet-demo-mysql.cid --replace --rm --cgroups=split --network=systemd-quadlet-demo --sdnotify=conmon -d -v systemd-quadlet-demo-mysql:/var/lib/mysql -v /tmp/quadlet_demo:/var/lib/quadlet_demo:Z --secret mysql-root-password-container,type=env,target=MYSQL_ROOT_PASSWORD --health-cmd /usr/bin/true --health-on-failure kill quay.io/linux-system-roles/mysql:5.6 ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStartEx": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman run --name=quadlet-demo-mysql --cidfile=/run/quadlet-demo-mysql.cid --replace --rm --cgroups=split --network=systemd-quadlet-demo --sdnotify=conmon -d -v systemd-quadlet-demo-mysql:/var/lib/mysql -v /tmp/quadlet_demo:/var/lib/quadlet_demo:Z --secret mysql-root-password-container,type=env,target=MYSQL_ROOT_PASSWORD --health-cmd /usr/bin/true --health-on-failure kill quay.io/linux-system-roles/mysql:5.6 ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStop": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman rm -v -f -i --cidfile=/run/quadlet-demo-mysql.cid ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStopEx": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman rm -v -f -i --cidfile=/run/quadlet-demo-mysql.cid ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStopPost": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman rm -v -f -i --cidfile=/run/quadlet-demo-mysql.cid ; ignore_errors=yes ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStopPostEx": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman rm -v -f -i --cidfile=/run/quadlet-demo-mysql.cid ; flags=ignore-failure ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExitType": "main", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FinalKillSignal": "9", "FragmentPath": "/run/systemd/generator/quadlet-demo-mysql.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOReadBytes": "18446744073709551615", "IOReadOperations": "18446744073709551615", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "IOWeight": "[not set]", "IOWriteBytes": "18446744073709551615", "IOWriteOperations": "18446744073709551615", "IPAccounting": "no", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "Id": "quadlet-demo-mysql.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestamp": "Sun 2024-02-04 20:07:16 UTC", "InactiveEnterTimestampMonotonic": "579154292", "InactiveExitTimestamp": "Sun 2024-02-04 20:07:16 UTC", "InactiveExitTimestampMonotonic": "579062953", "InvocationID": "c4411785157e4ac6b2ab9031198bdcfa", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "mixed", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "0", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "8388608", "LimitMEMLOCKSoft": "8388608", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "524288", "LimitNOFILESoft": "1024", "LimitNPROC": "13974", "LimitNPROCSoft": "13974", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "13974", "LimitSIGPENDINGSoft": "13974", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "0", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "ManagedOOMSwap": "auto", "MemoryAccounting": "yes", "MemoryAvailable": "infinity", "MemoryCurrent": "[not set]", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemorySwapMax": "infinity", "MountAPIVFS": "no", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAPolicy": "n/a", "Names": "quadlet-demo-mysql.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "all", "OOMPolicy": "continue", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "OnSuccessJobMode": "fail", "Perpetual": "no", "PrivateDevices": "no", "PrivateIPC": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProcSubset": "all", "ProtectClock": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectHostname": "no", "ProtectKernelLogs": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectProc": "default", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "ReloadResult": "success", "ReloadSignal": "1", "RemainAfterExit": "no", "RemoveIPC": "no", "Requires": "quadlet-demo-mysql-volume.service quadlet-demo-network.service sysinit.target system.slice -.mount", "RequiresMountsFor": "/tmp/quadlet_demo /run/containers", "Restart": "no", "RestartKillSignal": "15", "RestartUSec": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "exit-code", "RootDirectoryStartOnly": "no", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "Slice": "system.slice", "SourcePath": "/etc/containers/systemd/quadlet-demo-mysql.container", "StandardError": "inherit", "StandardInput": "null", "StandardOutput": "journal", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StateChangeTimestamp": "Sun 2024-02-04 20:07:16 UTC", "StateChangeTimestampMonotonic": "579154292", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "failed", "SuccessAction": "none", "SyslogFacility": "3", "SyslogIdentifier": "quadlet-demo-mysql", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "2147483646", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "[not set]", "TasksMax": "22359", "TimeoutAbortUSec": "1min 30s", "TimeoutCleanUSec": "infinity", "TimeoutStartFailureMode": "terminate", "TimeoutStartUSec": "1min 30s", "TimeoutStopFailureMode": "terminate", "TimeoutStopUSec": "1min 30s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "notify", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "disabled", "UnitFileState": "generated", "UtmpMode": "init", "WantedBy": "multi-user.target", "WatchdogSignal": "6", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "0" } } TASK [linux-system-roles.podman : Remove quadlet file] ************************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/cleanup_quadlet_spec.yml:22 Sunday 04 February 2024 20:07:29 +0000 (0:00:00.535) 0:01:43.659 ******* changed: [sut] => { "changed": true, "path": "/etc/containers/systemd/quadlet-demo-mysql.container", "state": "absent" } TASK [linux-system-roles.podman : Manage linger] ******************************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/cleanup_quadlet_spec.yml:28 Sunday 04 February 2024 20:07:29 +0000 (0:00:00.199) 0:01:43.859 ******* included: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/manage_linger.yml for sut TASK [linux-system-roles.podman : Enable linger if needed] ********************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/manage_linger.yml:12 Sunday 04 February 2024 20:07:29 +0000 (0:00:00.040) 0:01:43.899 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Mark user as not yet needing to cancel linger] *** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/manage_linger.yml:18 Sunday 04 February 2024 20:07:29 +0000 (0:00:00.024) 0:01:43.923 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Mark user for possible linger cancel] ******** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/manage_linger.yml:22 Sunday 04 February 2024 20:07:29 +0000 (0:00:00.024) 0:01:43.948 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Reload systemctl] **************************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/cleanup_quadlet_spec.yml:36 Sunday 04 February 2024 20:07:29 +0000 (0:00:00.023) 0:01:43.972 ******* ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.podman : Prune images no longer in use] *************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/cleanup_quadlet_spec.yml:45 Sunday 04 February 2024 20:07:30 +0000 (0:00:00.576) 0:01:44.548 ******* changed: [sut] => { "changed": true, "cmd": [ "podman", "image", "prune", "-f" ], "delta": "0:00:00.034632", "end": "2024-02-04 20:07:30.259273", "rc": 0, "start": "2024-02-04 20:07:30.224641" } TASK [linux-system-roles.podman : Remove network] ****************************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/cleanup_quadlet_spec.yml:53 Sunday 04 February 2024 20:07:30 +0000 (0:00:00.235) 0:01:44.783 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Create and update quadlets] ****************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:209 Sunday 04 February 2024 20:07:30 +0000 (0:00:00.026) 0:01:44.810 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Set per-container variables part 0] ********** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:14 Sunday 04 February 2024 20:07:30 +0000 (0:00:00.023) 0:01:44.833 ******* ok: [sut] => { "ansible_facts": { "__podman_quadlet_file_src": "envoy-proxy-configmap.yml", "__podman_quadlet_spec": {}, "__podman_quadlet_str": "---\napiVersion: v1\nkind: ConfigMap\nmetadata:\n name: envoy-proxy-config\ndata:\n envoy.yaml: |\n admin:\n address:\n socket_address:\n address: 0.0.0.0\n port_value: 9901\n\n static_resources:\n listeners:\n - name: listener_0\n address:\n socket_address:\n address: 0.0.0.0\n port_value: 8080\n filter_chains:\n - filters:\n - name: envoy.filters.network.http_connection_manager\n typed_config:\n \"@type\": type.googleapis.com/envoy.extensions.filters.network.http_connection_manager.v3.HttpConnectionManager\n stat_prefix: ingress_http\n codec_type: AUTO\n route_config:\n name: local_route\n virtual_hosts:\n - name: local_service\n domains: [\"*\"]\n routes:\n - match:\n prefix: \"/\"\n route:\n cluster: backend\n http_filters:\n - name: envoy.filters.http.router\n typed_config:\n \"@type\": type.googleapis.com/envoy.extensions.filters.http.router.v3.Router\n transport_socket:\n name: envoy.transport_sockets.tls\n typed_config:\n \"@type\": type.googleapis.com/envoy.extensions.transport_sockets.tls.v3.DownstreamTlsContext\n common_tls_context:\n tls_certificates:\n - certificate_chain:\n filename: /etc/envoy-certificates/certificate.pem\n private_key:\n filename: /etc/envoy-certificates/certificate.key\n clusters:\n - name: backend\n connect_timeout: 5s\n type: STATIC\n dns_refresh_rate: 1800s\n lb_policy: ROUND_ROBIN\n load_assignment:\n cluster_name: backend\n endpoints:\n - lb_endpoints:\n - endpoint:\n address:\n socket_address:\n address: 127.0.0.1\n port_value: 80", "__podman_quadlet_template_src": "" }, "changed": false } TASK [linux-system-roles.podman : Set per-container variables part 1] ********** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:40 Sunday 04 February 2024 20:07:30 +0000 (0:00:00.034) 0:01:44.868 ******* ok: [sut] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_pull_image": true, "__podman_state": "absent", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [linux-system-roles.podman : Fail if no quadlet spec is given] ************ task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:57 Sunday 04 February 2024 20:07:30 +0000 (0:00:00.030) 0:01:44.899 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Set per-container variables part 2] ********** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:70 Sunday 04 February 2024 20:07:30 +0000 (0:00:00.025) 0:01:44.924 ******* ok: [sut] => { "ansible_facts": { "__podman_quadlet_name": "envoy-proxy-configmap", "__podman_quadlet_type": "yml", "__podman_rootless": false }, "changed": false } TASK [linux-system-roles.podman : Check user and group information] ************ task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:97 Sunday 04 February 2024 20:07:30 +0000 (0:00:00.035) 0:01:44.960 ******* included: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml for sut TASK [linux-system-roles.podman : Get user information] ************************ task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:2 Sunday 04 February 2024 20:07:30 +0000 (0:00:00.044) 0:01:45.004 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Fail if user does not exist] ***************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:10 Sunday 04 February 2024 20:07:30 +0000 (0:00:00.034) 0:01:45.039 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Set group for podman user] ******************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:17 Sunday 04 February 2024 20:07:30 +0000 (0:00:00.029) 0:01:45.068 ******* ok: [sut] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [linux-system-roles.podman : Get group information] *********************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:28 Sunday 04 February 2024 20:07:30 +0000 (0:00:00.031) 0:01:45.099 ******* ok: [sut] => { "ansible_facts": { "getent_group": { "root": [ "x", "0", "" ] } }, "changed": false } TASK [linux-system-roles.podman : Set group name] ****************************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:36 Sunday 04 February 2024 20:07:30 +0000 (0:00:00.204) 0:01:45.304 ******* ok: [sut] => { "ansible_facts": { "__podman_group_name": "root" }, "changed": false } TASK [linux-system-roles.podman : See if getsubids exists] ********************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:41 Sunday 04 February 2024 20:07:30 +0000 (0:00:00.030) 0:01:45.334 ******* ok: [sut] => { "changed": false, "stat": { "atime": 1707077211.5674798, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "076f89863b474504550ab1b5c11216a77ed7123c", "ctime": 1707077179.8497021, "dev": 51713, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 4380701, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-pie-executable", "mode": "0755", "mtime": 1689166618.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 15496, "uid": 0, "version": "2573075764", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [linux-system-roles.podman : Check user with getsubids] ******************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:52 Sunday 04 February 2024 20:07:31 +0000 (0:00:00.196) 0:01:45.530 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Check group with getsubids] ****************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:56 Sunday 04 February 2024 20:07:31 +0000 (0:00:00.024) 0:01:45.555 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Check if user is in subuid file] ************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:66 Sunday 04 February 2024 20:07:31 +0000 (0:00:00.022) 0:01:45.577 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Fail if user not in subuid file] ************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:74 Sunday 04 February 2024 20:07:31 +0000 (0:00:00.023) 0:01:45.600 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Check if group is in subgid file] ************ task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:81 Sunday 04 February 2024 20:07:31 +0000 (0:00:00.022) 0:01:45.623 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Fail if group not in subgid file] ************ task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:89 Sunday 04 February 2024 20:07:31 +0000 (0:00:00.022) 0:01:45.645 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Set per-container variables part 3] ********** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:102 Sunday 04 February 2024 20:07:31 +0000 (0:00:00.023) 0:01:45.669 ******* ok: [sut] => { "ansible_facts": { "__podman_activate_systemd_unit": true, "__podman_images_found": [], "__podman_kube_yamls_raw": "", "__podman_service_name": "", "__podman_systemd_scope": "system", "__podman_user_home_dir": "/root", "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [linux-system-roles.podman : Set per-container variables part 4] ********** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:134 Sunday 04 February 2024 20:07:31 +0000 (0:00:00.039) 0:01:45.709 ******* ok: [sut] => { "ansible_facts": { "__podman_quadlet_path": "/etc/containers/systemd" }, "changed": false } TASK [linux-system-roles.podman : Get kube yaml contents] ********************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:140 Sunday 04 February 2024 20:07:31 +0000 (0:00:00.024) 0:01:45.733 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Set per-container variables part 5] ********** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:151 Sunday 04 February 2024 20:07:31 +0000 (0:00:00.023) 0:01:45.757 ******* ok: [sut] => { "ansible_facts": { "__podman_images": [], "__podman_quadlet_file": "/etc/containers/systemd/envoy-proxy-configmap.yml", "__podman_volumes": [] }, "changed": false } TASK [linux-system-roles.podman : Cleanup quadlets] **************************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:205 Sunday 04 February 2024 20:07:31 +0000 (0:00:00.063) 0:01:45.821 ******* included: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/cleanup_quadlet_spec.yml for sut TASK [linux-system-roles.podman : Stop and disable service] ******************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/cleanup_quadlet_spec.yml:4 Sunday 04 February 2024 20:07:31 +0000 (0:00:00.048) 0:01:45.870 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Remove quadlet file] ************************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/cleanup_quadlet_spec.yml:22 Sunday 04 February 2024 20:07:31 +0000 (0:00:00.025) 0:01:45.895 ******* ok: [sut] => { "changed": false, "path": "/etc/containers/systemd/envoy-proxy-configmap.yml", "state": "absent" } TASK [linux-system-roles.podman : Manage linger] ******************************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/cleanup_quadlet_spec.yml:28 Sunday 04 February 2024 20:07:31 +0000 (0:00:00.200) 0:01:46.095 ******* included: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/manage_linger.yml for sut TASK [linux-system-roles.podman : Enable linger if needed] ********************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/manage_linger.yml:12 Sunday 04 February 2024 20:07:31 +0000 (0:00:00.037) 0:01:46.132 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Mark user as not yet needing to cancel linger] *** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/manage_linger.yml:18 Sunday 04 February 2024 20:07:31 +0000 (0:00:00.023) 0:01:46.156 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Mark user for possible linger cancel] ******** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/manage_linger.yml:22 Sunday 04 February 2024 20:07:31 +0000 (0:00:00.059) 0:01:46.215 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Reload systemctl] **************************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/cleanup_quadlet_spec.yml:36 Sunday 04 February 2024 20:07:31 +0000 (0:00:00.024) 0:01:46.240 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Prune images no longer in use] *************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/cleanup_quadlet_spec.yml:45 Sunday 04 February 2024 20:07:31 +0000 (0:00:00.025) 0:01:46.266 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Remove network] ****************************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/cleanup_quadlet_spec.yml:53 Sunday 04 February 2024 20:07:31 +0000 (0:00:00.025) 0:01:46.292 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Create and update quadlets] ****************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:209 Sunday 04 February 2024 20:07:31 +0000 (0:00:00.023) 0:01:46.315 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Set per-container variables part 0] ********** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:14 Sunday 04 February 2024 20:07:31 +0000 (0:00:00.022) 0:01:46.337 ******* ok: [sut] => { "ansible_facts": { "__podman_quadlet_file_src": "", "__podman_quadlet_spec": {}, "__podman_quadlet_str": "---\napiVersion: v1\nkind: PersistentVolumeClaim\nmetadata:\n name: wp-pv-claim\n labels:\n app: wordpress\nspec:\n accessModes:\n - ReadWriteOnce\n resources:\n requests:\n storage: 20Gi\n---\napiVersion: v1\nkind: Pod\nmetadata:\n name: quadlet-demo\nspec:\n containers:\n - name: wordpress\n image: quay.io/linux-system-roles/wordpress:4.8-apache\n env:\n - name: WORDPRESS_DB_HOST\n value: quadlet-demo-mysql\n - name: WORDPRESS_DB_PASSWORD\n valueFrom:\n secretKeyRef:\n name: mysql-root-password-kube\n key: password\n volumeMounts:\n - name: wordpress-persistent-storage\n mountPath: /var/www/html\n resources:\n requests:\n memory: \"64Mi\"\n cpu: \"250m\"\n limits:\n memory: \"128Mi\"\n cpu: \"500m\"\n - name: envoy\n image: quay.io/linux-system-roles/envoyproxy:v1.25.0\n volumeMounts:\n - name: config-volume\n mountPath: /etc/envoy\n - name: certificates\n mountPath: /etc/envoy-certificates\n env:\n - name: ENVOY_UID\n value: \"0\"\n resources:\n requests:\n memory: \"64Mi\"\n cpu: \"250m\"\n limits:\n memory: \"128Mi\"\n cpu: \"500m\"\n volumes:\n - name: config-volume\n configMap:\n name: envoy-proxy-config\n - name: certificates\n secret:\n secretName: envoy-certificates\n - name: wordpress-persistent-storage\n persistentVolumeClaim:\n claimName: wp-pv-claim\n - name: www # not used - for testing hostpath\n hostPath:\n path: /tmp/httpd3\n - name: create # not used - for testing hostpath\n hostPath:\n path: /tmp/httpd3-create\n", "__podman_quadlet_template_src": "quadlet-demo.yml.j2" }, "changed": false } TASK [linux-system-roles.podman : Set per-container variables part 1] ********** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:40 Sunday 04 February 2024 20:07:31 +0000 (0:00:00.082) 0:01:46.420 ******* ok: [sut] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_pull_image": true, "__podman_state": "absent", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [linux-system-roles.podman : Fail if no quadlet spec is given] ************ task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:57 Sunday 04 February 2024 20:07:31 +0000 (0:00:00.031) 0:01:46.451 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Set per-container variables part 2] ********** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:70 Sunday 04 February 2024 20:07:31 +0000 (0:00:00.028) 0:01:46.480 ******* ok: [sut] => { "ansible_facts": { "__podman_quadlet_name": "quadlet-demo", "__podman_quadlet_type": "yml", "__podman_rootless": false }, "changed": false } TASK [linux-system-roles.podman : Check user and group information] ************ task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:97 Sunday 04 February 2024 20:07:32 +0000 (0:00:00.036) 0:01:46.517 ******* included: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml for sut TASK [linux-system-roles.podman : Get user information] ************************ task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:2 Sunday 04 February 2024 20:07:32 +0000 (0:00:00.043) 0:01:46.560 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Fail if user does not exist] ***************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:10 Sunday 04 February 2024 20:07:32 +0000 (0:00:00.028) 0:01:46.588 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Set group for podman user] ******************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:17 Sunday 04 February 2024 20:07:32 +0000 (0:00:00.026) 0:01:46.615 ******* ok: [sut] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [linux-system-roles.podman : Get group information] *********************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:28 Sunday 04 February 2024 20:07:32 +0000 (0:00:00.030) 0:01:46.646 ******* ok: [sut] => { "ansible_facts": { "getent_group": { "root": [ "x", "0", "" ] } }, "changed": false } TASK [linux-system-roles.podman : Set group name] ****************************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:36 Sunday 04 February 2024 20:07:32 +0000 (0:00:00.201) 0:01:46.848 ******* ok: [sut] => { "ansible_facts": { "__podman_group_name": "root" }, "changed": false } TASK [linux-system-roles.podman : See if getsubids exists] ********************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:41 Sunday 04 February 2024 20:07:32 +0000 (0:00:00.029) 0:01:46.878 ******* ok: [sut] => { "changed": false, "stat": { "atime": 1707077211.5674798, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "076f89863b474504550ab1b5c11216a77ed7123c", "ctime": 1707077179.8497021, "dev": 51713, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 4380701, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-pie-executable", "mode": "0755", "mtime": 1689166618.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 15496, "uid": 0, "version": "2573075764", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [linux-system-roles.podman : Check user with getsubids] ******************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:52 Sunday 04 February 2024 20:07:32 +0000 (0:00:00.228) 0:01:47.107 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Check group with getsubids] ****************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:56 Sunday 04 February 2024 20:07:32 +0000 (0:00:00.026) 0:01:47.134 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Check if user is in subuid file] ************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:66 Sunday 04 February 2024 20:07:32 +0000 (0:00:00.029) 0:01:47.163 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Fail if user not in subuid file] ************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:74 Sunday 04 February 2024 20:07:32 +0000 (0:00:00.023) 0:01:47.186 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Check if group is in subgid file] ************ task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:81 Sunday 04 February 2024 20:07:32 +0000 (0:00:00.023) 0:01:47.210 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Fail if group not in subgid file] ************ task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:89 Sunday 04 February 2024 20:07:32 +0000 (0:00:00.022) 0:01:47.232 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Set per-container variables part 3] ********** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:102 Sunday 04 February 2024 20:07:32 +0000 (0:00:00.023) 0:01:47.256 ******* ok: [sut] => { "ansible_facts": { "__podman_activate_systemd_unit": true, "__podman_images_found": [], "__podman_kube_yamls_raw": "", "__podman_service_name": "", "__podman_systemd_scope": "system", "__podman_user_home_dir": "/root", "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [linux-system-roles.podman : Set per-container variables part 4] ********** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:134 Sunday 04 February 2024 20:07:32 +0000 (0:00:00.042) 0:01:47.299 ******* ok: [sut] => { "ansible_facts": { "__podman_quadlet_path": "/etc/containers/systemd" }, "changed": false } TASK [linux-system-roles.podman : Get kube yaml contents] ********************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:140 Sunday 04 February 2024 20:07:32 +0000 (0:00:00.025) 0:01:47.324 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Set per-container variables part 5] ********** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:151 Sunday 04 February 2024 20:07:32 +0000 (0:00:00.024) 0:01:47.348 ******* ok: [sut] => { "ansible_facts": { "__podman_images": [], "__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo.yml", "__podman_volumes": [] }, "changed": false } TASK [linux-system-roles.podman : Cleanup quadlets] **************************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:205 Sunday 04 February 2024 20:07:32 +0000 (0:00:00.059) 0:01:47.407 ******* included: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/cleanup_quadlet_spec.yml for sut TASK [linux-system-roles.podman : Stop and disable service] ******************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/cleanup_quadlet_spec.yml:4 Sunday 04 February 2024 20:07:32 +0000 (0:00:00.048) 0:01:47.456 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Remove quadlet file] ************************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/cleanup_quadlet_spec.yml:22 Sunday 04 February 2024 20:07:32 +0000 (0:00:00.027) 0:01:47.483 ******* ok: [sut] => { "changed": false, "path": "/etc/containers/systemd/quadlet-demo.yml", "state": "absent" } TASK [linux-system-roles.podman : Manage linger] ******************************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/cleanup_quadlet_spec.yml:28 Sunday 04 February 2024 20:07:33 +0000 (0:00:00.201) 0:01:47.685 ******* included: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/manage_linger.yml for sut TASK [linux-system-roles.podman : Enable linger if needed] ********************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/manage_linger.yml:12 Sunday 04 February 2024 20:07:33 +0000 (0:00:00.039) 0:01:47.724 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Mark user as not yet needing to cancel linger] *** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/manage_linger.yml:18 Sunday 04 February 2024 20:07:33 +0000 (0:00:00.024) 0:01:47.749 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Mark user for possible linger cancel] ******** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/manage_linger.yml:22 Sunday 04 February 2024 20:07:33 +0000 (0:00:00.025) 0:01:47.775 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Reload systemctl] **************************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/cleanup_quadlet_spec.yml:36 Sunday 04 February 2024 20:07:33 +0000 (0:00:00.059) 0:01:47.834 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Prune images no longer in use] *************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/cleanup_quadlet_spec.yml:45 Sunday 04 February 2024 20:07:33 +0000 (0:00:00.025) 0:01:47.860 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Remove network] ****************************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/cleanup_quadlet_spec.yml:53 Sunday 04 February 2024 20:07:33 +0000 (0:00:00.026) 0:01:47.887 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Create and update quadlets] ****************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:209 Sunday 04 February 2024 20:07:33 +0000 (0:00:00.025) 0:01:47.912 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Set per-container variables part 0] ********** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:14 Sunday 04 February 2024 20:07:33 +0000 (0:00:00.025) 0:01:47.938 ******* ok: [sut] => { "ansible_facts": { "__podman_quadlet_file_src": "quadlet-demo.kube", "__podman_quadlet_spec": {}, "__podman_quadlet_str": "[Install]\nWantedBy=default.target\n\n[Unit]\nRequires=quadlet-demo-mysql.service\nAfter=quadlet-demo-mysql.service\n\n[Kube]\n# Point to the yaml file in the same directory\nYaml=quadlet-demo.yml\n# Use the quadlet-demo network\nNetwork=quadlet-demo.network\n# Publish the envoy proxy data port\nPublishPort=8000:8080\n# Publish the envoy proxy admin port\nPublishPort=9000:9901\n# Use the envoy proxy config map in the same directory\nConfigMap=envoy-proxy-configmap.yml", "__podman_quadlet_template_src": "" }, "changed": false } TASK [linux-system-roles.podman : Set per-container variables part 1] ********** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:40 Sunday 04 February 2024 20:07:33 +0000 (0:00:00.031) 0:01:47.970 ******* ok: [sut] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_pull_image": true, "__podman_state": "absent", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [linux-system-roles.podman : Fail if no quadlet spec is given] ************ task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:57 Sunday 04 February 2024 20:07:33 +0000 (0:00:00.029) 0:01:48.000 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Set per-container variables part 2] ********** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:70 Sunday 04 February 2024 20:07:33 +0000 (0:00:00.024) 0:01:48.024 ******* ok: [sut] => { "ansible_facts": { "__podman_quadlet_name": "quadlet-demo", "__podman_quadlet_type": "kube", "__podman_rootless": false }, "changed": false } TASK [linux-system-roles.podman : Check user and group information] ************ task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:97 Sunday 04 February 2024 20:07:33 +0000 (0:00:00.032) 0:01:48.056 ******* included: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml for sut TASK [linux-system-roles.podman : Get user information] ************************ task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:2 Sunday 04 February 2024 20:07:33 +0000 (0:00:00.041) 0:01:48.097 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Fail if user does not exist] ***************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:10 Sunday 04 February 2024 20:07:33 +0000 (0:00:00.027) 0:01:48.125 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Set group for podman user] ******************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:17 Sunday 04 February 2024 20:07:33 +0000 (0:00:00.026) 0:01:48.152 ******* ok: [sut] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [linux-system-roles.podman : Get group information] *********************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:28 Sunday 04 February 2024 20:07:33 +0000 (0:00:00.031) 0:01:48.183 ******* ok: [sut] => { "ansible_facts": { "getent_group": { "root": [ "x", "0", "" ] } }, "changed": false } TASK [linux-system-roles.podman : Set group name] ****************************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:36 Sunday 04 February 2024 20:07:33 +0000 (0:00:00.203) 0:01:48.387 ******* ok: [sut] => { "ansible_facts": { "__podman_group_name": "root" }, "changed": false } TASK [linux-system-roles.podman : See if getsubids exists] ********************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:41 Sunday 04 February 2024 20:07:33 +0000 (0:00:00.032) 0:01:48.419 ******* ok: [sut] => { "changed": false, "stat": { "atime": 1707077211.5674798, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "076f89863b474504550ab1b5c11216a77ed7123c", "ctime": 1707077179.8497021, "dev": 51713, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 4380701, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-pie-executable", "mode": "0755", "mtime": 1689166618.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 15496, "uid": 0, "version": "2573075764", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [linux-system-roles.podman : Check user with getsubids] ******************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:52 Sunday 04 February 2024 20:07:34 +0000 (0:00:00.200) 0:01:48.619 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Check group with getsubids] ****************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:56 Sunday 04 February 2024 20:07:34 +0000 (0:00:00.022) 0:01:48.641 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Check if user is in subuid file] ************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:66 Sunday 04 February 2024 20:07:34 +0000 (0:00:00.056) 0:01:48.697 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Fail if user not in subuid file] ************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:74 Sunday 04 February 2024 20:07:34 +0000 (0:00:00.025) 0:01:48.723 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Check if group is in subgid file] ************ task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:81 Sunday 04 February 2024 20:07:34 +0000 (0:00:00.027) 0:01:48.751 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Fail if group not in subgid file] ************ task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:89 Sunday 04 February 2024 20:07:34 +0000 (0:00:00.028) 0:01:48.779 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Set per-container variables part 3] ********** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:102 Sunday 04 February 2024 20:07:34 +0000 (0:00:00.026) 0:01:48.806 ******* ok: [sut] => { "ansible_facts": { "__podman_activate_systemd_unit": true, "__podman_images_found": [], "__podman_kube_yamls_raw": [ "quadlet-demo.yml" ], "__podman_service_name": "quadlet-demo.service", "__podman_systemd_scope": "system", "__podman_user_home_dir": "/root", "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [linux-system-roles.podman : Set per-container variables part 4] ********** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:134 Sunday 04 February 2024 20:07:34 +0000 (0:00:00.055) 0:01:48.861 ******* ok: [sut] => { "ansible_facts": { "__podman_quadlet_path": "/etc/containers/systemd" }, "changed": false } TASK [linux-system-roles.podman : Get kube yaml contents] ********************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:140 Sunday 04 February 2024 20:07:34 +0000 (0:00:00.027) 0:01:48.889 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Set per-container variables part 5] ********** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:151 Sunday 04 February 2024 20:07:34 +0000 (0:00:00.024) 0:01:48.913 ******* ok: [sut] => { "ansible_facts": { "__podman_images": [], "__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo.kube", "__podman_volumes": [] }, "changed": false } TASK [linux-system-roles.podman : Cleanup quadlets] **************************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:205 Sunday 04 February 2024 20:07:34 +0000 (0:00:00.058) 0:01:48.972 ******* included: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/cleanup_quadlet_spec.yml for sut TASK [linux-system-roles.podman : Stop and disable service] ******************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/cleanup_quadlet_spec.yml:4 Sunday 04 February 2024 20:07:34 +0000 (0:00:00.048) 0:01:49.021 ******* ok: [sut] => { "changed": false, "failed_when_result": false } MSG: Could not find the requested service quadlet-demo.service: host TASK [linux-system-roles.podman : Remove quadlet file] ************************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/cleanup_quadlet_spec.yml:22 Sunday 04 February 2024 20:07:34 +0000 (0:00:00.330) 0:01:49.352 ******* ok: [sut] => { "changed": false, "path": "/etc/containers/systemd/quadlet-demo.kube", "state": "absent" } TASK [linux-system-roles.podman : Manage linger] ******************************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/cleanup_quadlet_spec.yml:28 Sunday 04 February 2024 20:07:35 +0000 (0:00:00.201) 0:01:49.553 ******* included: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/manage_linger.yml for sut TASK [linux-system-roles.podman : Enable linger if needed] ********************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/manage_linger.yml:12 Sunday 04 February 2024 20:07:35 +0000 (0:00:00.041) 0:01:49.594 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Mark user as not yet needing to cancel linger] *** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/manage_linger.yml:18 Sunday 04 February 2024 20:07:35 +0000 (0:00:00.025) 0:01:49.619 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Mark user for possible linger cancel] ******** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/manage_linger.yml:22 Sunday 04 February 2024 20:07:35 +0000 (0:00:00.036) 0:01:49.656 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Reload systemctl] **************************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/cleanup_quadlet_spec.yml:36 Sunday 04 February 2024 20:07:35 +0000 (0:00:00.028) 0:01:49.685 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Prune images no longer in use] *************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/cleanup_quadlet_spec.yml:45 Sunday 04 February 2024 20:07:35 +0000 (0:00:00.074) 0:01:49.759 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Remove network] ****************************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/cleanup_quadlet_spec.yml:53 Sunday 04 February 2024 20:07:35 +0000 (0:00:00.038) 0:01:49.798 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Create and update quadlets] ****************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:209 Sunday 04 February 2024 20:07:35 +0000 (0:00:00.037) 0:01:49.835 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Set per-container variables part 0] ********** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:14 Sunday 04 February 2024 20:07:35 +0000 (0:00:00.026) 0:01:49.862 ******* ok: [sut] => { "ansible_facts": { "__podman_quadlet_file_src": "quadlet-demo.network", "__podman_quadlet_spec": {}, "__podman_quadlet_str": "[Network]\nSubnet=192.168.30.0/24\nGateway=192.168.30.1\nLabel=app=wordpress", "__podman_quadlet_template_src": "" }, "changed": false } TASK [linux-system-roles.podman : Set per-container variables part 1] ********** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:40 Sunday 04 February 2024 20:07:35 +0000 (0:00:00.040) 0:01:49.903 ******* ok: [sut] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_pull_image": true, "__podman_state": "absent", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [linux-system-roles.podman : Fail if no quadlet spec is given] ************ task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:57 Sunday 04 February 2024 20:07:35 +0000 (0:00:00.043) 0:01:49.946 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Set per-container variables part 2] ********** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:70 Sunday 04 February 2024 20:07:35 +0000 (0:00:00.037) 0:01:49.984 ******* ok: [sut] => { "ansible_facts": { "__podman_quadlet_name": "quadlet-demo", "__podman_quadlet_type": "network", "__podman_rootless": false }, "changed": false } TASK [linux-system-roles.podman : Check user and group information] ************ task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:97 Sunday 04 February 2024 20:07:35 +0000 (0:00:00.038) 0:01:50.023 ******* included: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml for sut TASK [linux-system-roles.podman : Get user information] ************************ task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:2 Sunday 04 February 2024 20:07:35 +0000 (0:00:00.047) 0:01:50.070 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Fail if user does not exist] ***************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:10 Sunday 04 February 2024 20:07:35 +0000 (0:00:00.038) 0:01:50.109 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Set group for podman user] ******************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:17 Sunday 04 February 2024 20:07:35 +0000 (0:00:00.033) 0:01:50.142 ******* ok: [sut] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [linux-system-roles.podman : Get group information] *********************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:28 Sunday 04 February 2024 20:07:35 +0000 (0:00:00.034) 0:01:50.177 ******* ok: [sut] => { "ansible_facts": { "getent_group": { "root": [ "x", "0", "" ] } }, "changed": false } TASK [linux-system-roles.podman : Set group name] ****************************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:36 Sunday 04 February 2024 20:07:35 +0000 (0:00:00.205) 0:01:50.382 ******* ok: [sut] => { "ansible_facts": { "__podman_group_name": "root" }, "changed": false } TASK [linux-system-roles.podman : See if getsubids exists] ********************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:41 Sunday 04 February 2024 20:07:35 +0000 (0:00:00.032) 0:01:50.415 ******* ok: [sut] => { "changed": false, "stat": { "atime": 1707077211.5674798, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "076f89863b474504550ab1b5c11216a77ed7123c", "ctime": 1707077179.8497021, "dev": 51713, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 4380701, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-pie-executable", "mode": "0755", "mtime": 1689166618.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 15496, "uid": 0, "version": "2573075764", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [linux-system-roles.podman : Check user with getsubids] ******************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:52 Sunday 04 February 2024 20:07:36 +0000 (0:00:00.202) 0:01:50.617 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Check group with getsubids] ****************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:56 Sunday 04 February 2024 20:07:36 +0000 (0:00:00.023) 0:01:50.640 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Check if user is in subuid file] ************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:66 Sunday 04 February 2024 20:07:36 +0000 (0:00:00.055) 0:01:50.696 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Fail if user not in subuid file] ************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:74 Sunday 04 February 2024 20:07:36 +0000 (0:00:00.027) 0:01:50.723 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Check if group is in subgid file] ************ task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:81 Sunday 04 February 2024 20:07:36 +0000 (0:00:00.024) 0:01:50.748 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Fail if group not in subgid file] ************ task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:89 Sunday 04 February 2024 20:07:36 +0000 (0:00:00.023) 0:01:50.771 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Set per-container variables part 3] ********** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:102 Sunday 04 February 2024 20:07:36 +0000 (0:00:00.023) 0:01:50.795 ******* ok: [sut] => { "ansible_facts": { "__podman_activate_systemd_unit": true, "__podman_images_found": [], "__podman_kube_yamls_raw": "", "__podman_service_name": "", "__podman_systemd_scope": "system", "__podman_user_home_dir": "/root", "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [linux-system-roles.podman : Set per-container variables part 4] ********** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:134 Sunday 04 February 2024 20:07:36 +0000 (0:00:00.041) 0:01:50.836 ******* ok: [sut] => { "ansible_facts": { "__podman_quadlet_path": "/etc/containers/systemd" }, "changed": false } TASK [linux-system-roles.podman : Get kube yaml contents] ********************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:140 Sunday 04 February 2024 20:07:36 +0000 (0:00:00.027) 0:01:50.863 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Set per-container variables part 5] ********** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:151 Sunday 04 February 2024 20:07:36 +0000 (0:00:00.023) 0:01:50.887 ******* ok: [sut] => { "ansible_facts": { "__podman_images": [], "__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo.network", "__podman_volumes": [] }, "changed": false } TASK [linux-system-roles.podman : Cleanup quadlets] **************************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:205 Sunday 04 February 2024 20:07:36 +0000 (0:00:00.063) 0:01:50.950 ******* included: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/cleanup_quadlet_spec.yml for sut TASK [linux-system-roles.podman : Stop and disable service] ******************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/cleanup_quadlet_spec.yml:4 Sunday 04 February 2024 20:07:36 +0000 (0:00:00.050) 0:01:51.000 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Remove quadlet file] ************************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/cleanup_quadlet_spec.yml:22 Sunday 04 February 2024 20:07:36 +0000 (0:00:00.026) 0:01:51.027 ******* changed: [sut] => { "changed": true, "path": "/etc/containers/systemd/quadlet-demo.network", "state": "absent" } TASK [linux-system-roles.podman : Manage linger] ******************************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/cleanup_quadlet_spec.yml:28 Sunday 04 February 2024 20:07:36 +0000 (0:00:00.202) 0:01:51.229 ******* included: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/manage_linger.yml for sut TASK [linux-system-roles.podman : Enable linger if needed] ********************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/manage_linger.yml:12 Sunday 04 February 2024 20:07:36 +0000 (0:00:00.042) 0:01:51.271 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Mark user as not yet needing to cancel linger] *** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/manage_linger.yml:18 Sunday 04 February 2024 20:07:36 +0000 (0:00:00.024) 0:01:51.296 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Mark user for possible linger cancel] ******** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/manage_linger.yml:22 Sunday 04 February 2024 20:07:36 +0000 (0:00:00.024) 0:01:51.320 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Reload systemctl] **************************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/cleanup_quadlet_spec.yml:36 Sunday 04 February 2024 20:07:36 +0000 (0:00:00.026) 0:01:51.347 ******* ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.podman : Prune images no longer in use] *************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/cleanup_quadlet_spec.yml:45 Sunday 04 February 2024 20:07:37 +0000 (0:00:00.507) 0:01:51.854 ******* changed: [sut] => { "changed": true, "cmd": [ "podman", "image", "prune", "-f" ], "delta": "0:00:00.034636", "end": "2024-02-04 20:07:37.565887", "rc": 0, "start": "2024-02-04 20:07:37.531251" } TASK [linux-system-roles.podman : Remove network] ****************************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/cleanup_quadlet_spec.yml:53 Sunday 04 February 2024 20:07:37 +0000 (0:00:00.235) 0:01:52.090 ******* changed: [sut] => { "changed": true, "cmd": [ "podman", "network", "rm", "systemd-quadlet-demo" ], "delta": "0:00:00.036757", "end": "2024-02-04 20:07:37.803357", "rc": 0, "start": "2024-02-04 20:07:37.766600" } STDOUT: systemd-quadlet-demo TASK [linux-system-roles.podman : Create and update quadlets] ****************** task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:209 Sunday 04 February 2024 20:07:37 +0000 (0:00:00.243) 0:01:52.333 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Cancel linger] ******************************* task path: /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/main.yml:132 Sunday 04 February 2024 20:07:37 +0000 (0:00:00.023) 0:01:52.357 ******* META: ran handlers META: ran handlers PLAY RECAP ********************************************************************* sut : ok=264 changed=25 unreachable=0 failed=0 skipped=249 rescued=1 ignored=0 Sunday 04 February 2024 20:07:37 +0000 (0:00:00.007) 0:01:52.364 ******* =============================================================================== linux-system-roles.podman : Ensure required packages are installed ----- 49.70s /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/main.yml:14 linux-system-roles.podman : Ensure container images are present --------- 8.33s /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/create_update_quadlet_spec.yml:25 fedora.linux_system_roles.certificate : Ensure certificate role dependencies are installed --- 5.67s /WORKDIR/git-fix-linger-handling7nz_rqha/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:5 fedora.linux_system_roles.certificate : Ensure provider packages are installed --- 3.27s /WORKDIR/git-fix-linger-handling7nz_rqha/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:23 fedora.linux_system_roles.firewall : Configure firewall ----------------- 2.35s /WORKDIR/git-fix-linger-handling7nz_rqha/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:71 fedora.linux_system_roles.firewall : Enable and start firewalld service --- 2.02s /WORKDIR/git-fix-linger-handling7nz_rqha/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:28 linux-system-roles.podman : Gather the package facts -------------------- 1.13s /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/main.yml:6 linux-system-roles.podman : Gather the package facts -------------------- 0.95s /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/main.yml:6 fedora.linux_system_roles.firewall : Install firewalld ------------------ 0.89s /WORKDIR/git-fix-linger-handling7nz_rqha/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:19 fedora.linux_system_roles.firewall : Install firewalld ------------------ 0.89s /WORKDIR/git-fix-linger-handling7nz_rqha/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:19 fedora.linux_system_roles.certificate : Ensure certificate requests ----- 0.85s /WORKDIR/git-fix-linger-handling7nz_rqha/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:101 fedora.linux_system_roles.certificate : Ensure provider service is running --- 0.74s /WORKDIR/git-fix-linger-handling7nz_rqha/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:90 fedora.linux_system_roles.firewall : Unmask firewalld service ----------- 0.71s /WORKDIR/git-fix-linger-handling7nz_rqha/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:22 Gathering Facts --------------------------------------------------------- 0.70s /WORKDIR/git-fix-linger-handling7nz_rqha/tests/tests_quadlet_demo.yml:9 ------- fedora.linux_system_roles.certificate : Slurp the contents of the files --- 0.61s /WORKDIR/git-fix-linger-handling7nz_rqha/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:152 linux-system-roles.podman : Ensure quadlet file is copied --------------- 0.61s /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/create_update_quadlet_spec.yml:51 linux-system-roles.podman : Reload systemctl ---------------------------- 0.58s /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/cleanup_quadlet_spec.yml:36 fedora.linux_system_roles.certificate : Remove files -------------------- 0.56s /WORKDIR/git-fix-linger-handling7nz_rqha/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:181 linux-system-roles.podman : Stop and disable service -------------------- 0.55s /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/cleanup_quadlet_spec.yml:4 linux-system-roles.podman : Stop and disable service -------------------- 0.54s /WORKDIR/git-fix-linger-handling7nz_rqha/tests/roles/linux-system-roles.podman/tasks/cleanup_quadlet_spec.yml:4 ---^---^---^---^---^--- # STDERR: ---v---v---v---v---v--- [WARNING]: Using a variable for a task's 'args' is unsafe in some situations (see https://docs.ansible.com/ansible/devel/reference_appendices/faq.html#argsplat- unsafe) ---^---^---^---^---^---