# STDOUT: ---v---v---v---v---v--- ansible-playbook [core 2.16.0] config file = /etc/ansible/ansible.cfg configured module search path = ['/home/jenkins/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules'] ansible python module location = /opt/ansible-2.16/lib/python3.11/site-packages/ansible ansible collection location = /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection executable location = /opt/ansible-2.16/bin/ansible-playbook python version = 3.11.5 (main, Sep 7 2023, 00:00:00) [GCC 11.4.1 20230605 (Red Hat 11.4.1-2)] (/opt/ansible-2.16/bin/python) jinja version = 3.1.2 libyaml = True Using /etc/ansible/ansible.cfg as config file Skipping callback 'debug', as we already have a stdout callback. Skipping callback 'default', as we already have a stdout callback. Skipping callback 'minimal', as we already have a stdout callback. Skipping callback 'oneline', as we already have a stdout callback. PLAYBOOK: tests_quadlet_demo.yml *********************************************** 2 plays in /WORKDIR/git-fix-linger-handlinggb26lsz1/tests/tests_quadlet_demo.yml PLAY [all] ********************************************************************* TASK [Include vault variables] ************************************************* task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/tests/tests_quadlet_demo.yml:4 Monday 05 February 2024 04:35:37 +0000 (0:00:00.011) 0:00:00.011 ******* ok: [sut] => { "ansible_facts": { "mysql_container_root_password": { "__ansible_vault": "$ANSIBLE_VAULT;1.1;AES256\n61333932373230333539663035366431326163363166363036323963623131363530326231303634\n6635326161643165363366323062333334363730376631660a393566366139353861656364656661\n38653463363837336639363032646433666361646535366137303464623261313663643336306465\n6264663730656337310a343962353137386238383064646533366433333437303566656433386233\n34343235326665646661623131643335313236313131353661386338343366316261643634653633\n3832313034366536616531323963333234326461353130303532\n" } }, "ansible_included_var_files": [ "/WORKDIR/git-fix-linger-handlinggb26lsz1/tests/vars/vault-variables.yml" ], "changed": false } PLAY [Deploy the quadlet demo app] ********************************************* TASK [Gathering Facts] ********************************************************* task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/tests/tests_quadlet_demo.yml:9 Monday 05 February 2024 04:35:37 +0000 (0:00:00.012) 0:00:00.024 ******* ok: [sut] TASK [Generate certificates] *************************************************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/tests/tests_quadlet_demo.yml:39 Monday 05 February 2024 04:35:38 +0000 (0:00:00.761) 0:00:00.785 ******* TASK [fedora.linux_system_roles.certificate : Set version specific variables] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:2 Monday 05 February 2024 04:35:38 +0000 (0:00:00.022) 0:00:00.808 ******* included: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/set_vars.yml for sut TASK [fedora.linux_system_roles.certificate : Ensure ansible_facts used by role] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/set_vars.yml:2 Monday 05 February 2024 04:35:38 +0000 (0:00:00.012) 0:00:00.820 ******* skipping: [sut] => { "changed": false, "false_condition": "__certificate_required_facts | difference(ansible_facts.keys() | list) | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.certificate : Check if system is ostree] ******* task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/set_vars.yml:10 Monday 05 February 2024 04:35:38 +0000 (0:00:00.018) 0:00:00.839 ******* ok: [sut] => { "changed": false, "stat": { "exists": false } } TASK [fedora.linux_system_roles.certificate : Set flag to indicate system is ostree] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/set_vars.yml:15 Monday 05 February 2024 04:35:38 +0000 (0:00:00.240) 0:00:01.079 ******* ok: [sut] => { "ansible_facts": { "__certificate_is_ostree": false }, "changed": false } TASK [fedora.linux_system_roles.certificate : Set platform/version specific variables] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/set_vars.yml:19 Monday 05 February 2024 04:35:38 +0000 (0:00:00.015) 0:00:01.095 ******* skipping: [sut] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "false_condition": "__vars_file is file", "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=CentOS.yml) => { "ansible_loop_var": "item", "changed": false, "false_condition": "__vars_file is file", "item": "CentOS.yml", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=CentOS_8.yml) => { "ansible_loop_var": "item", "changed": false, "false_condition": "__vars_file is file", "item": "CentOS_8.yml", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=CentOS_8.yml) => { "ansible_loop_var": "item", "changed": false, "false_condition": "__vars_file is file", "item": "CentOS_8.yml", "skip_reason": "Conditional result was False" } skipping: [sut] => { "changed": false } MSG: All items skipped TASK [fedora.linux_system_roles.certificate : Ensure certificate role dependencies are installed] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:5 Monday 05 February 2024 04:35:39 +0000 (0:00:00.019) 0:00:01.114 ******* changed: [sut] => { "changed": true, "rc": 0, "results": [ "Installed: python3-pyasn1-0.3.7-6.el8.noarch" ] } TASK [fedora.linux_system_roles.certificate : Ensure provider packages are installed] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:23 Monday 05 February 2024 04:35:42 +0000 (0:00:03.530) 0:00:04.645 ******* changed: [sut] => (item=certmonger) => { "__certificate_provider": "certmonger", "ansible_loop_var": "__certificate_provider", "changed": true, "rc": 0, "results": [ "Installed: certmonger-0.79.17-2.el8.x86_64", "Installed: xmlrpc-c-1.51.0-8.el8.x86_64", "Installed: xmlrpc-c-client-1.51.0-8.el8.x86_64" ] } TASK [fedora.linux_system_roles.certificate : Ensure pre-scripts hooks directory exists] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:35 Monday 05 February 2024 04:35:46 +0000 (0:00:04.200) 0:00:08.846 ******* changed: [sut] => (item=certmonger) => { "__certificate_provider": "certmonger", "ansible_loop_var": "__certificate_provider", "changed": true, "gid": 0, "group": "root", "mode": "0700", "owner": "root", "path": "/etc/certmonger//pre-scripts", "secontext": "unconfined_u:object_r:etc_t:s0", "size": 6, "state": "directory", "uid": 0 } TASK [fedora.linux_system_roles.certificate : Ensure post-scripts hooks directory exists] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:61 Monday 05 February 2024 04:35:47 +0000 (0:00:00.272) 0:00:09.118 ******* changed: [sut] => (item=certmonger) => { "__certificate_provider": "certmonger", "ansible_loop_var": "__certificate_provider", "changed": true, "gid": 0, "group": "root", "mode": "0700", "owner": "root", "path": "/etc/certmonger//post-scripts", "secontext": "unconfined_u:object_r:etc_t:s0", "size": 6, "state": "directory", "uid": 0 } TASK [fedora.linux_system_roles.certificate : Ensure provider service is running] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:90 Monday 05 February 2024 04:35:47 +0000 (0:00:00.213) 0:00:09.332 ******* changed: [sut] => (item=certmonger) => { "__certificate_provider": "certmonger", "ansible_loop_var": "__certificate_provider", "changed": true, "enabled": true, "name": "certmonger", "state": "started", "status": { "ActiveEnterTimestampMonotonic": "0", "ActiveExitTimestampMonotonic": "0", "ActiveState": "inactive", "After": "basic.target dbus.socket sysinit.target syslog.target dbus.service network.target system.slice systemd-journald.socket", "AllowIsolate": "no", "AllowedCPUs": "", "AllowedMemoryNodes": "", "AmbientCapabilities": "", "AssertResult": "no", "AssertTimestampMonotonic": "0", "Before": "shutdown.target", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "BusName": "org.fedorahosted.certmonger", "CPUAccounting": "no", "CPUAffinity": "", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "[not set]", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "no", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf", "CollectMode": "inactive", "ConditionResult": "no", "ConditionTimestampMonotonic": "0", "ConfigurationDirectoryMode": "0755", "Conflicts": "shutdown.target", "ControlPID": "0", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "Delegate": "no", "Description": "Certificate monitoring and PKI enrollment", "DevicePolicy": "auto", "DynamicUser": "no", "EffectiveCPUs": "", "EffectiveMemoryNodes": "", "EnvironmentFiles": "/etc/sysconfig/certmonger (ignore_errors=yes)", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "0", "ExecMainStartTimestampMonotonic": "0", "ExecMainStatus": "0", "ExecStart": "{ path=/usr/sbin/certmonger ; argv[]=/usr/sbin/certmonger -S -p /run/certmonger.pid -n $OPTS ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FragmentPath": "/usr/lib/systemd/system/certmonger.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOSchedulingClass": "0", "IOSchedulingPriority": "0", "IOWeight": "[not set]", "IPAccounting": "no", "IPEgressBytes": "18446744073709551615", "IPEgressPackets": "18446744073709551615", "IPIngressBytes": "18446744073709551615", "IPIngressPackets": "18446744073709551615", "Id": "certmonger.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestampMonotonic": "0", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "control-group", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "0", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "65536", "LimitMEMLOCKSoft": "65536", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "262144", "LimitNOFILESoft": "1024", "LimitNPROC": "14004", "LimitNPROCSoft": "14004", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "14004", "LimitSIGPENDINGSoft": "14004", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "0", "MemoryAccounting": "yes", "MemoryCurrent": "[not set]", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemorySwapMax": "infinity", "MountAPIVFS": "no", "MountFlags": "", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAMask": "", "NUMAPolicy": "n/a", "Names": "certmonger.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "PIDFile": "/run/certmonger.pid", "PartOf": "dbus.service", "PermissionsStartOnly": "no", "Perpetual": "no", "PrivateDevices": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "RemainAfterExit": "no", "RemoveIPC": "no", "Requires": "sysinit.target dbus.socket system.slice", "Restart": "no", "RestartUSec": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "Slice": "system.slice", "StandardError": "inherit", "StandardInput": "null", "StandardInputData": "", "StandardOutput": "journal", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StateChangeTimestampMonotonic": "0", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "dead", "SuccessAction": "none", "SyslogFacility": "3", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "0", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "[not set]", "TasksMax": "22406", "TimeoutStartUSec": "1min 30s", "TimeoutStopUSec": "1min 30s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "dbus", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "disabled", "UnitFileState": "disabled", "UtmpMode": "init", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "0" } } TASK [fedora.linux_system_roles.certificate : Ensure certificate requests] ***** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:101 Monday 05 February 2024 04:35:47 +0000 (0:00:00.676) 0:00:10.008 ******* changed: [sut] => (item={'name': 'quadlet_demo', 'dns': ['localhost'], 'ca': 'self-sign'}) => { "ansible_loop_var": "item", "changed": true, "item": { "ca": "self-sign", "dns": [ "localhost" ], "name": "quadlet_demo" } } MSG: Certificate requested (new). TASK [fedora.linux_system_roles.certificate : Slurp the contents of the files] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:152 Monday 05 February 2024 04:35:48 +0000 (0:00:00.714) 0:00:10.722 ******* ok: [sut] => (item=['cert', {'name': 'quadlet_demo', 'dns': ['localhost'], 'ca': 'self-sign'}]) => { "ansible_loop_var": "item", "changed": false, "content": "LS0tLS1CRUdJTiBDRVJUSUZJQ0FURS0tLS0tCk1JSURnakNDQW1xZ0F3SUJBZ0lRQ3Q0aW1GYjFUR3VlcmJwOFlFeHo2REFOQmdrcWhraUc5dzBCQVFzRkFEQlEKTVNBd0hnWURWUVFEREJkTWIyTmhiQ0JUYVdkdWFXNW5JRUYxZEdodmNtbDBlVEVzTUNvR0ExVUVBd3dqTUdGawpaVEl5T1RndE5UWm1OVFJqTm1JdE9XVmhaR0poTjJNdE5qQTBZemN6WlRjd0hoY05NalF3TWpBMU1EUXpOVFE0CldoY05NalV3TWpBMU1EUXpOVFE0V2pBVU1SSXdFQVlEVlFRREV3bHNiMk5oYkdodmMzUXdnZ0VpTUEwR0NTcUcKU0liM0RRRUJBUVVBQTRJQkR3QXdnZ0VLQW9JQkFRQy9SZGhoNlZFVjVhdzBYQlBFaGpoZ2JsbXJobUwxQ1gwRgpYV2hvRHBBQ25XOExtY3F1c2tkM25abzVqc1dmR25vZm80MGdnc0gyMUF6cHphdjRLNHFiRVVoSExmRU42cWUxCjhlYXlLcisrOENOZHlrU2Ixckl1ejhJVXlzRTduM0ttVVY2b2xnZUpxdzB6STg5Vi9SN0Z4MDFudThsKzdydHUKMWRCN25JNFVKTDA5cGhabEZYdE5wdi82b2tuNGllOEd5VEJ1WUhBODVRR285cUJBcG5NdnRwN1lVQTB5WmdXWQpWTGVSSGVrUFBZSFRSdHhYTzdlNE13MTU4UGdmRkppZ3dyc0h0TEF0dEhGWnZMV2lXVkpKR0VLZ25ya09IQko5ClB4LzZoaGtXeGRpZVpBYWJKcklHcloxZU5UQWVIcjNiUkFOTUpFZjlPQzl4UHZtVDU5RWxBZ01CQUFHamdaTXcKZ1pBd0N3WURWUjBQQkFRREFnV2dNQlFHQTFVZEVRUU5NQXVDQ1d4dlkyRnNhRzl6ZERBZEJnTlZIU1VFRmpBVQpCZ2dyQmdFRkJRY0RBUVlJS3dZQkJRVUhBd0l3REFZRFZSMFRBUUgvQkFJd0FEQWRCZ05WSFE0RUZnUVVUYU5wClFaVmZtN1RCSG5lTkM3dXhjS0ZsZWhvd0h3WURWUjBqQkJnd0ZvQVU0eEhxUEE0QWZEcVFQRHdHcTlKQ3FtSjkKWGdvd0RRWUpLb1pJaHZjTkFRRUxCUUFEZ2dFQkFBOWxyb0xGbkNtcEdGUjRsZDdBdHoyd08zRXA4eDB1U0VLcQpONVZYOCt4OERacCtJVnhpOG9SNlhNSXUyTkJpRk1qMmtrM242ZzBDVml6Y3M3MWFyc3l0MjVvcElqWFdpVGtRCjJUSFhuQVJ2MUZ2RFJ4TVRJNDlITHVORTFDalVOZmpLeXR4VTkyd3ovbUdydERKai9xTjAyVytCV0ZUdWVyY3oKVklmMW5yOFdSaGZqTVE4SzZ1VVZ3MVFnN1ROYWlaakR5ZkFmZndqSkV2bklHMHo2Nm5GbkcxYXVWTCtBNnoyNwpRUUx6M1ZWY1U4dmdML3VrL3Iwczg0T1MzUHY2UnhVZUVpbnI4enpGOHlWQURXMVRudWtTUThkaWlld0NKM1BvCjh1SUoyRmpCWEh5U2RyMUFMN0FMNWc3N051M2JwOVMzVmUzREErbzBveElhaFFUSHYyYz0KLS0tLS1FTkQgQ0VSVElGSUNBVEUtLS0tLQo=", "encoding": "base64", "item": [ "cert", { "ca": "self-sign", "dns": [ "localhost" ], "name": "quadlet_demo" } ], "source": "/etc/pki/tls/certs/quadlet_demo.crt" } ok: [sut] => (item=['key', {'name': 'quadlet_demo', 'dns': ['localhost'], 'ca': 'self-sign'}]) => { "ansible_loop_var": "item", "changed": false, "content": "LS0tLS1CRUdJTiBQUklWQVRFIEtFWS0tLS0tCk1JSUV2d0lCQURBTkJna3Foa2lHOXcwQkFRRUZBQVNDQktrd2dnU2xBZ0VBQW9JQkFRQy9SZGhoNlZFVjVhdzAKWEJQRWhqaGdibG1yaG1MMUNYMEZYV2hvRHBBQ25XOExtY3F1c2tkM25abzVqc1dmR25vZm80MGdnc0gyMUF6cAp6YXY0SzRxYkVVaEhMZkVONnFlMThlYXlLcisrOENOZHlrU2Ixckl1ejhJVXlzRTduM0ttVVY2b2xnZUpxdzB6Ckk4OVYvUjdGeDAxbnU4bCs3cnR1MWRCN25JNFVKTDA5cGhabEZYdE5wdi82b2tuNGllOEd5VEJ1WUhBODVRR28KOXFCQXBuTXZ0cDdZVUEweVpnV1lWTGVSSGVrUFBZSFRSdHhYTzdlNE13MTU4UGdmRkppZ3dyc0h0TEF0dEhGWgp2TFdpV1ZKSkdFS2ducmtPSEJKOVB4LzZoaGtXeGRpZVpBYWJKcklHcloxZU5UQWVIcjNiUkFOTUpFZjlPQzl4ClB2bVQ1OUVsQWdNQkFBRUNnZ0VCQUlieDR3aXBFSEVQZWQveSttdFRlcURtR2ROKzI0QWZjcG4vWGJMVVcxdk0KOE5SUmE0SzIrdDcwVFNQQWk4S0ZicFcxME5LWURmcHpmcXpSVXp1aWxJWjB1ZmFsL2dTZ2dqUkdkMGxLQ2IvUApNbFJ3YU5rOXJyYUNFMTZSOGFSOENPbzRURDFLcEFUSHJtZGZJMTk2Vm1TWkNRNWZtWGc1Ty9tbFZoSnYwSGw2CmU4R1pCaEtBVldMYmRLR0lnRldlL2ZScEVRSi9uTlBPNTd2NnI1ZEU1a1owSDZSS1o1WjRZWWY2cFpMWlozd20KWUdkSUswUmhLek1nT3NwZzR1VVVXWjRaY0lJRE5uZjArcTlGK2oyamNHUWxxUFBqaHdUb1RjbEJhUDNaOTFrSQpVaUdhN0gxWFNmLzlaK05NQVhtd2ljcHFOa1NkWjJrN3dwRk1yY2ZMRm4wQ2dZRUErSnBpOGg3TnNRTTRkQVROCndpQmVjOSs2Tk01WlZRMXk1NW9wRU1xL240U05mMEFGbWNzOFFybDkraWRtT09kb2hiUDA5aklNYXFib01laDYKSVpwQnJOYzEzVitkTGF4SHJ3cVFPeUFsN2FVTGpNSG9hd05BUi9SMjZDSDlhd3lVTGE3RFk2d0MzbXk0dmFSKwpqbXJ4TjIxQk9nZ1VQS0JGK1RGWElqNklVOGNDZ1lFQXhQYkYvZjJ1VzZmUWhYUm95Q1daUW1lZ2x3eElnbU1PCkFEMnB4M1Y2YU95OWxBU2t4QmNDTlFOTmNCNDFVRlE1MXBHZ2xWczUyemNGeUVxb1FjZ0c0OTRLam5PazhTQ3kKUk1UWXo0M216R0pVTHRyT1hMN1dod2NpUS9tcjFqUDI4MEJOWkV2TU5rcVhZemZORXJkdmV5YlpCN3pBMTRUOQorUk1MeGhuRzI3TUNnWUVBbTBISDQ1QUxRNjVRTmx5OXA2d1JjT3BiWHhrenZ5TmNESHBCaFVRcDhsU1ZEZENNCjd1UjNLUEN4VldYL3VEQUZIZU1xYkFzTHFCdSt6VlJ6MmNKOW01VDRCZDN0eXBWYVcxNE1mM0J2WDRBQXU1SnEKTlZMZGRWS2VaQkMrTkhteG5Pb1ZDbzFsYVVXYkJvelJvTjMyOFhlaEViM3o3WHptMWw3VEhRYlNtd2NDZ1lBRApuWkFHMkloQjBIUEliMW1aQ3Ura1EyaTBQUnA3b2p2WWowazFpeGVESTQxYlVUYUJEdmNRYUE0YzE4RWN4MThFCmJFbnBVL0pjN3NhaFpaVnZGNi9nc091NnI2N0FhbHc3WE8ycDZhM3A1SWd0RUFFbnZLME5iM2tXakZlMkorQUoKbVEyU29MMlA1WkNmMXA5aUQ1SHhUY1pBR1J2bEdPRGNtSmZudkFObGR3S0JnUUR2Wi90WWtSd0JESkwxaG9sNQpRSmNFL29tNG5GbGZNclZ2UElFYnZGZU1ESHJXTkdWTDh1cy9VWGxEVG96Ym5JaDBxaXNZbkJYSmdzbXhKaSs2Cnpoc25kK0Q3TFd6ak0rL3F4bHkxSllrM1A1RFB2WFcvUkFiYnZsUUpPWGNJUzI0eFowLzlsZU9xTDA5RENhdDEKbE8vdTFJV2VkN2NGMFFQY2dlUFlVc0NBYlE9PQotLS0tLUVORCBQUklWQVRFIEtFWS0tLS0tCg==", "encoding": "base64", "item": [ "key", { "ca": "self-sign", "dns": [ "localhost" ], "name": "quadlet_demo" } ], "source": "/etc/pki/tls/private/quadlet_demo.key" } ok: [sut] => (item=['ca', {'name': 'quadlet_demo', 'dns': ['localhost'], 'ca': 'self-sign'}]) => { "ansible_loop_var": "item", "changed": false, "content": "LS0tLS1CRUdJTiBDRVJUSUZJQ0FURS0tLS0tCk1JSURnakNDQW1xZ0F3SUJBZ0lRQ3Q0aW1GYjFUR3VlcmJwOFlFeHo2REFOQmdrcWhraUc5dzBCQVFzRkFEQlEKTVNBd0hnWURWUVFEREJkTWIyTmhiQ0JUYVdkdWFXNW5JRUYxZEdodmNtbDBlVEVzTUNvR0ExVUVBd3dqTUdGawpaVEl5T1RndE5UWm1OVFJqTm1JdE9XVmhaR0poTjJNdE5qQTBZemN6WlRjd0hoY05NalF3TWpBMU1EUXpOVFE0CldoY05NalV3TWpBMU1EUXpOVFE0V2pBVU1SSXdFQVlEVlFRREV3bHNiMk5oYkdodmMzUXdnZ0VpTUEwR0NTcUcKU0liM0RRRUJBUVVBQTRJQkR3QXdnZ0VLQW9JQkFRQy9SZGhoNlZFVjVhdzBYQlBFaGpoZ2JsbXJobUwxQ1gwRgpYV2hvRHBBQ25XOExtY3F1c2tkM25abzVqc1dmR25vZm80MGdnc0gyMUF6cHphdjRLNHFiRVVoSExmRU42cWUxCjhlYXlLcisrOENOZHlrU2Ixckl1ejhJVXlzRTduM0ttVVY2b2xnZUpxdzB6STg5Vi9SN0Z4MDFudThsKzdydHUKMWRCN25JNFVKTDA5cGhabEZYdE5wdi82b2tuNGllOEd5VEJ1WUhBODVRR285cUJBcG5NdnRwN1lVQTB5WmdXWQpWTGVSSGVrUFBZSFRSdHhYTzdlNE13MTU4UGdmRkppZ3dyc0h0TEF0dEhGWnZMV2lXVkpKR0VLZ25ya09IQko5ClB4LzZoaGtXeGRpZVpBYWJKcklHcloxZU5UQWVIcjNiUkFOTUpFZjlPQzl4UHZtVDU5RWxBZ01CQUFHamdaTXcKZ1pBd0N3WURWUjBQQkFRREFnV2dNQlFHQTFVZEVRUU5NQXVDQ1d4dlkyRnNhRzl6ZERBZEJnTlZIU1VFRmpBVQpCZ2dyQmdFRkJRY0RBUVlJS3dZQkJRVUhBd0l3REFZRFZSMFRBUUgvQkFJd0FEQWRCZ05WSFE0RUZnUVVUYU5wClFaVmZtN1RCSG5lTkM3dXhjS0ZsZWhvd0h3WURWUjBqQkJnd0ZvQVU0eEhxUEE0QWZEcVFQRHdHcTlKQ3FtSjkKWGdvd0RRWUpLb1pJaHZjTkFRRUxCUUFEZ2dFQkFBOWxyb0xGbkNtcEdGUjRsZDdBdHoyd08zRXA4eDB1U0VLcQpONVZYOCt4OERacCtJVnhpOG9SNlhNSXUyTkJpRk1qMmtrM242ZzBDVml6Y3M3MWFyc3l0MjVvcElqWFdpVGtRCjJUSFhuQVJ2MUZ2RFJ4TVRJNDlITHVORTFDalVOZmpLeXR4VTkyd3ovbUdydERKai9xTjAyVytCV0ZUdWVyY3oKVklmMW5yOFdSaGZqTVE4SzZ1VVZ3MVFnN1ROYWlaakR5ZkFmZndqSkV2bklHMHo2Nm5GbkcxYXVWTCtBNnoyNwpRUUx6M1ZWY1U4dmdML3VrL3Iwczg0T1MzUHY2UnhVZUVpbnI4enpGOHlWQURXMVRudWtTUThkaWlld0NKM1BvCjh1SUoyRmpCWEh5U2RyMUFMN0FMNWc3N051M2JwOVMzVmUzREErbzBveElhaFFUSHYyYz0KLS0tLS1FTkQgQ0VSVElGSUNBVEUtLS0tLQo=", "encoding": "base64", "item": [ "ca", { "ca": "self-sign", "dns": [ "localhost" ], "name": "quadlet_demo" } ], "source": "/etc/pki/tls/certs/quadlet_demo.crt" } TASK [fedora.linux_system_roles.certificate : Create return data] ************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:160 Monday 05 February 2024 04:35:49 +0000 (0:00:00.601) 0:00:11.324 ******* ok: [sut] => { "ansible_facts": { "certificate_test_certs": { "quadlet_demo": { "ca": "/etc/pki/tls/certs/quadlet_demo.crt", "ca_content": "-----BEGIN CERTIFICATE-----\nMIIDgjCCAmqgAwIBAgIQCt4imFb1TGuerbp8YExz6DANBgkqhkiG9w0BAQsFADBQ\nMSAwHgYDVQQDDBdMb2NhbCBTaWduaW5nIEF1dGhvcml0eTEsMCoGA1UEAwwjMGFk\nZTIyOTgtNTZmNTRjNmItOWVhZGJhN2MtNjA0YzczZTcwHhcNMjQwMjA1MDQzNTQ4\nWhcNMjUwMjA1MDQzNTQ4WjAUMRIwEAYDVQQDEwlsb2NhbGhvc3QwggEiMA0GCSqG\nSIb3DQEBAQUAA4IBDwAwggEKAoIBAQC/Rdhh6VEV5aw0XBPEhjhgblmrhmL1CX0F\nXWhoDpACnW8Lmcquskd3nZo5jsWfGnofo40ggsH21Azpzav4K4qbEUhHLfEN6qe1\n8eayKr++8CNdykSb1rIuz8IUysE7n3KmUV6olgeJqw0zI89V/R7Fx01nu8l+7rtu\n1dB7nI4UJL09phZlFXtNpv/6okn4ie8GyTBuYHA85QGo9qBApnMvtp7YUA0yZgWY\nVLeRHekPPYHTRtxXO7e4Mw158PgfFJigwrsHtLAttHFZvLWiWVJJGEKgnrkOHBJ9\nPx/6hhkWxdieZAabJrIGrZ1eNTAeHr3bRANMJEf9OC9xPvmT59ElAgMBAAGjgZMw\ngZAwCwYDVR0PBAQDAgWgMBQGA1UdEQQNMAuCCWxvY2FsaG9zdDAdBgNVHSUEFjAU\nBggrBgEFBQcDAQYIKwYBBQUHAwIwDAYDVR0TAQH/BAIwADAdBgNVHQ4EFgQUTaNp\nQZVfm7TBHneNC7uxcKFlehowHwYDVR0jBBgwFoAU4xHqPA4AfDqQPDwGq9JCqmJ9\nXgowDQYJKoZIhvcNAQELBQADggEBAA9lroLFnCmpGFR4ld7Atz2wO3Ep8x0uSEKq\nN5VX8+x8DZp+IVxi8oR6XMIu2NBiFMj2kk3n6g0CVizcs71arsyt25opIjXWiTkQ\n2THXnARv1FvDRxMTI49HLuNE1CjUNfjKytxU92wz/mGrtDJj/qN02W+BWFTuercz\nVIf1nr8WRhfjMQ8K6uUVw1Qg7TNaiZjDyfAffwjJEvnIG0z66nFnG1auVL+A6z27\nQQLz3VVcU8vgL/uk/r0s84OS3Pv6RxUeEinr8zzF8yVADW1TnukSQ8diiewCJ3Po\n8uIJ2FjBXHySdr1AL7AL5g77Nu3bp9S3Ve3DA+o0oxIahQTHv2c=\n-----END CERTIFICATE-----\n", "cert": "/etc/pki/tls/certs/quadlet_demo.crt", "cert_content": "-----BEGIN CERTIFICATE-----\nMIIDgjCCAmqgAwIBAgIQCt4imFb1TGuerbp8YExz6DANBgkqhkiG9w0BAQsFADBQ\nMSAwHgYDVQQDDBdMb2NhbCBTaWduaW5nIEF1dGhvcml0eTEsMCoGA1UEAwwjMGFk\nZTIyOTgtNTZmNTRjNmItOWVhZGJhN2MtNjA0YzczZTcwHhcNMjQwMjA1MDQzNTQ4\nWhcNMjUwMjA1MDQzNTQ4WjAUMRIwEAYDVQQDEwlsb2NhbGhvc3QwggEiMA0GCSqG\nSIb3DQEBAQUAA4IBDwAwggEKAoIBAQC/Rdhh6VEV5aw0XBPEhjhgblmrhmL1CX0F\nXWhoDpACnW8Lmcquskd3nZo5jsWfGnofo40ggsH21Azpzav4K4qbEUhHLfEN6qe1\n8eayKr++8CNdykSb1rIuz8IUysE7n3KmUV6olgeJqw0zI89V/R7Fx01nu8l+7rtu\n1dB7nI4UJL09phZlFXtNpv/6okn4ie8GyTBuYHA85QGo9qBApnMvtp7YUA0yZgWY\nVLeRHekPPYHTRtxXO7e4Mw158PgfFJigwrsHtLAttHFZvLWiWVJJGEKgnrkOHBJ9\nPx/6hhkWxdieZAabJrIGrZ1eNTAeHr3bRANMJEf9OC9xPvmT59ElAgMBAAGjgZMw\ngZAwCwYDVR0PBAQDAgWgMBQGA1UdEQQNMAuCCWxvY2FsaG9zdDAdBgNVHSUEFjAU\nBggrBgEFBQcDAQYIKwYBBQUHAwIwDAYDVR0TAQH/BAIwADAdBgNVHQ4EFgQUTaNp\nQZVfm7TBHneNC7uxcKFlehowHwYDVR0jBBgwFoAU4xHqPA4AfDqQPDwGq9JCqmJ9\nXgowDQYJKoZIhvcNAQELBQADggEBAA9lroLFnCmpGFR4ld7Atz2wO3Ep8x0uSEKq\nN5VX8+x8DZp+IVxi8oR6XMIu2NBiFMj2kk3n6g0CVizcs71arsyt25opIjXWiTkQ\n2THXnARv1FvDRxMTI49HLuNE1CjUNfjKytxU92wz/mGrtDJj/qN02W+BWFTuercz\nVIf1nr8WRhfjMQ8K6uUVw1Qg7TNaiZjDyfAffwjJEvnIG0z66nFnG1auVL+A6z27\nQQLz3VVcU8vgL/uk/r0s84OS3Pv6RxUeEinr8zzF8yVADW1TnukSQ8diiewCJ3Po\n8uIJ2FjBXHySdr1AL7AL5g77Nu3bp9S3Ve3DA+o0oxIahQTHv2c=\n-----END CERTIFICATE-----\n", "key": "/etc/pki/tls/private/quadlet_demo.key", "key_content": "-----BEGIN PRIVATE KEY-----\nMIIEvwIBADANBgkqhkiG9w0BAQEFAASCBKkwggSlAgEAAoIBAQC/Rdhh6VEV5aw0\nXBPEhjhgblmrhmL1CX0FXWhoDpACnW8Lmcquskd3nZo5jsWfGnofo40ggsH21Azp\nzav4K4qbEUhHLfEN6qe18eayKr++8CNdykSb1rIuz8IUysE7n3KmUV6olgeJqw0z\nI89V/R7Fx01nu8l+7rtu1dB7nI4UJL09phZlFXtNpv/6okn4ie8GyTBuYHA85QGo\n9qBApnMvtp7YUA0yZgWYVLeRHekPPYHTRtxXO7e4Mw158PgfFJigwrsHtLAttHFZ\nvLWiWVJJGEKgnrkOHBJ9Px/6hhkWxdieZAabJrIGrZ1eNTAeHr3bRANMJEf9OC9x\nPvmT59ElAgMBAAECggEBAIbx4wipEHEPed/y+mtTeqDmGdN+24Afcpn/XbLUW1vM\n8NRRa4K2+t70TSPAi8KFbpW10NKYDfpzfqzRUzuilIZ0ufal/gSggjRGd0lKCb/P\nMlRwaNk9rraCE16R8aR8COo4TD1KpATHrmdfI196VmSZCQ5fmXg5O/mlVhJv0Hl6\ne8GZBhKAVWLbdKGIgFWe/fRpEQJ/nNPO57v6r5dE5kZ0H6RKZ5Z4YYf6pZLZZ3wm\nYGdIK0RhKzMgOspg4uUUWZ4ZcIIDNnf0+q9F+j2jcGQlqPPjhwToTclBaP3Z91kI\nUiGa7H1XSf/9Z+NMAXmwicpqNkSdZ2k7wpFMrcfLFn0CgYEA+Jpi8h7NsQM4dATN\nwiBec9+6NM5ZVQ1y55opEMq/n4SNf0AFmcs8Qrl9+idmOOdohbP09jIMaqboMeh6\nIZpBrNc13V+dLaxHrwqQOyAl7aULjMHoawNAR/R26CH9awyULa7DY6wC3my4vaR+\njmrxN21BOggUPKBF+TFXIj6IU8cCgYEAxPbF/f2uW6fQhXRoyCWZQmeglwxIgmMO\nAD2px3V6aOy9lASkxBcCNQNNcB41UFQ51pGglVs52zcFyEqoQcgG494KjnOk8SCy\nRMTYz43mzGJULtrOXL7WhwciQ/mr1jP280BNZEvMNkqXYzfNErdveybZB7zA14T9\n+RMLxhnG27MCgYEAm0HH45ALQ65QNly9p6wRcOpbXxkzvyNcDHpBhUQp8lSVDdCM\n7uR3KPCxVWX/uDAFHeMqbAsLqBu+zVRz2cJ9m5T4Bd3typVaW14Mf3BvX4AAu5Jq\nNVLddVKeZBC+NHmxnOoVCo1laUWbBozRoN328XehEb3z7Xzm1l7THQbSmwcCgYAD\nnZAG2IhB0HPIb1mZCu+kQ2i0PRp7ojvYj0k1ixeDI41bUTaBDvcQaA4c18Ecx18E\nbEnpU/Jc7sahZZVvF6/gsOu6r67Aalw7XO2p6a3p5IgtEAEnvK0Nb3kWjFe2J+AJ\nmQ2SoL2P5ZCf1p9iD5HxTcZAGRvlGODcmJfnvANldwKBgQDvZ/tYkRwBDJL1hol5\nQJcE/om4nFlfMrVvPIEbvFeMDHrWNGVL8us/UXlDTozbnIh0qisYnBXJgsmxJi+6\nzhsnd+D7LWzjM+/qxly1JYk3P5DPvXW/RAbbvlQJOXcIS24xZ0/9leOqL09DCat1\nlO/u1IWed7cF0QPcgePYUsCAbQ==\n-----END PRIVATE KEY-----\n" } } }, "changed": false } TASK [fedora.linux_system_roles.certificate : Stop tracking certificates] ****** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:176 Monday 05 February 2024 04:35:49 +0000 (0:00:00.022) 0:00:11.347 ******* ok: [sut] => (item={'cert': '/etc/pki/tls/certs/quadlet_demo.crt', 'cert_content': '-----BEGIN CERTIFICATE-----\nMIIDgjCCAmqgAwIBAgIQCt4imFb1TGuerbp8YExz6DANBgkqhkiG9w0BAQsFADBQ\nMSAwHgYDVQQDDBdMb2NhbCBTaWduaW5nIEF1dGhvcml0eTEsMCoGA1UEAwwjMGFk\nZTIyOTgtNTZmNTRjNmItOWVhZGJhN2MtNjA0YzczZTcwHhcNMjQwMjA1MDQzNTQ4\nWhcNMjUwMjA1MDQzNTQ4WjAUMRIwEAYDVQQDEwlsb2NhbGhvc3QwggEiMA0GCSqG\nSIb3DQEBAQUAA4IBDwAwggEKAoIBAQC/Rdhh6VEV5aw0XBPEhjhgblmrhmL1CX0F\nXWhoDpACnW8Lmcquskd3nZo5jsWfGnofo40ggsH21Azpzav4K4qbEUhHLfEN6qe1\n8eayKr++8CNdykSb1rIuz8IUysE7n3KmUV6olgeJqw0zI89V/R7Fx01nu8l+7rtu\n1dB7nI4UJL09phZlFXtNpv/6okn4ie8GyTBuYHA85QGo9qBApnMvtp7YUA0yZgWY\nVLeRHekPPYHTRtxXO7e4Mw158PgfFJigwrsHtLAttHFZvLWiWVJJGEKgnrkOHBJ9\nPx/6hhkWxdieZAabJrIGrZ1eNTAeHr3bRANMJEf9OC9xPvmT59ElAgMBAAGjgZMw\ngZAwCwYDVR0PBAQDAgWgMBQGA1UdEQQNMAuCCWxvY2FsaG9zdDAdBgNVHSUEFjAU\nBggrBgEFBQcDAQYIKwYBBQUHAwIwDAYDVR0TAQH/BAIwADAdBgNVHQ4EFgQUTaNp\nQZVfm7TBHneNC7uxcKFlehowHwYDVR0jBBgwFoAU4xHqPA4AfDqQPDwGq9JCqmJ9\nXgowDQYJKoZIhvcNAQELBQADggEBAA9lroLFnCmpGFR4ld7Atz2wO3Ep8x0uSEKq\nN5VX8+x8DZp+IVxi8oR6XMIu2NBiFMj2kk3n6g0CVizcs71arsyt25opIjXWiTkQ\n2THXnARv1FvDRxMTI49HLuNE1CjUNfjKytxU92wz/mGrtDJj/qN02W+BWFTuercz\nVIf1nr8WRhfjMQ8K6uUVw1Qg7TNaiZjDyfAffwjJEvnIG0z66nFnG1auVL+A6z27\nQQLz3VVcU8vgL/uk/r0s84OS3Pv6RxUeEinr8zzF8yVADW1TnukSQ8diiewCJ3Po\n8uIJ2FjBXHySdr1AL7AL5g77Nu3bp9S3Ve3DA+o0oxIahQTHv2c=\n-----END CERTIFICATE-----\n', 'key': '/etc/pki/tls/private/quadlet_demo.key', 'key_content': '-----BEGIN PRIVATE KEY-----\nMIIEvwIBADANBgkqhkiG9w0BAQEFAASCBKkwggSlAgEAAoIBAQC/Rdhh6VEV5aw0\nXBPEhjhgblmrhmL1CX0FXWhoDpACnW8Lmcquskd3nZo5jsWfGnofo40ggsH21Azp\nzav4K4qbEUhHLfEN6qe18eayKr++8CNdykSb1rIuz8IUysE7n3KmUV6olgeJqw0z\nI89V/R7Fx01nu8l+7rtu1dB7nI4UJL09phZlFXtNpv/6okn4ie8GyTBuYHA85QGo\n9qBApnMvtp7YUA0yZgWYVLeRHekPPYHTRtxXO7e4Mw158PgfFJigwrsHtLAttHFZ\nvLWiWVJJGEKgnrkOHBJ9Px/6hhkWxdieZAabJrIGrZ1eNTAeHr3bRANMJEf9OC9x\nPvmT59ElAgMBAAECggEBAIbx4wipEHEPed/y+mtTeqDmGdN+24Afcpn/XbLUW1vM\n8NRRa4K2+t70TSPAi8KFbpW10NKYDfpzfqzRUzuilIZ0ufal/gSggjRGd0lKCb/P\nMlRwaNk9rraCE16R8aR8COo4TD1KpATHrmdfI196VmSZCQ5fmXg5O/mlVhJv0Hl6\ne8GZBhKAVWLbdKGIgFWe/fRpEQJ/nNPO57v6r5dE5kZ0H6RKZ5Z4YYf6pZLZZ3wm\nYGdIK0RhKzMgOspg4uUUWZ4ZcIIDNnf0+q9F+j2jcGQlqPPjhwToTclBaP3Z91kI\nUiGa7H1XSf/9Z+NMAXmwicpqNkSdZ2k7wpFMrcfLFn0CgYEA+Jpi8h7NsQM4dATN\nwiBec9+6NM5ZVQ1y55opEMq/n4SNf0AFmcs8Qrl9+idmOOdohbP09jIMaqboMeh6\nIZpBrNc13V+dLaxHrwqQOyAl7aULjMHoawNAR/R26CH9awyULa7DY6wC3my4vaR+\njmrxN21BOggUPKBF+TFXIj6IU8cCgYEAxPbF/f2uW6fQhXRoyCWZQmeglwxIgmMO\nAD2px3V6aOy9lASkxBcCNQNNcB41UFQ51pGglVs52zcFyEqoQcgG494KjnOk8SCy\nRMTYz43mzGJULtrOXL7WhwciQ/mr1jP280BNZEvMNkqXYzfNErdveybZB7zA14T9\n+RMLxhnG27MCgYEAm0HH45ALQ65QNly9p6wRcOpbXxkzvyNcDHpBhUQp8lSVDdCM\n7uR3KPCxVWX/uDAFHeMqbAsLqBu+zVRz2cJ9m5T4Bd3typVaW14Mf3BvX4AAu5Jq\nNVLddVKeZBC+NHmxnOoVCo1laUWbBozRoN328XehEb3z7Xzm1l7THQbSmwcCgYAD\nnZAG2IhB0HPIb1mZCu+kQ2i0PRp7ojvYj0k1ixeDI41bUTaBDvcQaA4c18Ecx18E\nbEnpU/Jc7sahZZVvF6/gsOu6r67Aalw7XO2p6a3p5IgtEAEnvK0Nb3kWjFe2J+AJ\nmQ2SoL2P5ZCf1p9iD5HxTcZAGRvlGODcmJfnvANldwKBgQDvZ/tYkRwBDJL1hol5\nQJcE/om4nFlfMrVvPIEbvFeMDHrWNGVL8us/UXlDTozbnIh0qisYnBXJgsmxJi+6\nzhsnd+D7LWzjM+/qxly1JYk3P5DPvXW/RAbbvlQJOXcIS24xZ0/9leOqL09DCat1\nlO/u1IWed7cF0QPcgePYUsCAbQ==\n-----END PRIVATE KEY-----\n', 'ca': '/etc/pki/tls/certs/quadlet_demo.crt', 'ca_content': '-----BEGIN CERTIFICATE-----\nMIIDgjCCAmqgAwIBAgIQCt4imFb1TGuerbp8YExz6DANBgkqhkiG9w0BAQsFADBQ\nMSAwHgYDVQQDDBdMb2NhbCBTaWduaW5nIEF1dGhvcml0eTEsMCoGA1UEAwwjMGFk\nZTIyOTgtNTZmNTRjNmItOWVhZGJhN2MtNjA0YzczZTcwHhcNMjQwMjA1MDQzNTQ4\nWhcNMjUwMjA1MDQzNTQ4WjAUMRIwEAYDVQQDEwlsb2NhbGhvc3QwggEiMA0GCSqG\nSIb3DQEBAQUAA4IBDwAwggEKAoIBAQC/Rdhh6VEV5aw0XBPEhjhgblmrhmL1CX0F\nXWhoDpACnW8Lmcquskd3nZo5jsWfGnofo40ggsH21Azpzav4K4qbEUhHLfEN6qe1\n8eayKr++8CNdykSb1rIuz8IUysE7n3KmUV6olgeJqw0zI89V/R7Fx01nu8l+7rtu\n1dB7nI4UJL09phZlFXtNpv/6okn4ie8GyTBuYHA85QGo9qBApnMvtp7YUA0yZgWY\nVLeRHekPPYHTRtxXO7e4Mw158PgfFJigwrsHtLAttHFZvLWiWVJJGEKgnrkOHBJ9\nPx/6hhkWxdieZAabJrIGrZ1eNTAeHr3bRANMJEf9OC9xPvmT59ElAgMBAAGjgZMw\ngZAwCwYDVR0PBAQDAgWgMBQGA1UdEQQNMAuCCWxvY2FsaG9zdDAdBgNVHSUEFjAU\nBggrBgEFBQcDAQYIKwYBBQUHAwIwDAYDVR0TAQH/BAIwADAdBgNVHQ4EFgQUTaNp\nQZVfm7TBHneNC7uxcKFlehowHwYDVR0jBBgwFoAU4xHqPA4AfDqQPDwGq9JCqmJ9\nXgowDQYJKoZIhvcNAQELBQADggEBAA9lroLFnCmpGFR4ld7Atz2wO3Ep8x0uSEKq\nN5VX8+x8DZp+IVxi8oR6XMIu2NBiFMj2kk3n6g0CVizcs71arsyt25opIjXWiTkQ\n2THXnARv1FvDRxMTI49HLuNE1CjUNfjKytxU92wz/mGrtDJj/qN02W+BWFTuercz\nVIf1nr8WRhfjMQ8K6uUVw1Qg7TNaiZjDyfAffwjJEvnIG0z66nFnG1auVL+A6z27\nQQLz3VVcU8vgL/uk/r0s84OS3Pv6RxUeEinr8zzF8yVADW1TnukSQ8diiewCJ3Po\n8uIJ2FjBXHySdr1AL7AL5g77Nu3bp9S3Ve3DA+o0oxIahQTHv2c=\n-----END CERTIFICATE-----\n'}) => { "ansible_loop_var": "item", "changed": false, "cmd": [ "getcert", "stop-tracking", "-f", "/etc/pki/tls/certs/quadlet_demo.crt" ], "delta": "0:00:00.034662", "end": "2024-02-05 04:35:49.516605", "item": { "ca": "/etc/pki/tls/certs/quadlet_demo.crt", "ca_content": "-----BEGIN CERTIFICATE-----\nMIIDgjCCAmqgAwIBAgIQCt4imFb1TGuerbp8YExz6DANBgkqhkiG9w0BAQsFADBQ\nMSAwHgYDVQQDDBdMb2NhbCBTaWduaW5nIEF1dGhvcml0eTEsMCoGA1UEAwwjMGFk\nZTIyOTgtNTZmNTRjNmItOWVhZGJhN2MtNjA0YzczZTcwHhcNMjQwMjA1MDQzNTQ4\nWhcNMjUwMjA1MDQzNTQ4WjAUMRIwEAYDVQQDEwlsb2NhbGhvc3QwggEiMA0GCSqG\nSIb3DQEBAQUAA4IBDwAwggEKAoIBAQC/Rdhh6VEV5aw0XBPEhjhgblmrhmL1CX0F\nXWhoDpACnW8Lmcquskd3nZo5jsWfGnofo40ggsH21Azpzav4K4qbEUhHLfEN6qe1\n8eayKr++8CNdykSb1rIuz8IUysE7n3KmUV6olgeJqw0zI89V/R7Fx01nu8l+7rtu\n1dB7nI4UJL09phZlFXtNpv/6okn4ie8GyTBuYHA85QGo9qBApnMvtp7YUA0yZgWY\nVLeRHekPPYHTRtxXO7e4Mw158PgfFJigwrsHtLAttHFZvLWiWVJJGEKgnrkOHBJ9\nPx/6hhkWxdieZAabJrIGrZ1eNTAeHr3bRANMJEf9OC9xPvmT59ElAgMBAAGjgZMw\ngZAwCwYDVR0PBAQDAgWgMBQGA1UdEQQNMAuCCWxvY2FsaG9zdDAdBgNVHSUEFjAU\nBggrBgEFBQcDAQYIKwYBBQUHAwIwDAYDVR0TAQH/BAIwADAdBgNVHQ4EFgQUTaNp\nQZVfm7TBHneNC7uxcKFlehowHwYDVR0jBBgwFoAU4xHqPA4AfDqQPDwGq9JCqmJ9\nXgowDQYJKoZIhvcNAQELBQADggEBAA9lroLFnCmpGFR4ld7Atz2wO3Ep8x0uSEKq\nN5VX8+x8DZp+IVxi8oR6XMIu2NBiFMj2kk3n6g0CVizcs71arsyt25opIjXWiTkQ\n2THXnARv1FvDRxMTI49HLuNE1CjUNfjKytxU92wz/mGrtDJj/qN02W+BWFTuercz\nVIf1nr8WRhfjMQ8K6uUVw1Qg7TNaiZjDyfAffwjJEvnIG0z66nFnG1auVL+A6z27\nQQLz3VVcU8vgL/uk/r0s84OS3Pv6RxUeEinr8zzF8yVADW1TnukSQ8diiewCJ3Po\n8uIJ2FjBXHySdr1AL7AL5g77Nu3bp9S3Ve3DA+o0oxIahQTHv2c=\n-----END CERTIFICATE-----\n", "cert": "/etc/pki/tls/certs/quadlet_demo.crt", "cert_content": "-----BEGIN CERTIFICATE-----\nMIIDgjCCAmqgAwIBAgIQCt4imFb1TGuerbp8YExz6DANBgkqhkiG9w0BAQsFADBQ\nMSAwHgYDVQQDDBdMb2NhbCBTaWduaW5nIEF1dGhvcml0eTEsMCoGA1UEAwwjMGFk\nZTIyOTgtNTZmNTRjNmItOWVhZGJhN2MtNjA0YzczZTcwHhcNMjQwMjA1MDQzNTQ4\nWhcNMjUwMjA1MDQzNTQ4WjAUMRIwEAYDVQQDEwlsb2NhbGhvc3QwggEiMA0GCSqG\nSIb3DQEBAQUAA4IBDwAwggEKAoIBAQC/Rdhh6VEV5aw0XBPEhjhgblmrhmL1CX0F\nXWhoDpACnW8Lmcquskd3nZo5jsWfGnofo40ggsH21Azpzav4K4qbEUhHLfEN6qe1\n8eayKr++8CNdykSb1rIuz8IUysE7n3KmUV6olgeJqw0zI89V/R7Fx01nu8l+7rtu\n1dB7nI4UJL09phZlFXtNpv/6okn4ie8GyTBuYHA85QGo9qBApnMvtp7YUA0yZgWY\nVLeRHekPPYHTRtxXO7e4Mw158PgfFJigwrsHtLAttHFZvLWiWVJJGEKgnrkOHBJ9\nPx/6hhkWxdieZAabJrIGrZ1eNTAeHr3bRANMJEf9OC9xPvmT59ElAgMBAAGjgZMw\ngZAwCwYDVR0PBAQDAgWgMBQGA1UdEQQNMAuCCWxvY2FsaG9zdDAdBgNVHSUEFjAU\nBggrBgEFBQcDAQYIKwYBBQUHAwIwDAYDVR0TAQH/BAIwADAdBgNVHQ4EFgQUTaNp\nQZVfm7TBHneNC7uxcKFlehowHwYDVR0jBBgwFoAU4xHqPA4AfDqQPDwGq9JCqmJ9\nXgowDQYJKoZIhvcNAQELBQADggEBAA9lroLFnCmpGFR4ld7Atz2wO3Ep8x0uSEKq\nN5VX8+x8DZp+IVxi8oR6XMIu2NBiFMj2kk3n6g0CVizcs71arsyt25opIjXWiTkQ\n2THXnARv1FvDRxMTI49HLuNE1CjUNfjKytxU92wz/mGrtDJj/qN02W+BWFTuercz\nVIf1nr8WRhfjMQ8K6uUVw1Qg7TNaiZjDyfAffwjJEvnIG0z66nFnG1auVL+A6z27\nQQLz3VVcU8vgL/uk/r0s84OS3Pv6RxUeEinr8zzF8yVADW1TnukSQ8diiewCJ3Po\n8uIJ2FjBXHySdr1AL7AL5g77Nu3bp9S3Ve3DA+o0oxIahQTHv2c=\n-----END CERTIFICATE-----\n", "key": "/etc/pki/tls/private/quadlet_demo.key", "key_content": "-----BEGIN PRIVATE KEY-----\nMIIEvwIBADANBgkqhkiG9w0BAQEFAASCBKkwggSlAgEAAoIBAQC/Rdhh6VEV5aw0\nXBPEhjhgblmrhmL1CX0FXWhoDpACnW8Lmcquskd3nZo5jsWfGnofo40ggsH21Azp\nzav4K4qbEUhHLfEN6qe18eayKr++8CNdykSb1rIuz8IUysE7n3KmUV6olgeJqw0z\nI89V/R7Fx01nu8l+7rtu1dB7nI4UJL09phZlFXtNpv/6okn4ie8GyTBuYHA85QGo\n9qBApnMvtp7YUA0yZgWYVLeRHekPPYHTRtxXO7e4Mw158PgfFJigwrsHtLAttHFZ\nvLWiWVJJGEKgnrkOHBJ9Px/6hhkWxdieZAabJrIGrZ1eNTAeHr3bRANMJEf9OC9x\nPvmT59ElAgMBAAECggEBAIbx4wipEHEPed/y+mtTeqDmGdN+24Afcpn/XbLUW1vM\n8NRRa4K2+t70TSPAi8KFbpW10NKYDfpzfqzRUzuilIZ0ufal/gSggjRGd0lKCb/P\nMlRwaNk9rraCE16R8aR8COo4TD1KpATHrmdfI196VmSZCQ5fmXg5O/mlVhJv0Hl6\ne8GZBhKAVWLbdKGIgFWe/fRpEQJ/nNPO57v6r5dE5kZ0H6RKZ5Z4YYf6pZLZZ3wm\nYGdIK0RhKzMgOspg4uUUWZ4ZcIIDNnf0+q9F+j2jcGQlqPPjhwToTclBaP3Z91kI\nUiGa7H1XSf/9Z+NMAXmwicpqNkSdZ2k7wpFMrcfLFn0CgYEA+Jpi8h7NsQM4dATN\nwiBec9+6NM5ZVQ1y55opEMq/n4SNf0AFmcs8Qrl9+idmOOdohbP09jIMaqboMeh6\nIZpBrNc13V+dLaxHrwqQOyAl7aULjMHoawNAR/R26CH9awyULa7DY6wC3my4vaR+\njmrxN21BOggUPKBF+TFXIj6IU8cCgYEAxPbF/f2uW6fQhXRoyCWZQmeglwxIgmMO\nAD2px3V6aOy9lASkxBcCNQNNcB41UFQ51pGglVs52zcFyEqoQcgG494KjnOk8SCy\nRMTYz43mzGJULtrOXL7WhwciQ/mr1jP280BNZEvMNkqXYzfNErdveybZB7zA14T9\n+RMLxhnG27MCgYEAm0HH45ALQ65QNly9p6wRcOpbXxkzvyNcDHpBhUQp8lSVDdCM\n7uR3KPCxVWX/uDAFHeMqbAsLqBu+zVRz2cJ9m5T4Bd3typVaW14Mf3BvX4AAu5Jq\nNVLddVKeZBC+NHmxnOoVCo1laUWbBozRoN328XehEb3z7Xzm1l7THQbSmwcCgYAD\nnZAG2IhB0HPIb1mZCu+kQ2i0PRp7ojvYj0k1ixeDI41bUTaBDvcQaA4c18Ecx18E\nbEnpU/Jc7sahZZVvF6/gsOu6r67Aalw7XO2p6a3p5IgtEAEnvK0Nb3kWjFe2J+AJ\nmQ2SoL2P5ZCf1p9iD5HxTcZAGRvlGODcmJfnvANldwKBgQDvZ/tYkRwBDJL1hol5\nQJcE/om4nFlfMrVvPIEbvFeMDHrWNGVL8us/UXlDTozbnIh0qisYnBXJgsmxJi+6\nzhsnd+D7LWzjM+/qxly1JYk3P5DPvXW/RAbbvlQJOXcIS24xZ0/9leOqL09DCat1\nlO/u1IWed7cF0QPcgePYUsCAbQ==\n-----END PRIVATE KEY-----\n" }, "rc": 0, "start": "2024-02-05 04:35:49.481943" } STDOUT: Request "20240205043548" removed. TASK [fedora.linux_system_roles.certificate : Remove files] ******************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:181 Monday 05 February 2024 04:35:49 +0000 (0:00:00.285) 0:00:11.632 ******* changed: [sut] => (item=/etc/pki/tls/certs/quadlet_demo.crt) => { "ansible_loop_var": "item", "changed": true, "item": "/etc/pki/tls/certs/quadlet_demo.crt", "path": "/etc/pki/tls/certs/quadlet_demo.crt", "state": "absent" } changed: [sut] => (item=/etc/pki/tls/private/quadlet_demo.key) => { "ansible_loop_var": "item", "changed": true, "item": "/etc/pki/tls/private/quadlet_demo.key", "path": "/etc/pki/tls/private/quadlet_demo.key", "state": "absent" } ok: [sut] => (item=/etc/pki/tls/certs/quadlet_demo.crt) => { "ansible_loop_var": "item", "changed": false, "item": "/etc/pki/tls/certs/quadlet_demo.crt", "path": "/etc/pki/tls/certs/quadlet_demo.crt", "state": "absent" } TASK [Run the role] ************************************************************ task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/tests/tests_quadlet_demo.yml:50 Monday 05 February 2024 04:35:50 +0000 (0:00:00.571) 0:00:12.204 ******* TASK [fedora.linux_system_roles.podman : Set platform/version specific variables] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:3 Monday 05 February 2024 04:35:50 +0000 (0:00:00.029) 0:00:12.234 ******* included: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml for sut TASK [fedora.linux_system_roles.podman : Ensure ansible_facts used by role] **** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:3 Monday 05 February 2024 04:35:50 +0000 (0:00:00.015) 0:00:12.249 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_required_facts | difference(ansible_facts.keys() | list) | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if system is ostree] ************ task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:11 Monday 05 February 2024 04:35:50 +0000 (0:00:00.017) 0:00:12.266 ******* ok: [sut] => { "changed": false, "stat": { "exists": false } } TASK [fedora.linux_system_roles.podman : Set flag to indicate system is ostree] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:16 Monday 05 February 2024 04:35:50 +0000 (0:00:00.184) 0:00:12.451 ******* ok: [sut] => { "ansible_facts": { "__podman_is_ostree": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Set platform/version specific variables] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:20 Monday 05 February 2024 04:35:50 +0000 (0:00:00.014) 0:00:12.465 ******* ok: [sut] => (item=RedHat.yml) => { "ansible_facts": { "__podman_packages": [ "podman", "shadow-utils-subid" ] }, "ansible_included_var_files": [ "/WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/vars/RedHat.yml" ], "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml" } skipping: [sut] => (item=CentOS.yml) => { "ansible_loop_var": "item", "changed": false, "false_condition": "__vars_file is file", "item": "CentOS.yml", "skip_reason": "Conditional result was False" } ok: [sut] => (item=CentOS_8.yml) => { "ansible_facts": { "__podman_packages": [ "crun", "podman", "podman-plugins", "shadow-utils-subid" ] }, "ansible_included_var_files": [ "/WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/vars/CentOS_8.yml" ], "ansible_loop_var": "item", "changed": false, "item": "CentOS_8.yml" } ok: [sut] => (item=CentOS_8.yml) => { "ansible_facts": { "__podman_packages": [ "crun", "podman", "podman-plugins", "shadow-utils-subid" ] }, "ansible_included_var_files": [ "/WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/vars/CentOS_8.yml" ], "ansible_loop_var": "item", "changed": false, "item": "CentOS_8.yml" } TASK [fedora.linux_system_roles.podman : Gather the package facts] ************* task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:6 Monday 05 February 2024 04:35:50 +0000 (0:00:00.027) 0:00:12.492 ******* ok: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Enable copr if requested] ************* task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:10 Monday 05 February 2024 04:35:51 +0000 (0:00:01.457) 0:00:13.950 ******* skipping: [sut] => { "changed": false, "false_condition": "podman_use_copr | d(false)", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Ensure required packages are installed] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:14 Monday 05 February 2024 04:35:51 +0000 (0:00:00.024) 0:00:13.974 ******* changed: [sut] => { "changed": true, "rc": 0, "results": [ "Installed: fuse-common-3.3.0-17.el8.x86_64", "Installed: container-selinux-2:2.227.0-1.module_el8+804+f131391c.noarch", "Installed: fuse-overlayfs-1.13-1.module_el8+804+f131391c.x86_64", "Installed: yajl-2.1.0-12.el8.x86_64", "Installed: runc-1:1.1.11-1.module_el8+804+f131391c.x86_64", "Installed: shadow-utils-subid-2:4.6-22.el8.x86_64", "Installed: containernetworking-plugins-1:1.4.0-1.module_el8+804+f131391c.x86_64", "Installed: fuse3-3.3.0-17.el8.x86_64", "Installed: podman-gvproxy-3:4.8.3-1.module_el8+804+f131391c.x86_64", "Installed: criu-3.18-4.module_el8+804+f131391c.x86_64", "Installed: containers-common-2:1-58.module_el8+804+f131391c.x86_64", "Installed: fuse3-libs-3.3.0-17.el8.x86_64", "Installed: libnet-1.1.6-15.el8.x86_64", "Installed: criu-libs-3.18-4.module_el8+804+f131391c.x86_64", "Installed: podman-plugins-3:4.8.3-1.module_el8+804+f131391c.x86_64", "Installed: crun-1.12-1.module_el8+804+f131391c.x86_64", "Installed: podman-3:4.8.3-1.module_el8+804+f131391c.x86_64", "Installed: slirp4netns-1.2.2-1.module_el8+804+f131391c.x86_64", "Installed: conmon-3:2.1.10-1.module_el8+804+f131391c.x86_64", "Installed: libslirp-4.4.0-1.module_el8+804+f131391c.x86_64", "Installed: podman-catatonit-3:4.8.3-1.module_el8+804+f131391c.x86_64", "Installed: protobuf-c-1.3.0-8.el8.x86_64", "Installed: dnsmasq-2.79-32.el8.x86_64" ] } TASK [fedora.linux_system_roles.podman : Get podman version] ******************* task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:22 Monday 05 February 2024 04:36:51 +0000 (0:00:59.692) 0:01:13.667 ******* ok: [sut] => { "changed": false, "cmd": [ "podman", "--version" ], "delta": "0:00:00.043822", "end": "2024-02-05 04:36:51.871571", "rc": 0, "start": "2024-02-05 04:36:51.827749" } STDOUT: podman version 4.8.3 TASK [fedora.linux_system_roles.podman : Set podman version] ******************* task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:27 Monday 05 February 2024 04:36:51 +0000 (0:00:00.328) 0:01:13.996 ******* ok: [sut] => { "ansible_facts": { "podman_version": "4.8.3" }, "changed": false } TASK [fedora.linux_system_roles.podman : Podman package version must be 4.2 or later] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:31 Monday 05 February 2024 04:36:51 +0000 (0:00:00.018) 0:01:14.015 ******* skipping: [sut] => { "changed": false, "false_condition": "podman_version is version(\"4.2\", \"<\")", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Podman package version must be 4.4 or later for quadlet, secrets] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:38 Monday 05 February 2024 04:36:51 +0000 (0:00:00.018) 0:01:14.033 ******* skipping: [sut] => { "changed": false, "false_condition": "podman_version is version(\"4.4\", \"<\")", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Podman package version must be 4.4 or later for quadlet, secrets] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:48 Monday 05 February 2024 04:36:51 +0000 (0:00:00.043) 0:01:14.076 ******* META: end_host conditional evaluated to False, continuing execution for sut skipping: [sut] => { "skip_reason": "end_host conditional evaluated to False, continuing execution for sut" } MSG: end_host conditional evaluated to false, continuing execution for sut TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:55 Monday 05 February 2024 04:36:52 +0000 (0:00:00.044) 0:01:14.121 ******* included: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for sut TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Monday 05 February 2024 04:36:52 +0000 (0:00:00.035) 0:01:14.157 ******* ok: [sut] => { "ansible_facts": { "getent_passwd": { "root": [ "x", "0", "0", "root", "/root", "/bin/bash" ] } }, "changed": false } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Monday 05 February 2024 04:36:53 +0000 (0:00:01.298) 0:01:15.455 ******* skipping: [sut] => { "changed": false, "false_condition": "not ansible_facts[\"getent_passwd\"][__podman_user]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Monday 05 February 2024 04:36:53 +0000 (0:00:00.020) 0:01:15.476 ******* ok: [sut] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get group information] **************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:27 Monday 05 February 2024 04:36:53 +0000 (0:00:00.024) 0:01:15.500 ******* ok: [sut] => { "ansible_facts": { "getent_group": { "root": [ "x", "0", "" ] } }, "changed": false } TASK [fedora.linux_system_roles.podman : Set group name] *********************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:34 Monday 05 February 2024 04:36:53 +0000 (0:00:00.199) 0:01:15.700 ******* ok: [sut] => { "ansible_facts": { "__podman_group_name": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:38 Monday 05 February 2024 04:36:53 +0000 (0:00:00.023) 0:01:15.724 ******* ok: [sut] => { "changed": false, "stat": { "atime": 1700557386.0, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "bb5b46ffbafcaa8c4021f3c8b3cb8594f48ef34b", "ctime": 1707107779.304328, "dev": 51713, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 6792906, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-sharedlib", "mode": "0755", "mtime": 1700557386.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 12640, "uid": 0, "version": "3297496774", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************ task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:49 Monday 05 February 2024 04:36:53 +0000 (0:00:00.198) 0:01:15.922 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check group with getsubids] *********** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:53 Monday 05 February 2024 04:36:53 +0000 (0:00:00.017) 0:01:15.939 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if user is in subuid file] ****** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:63 Monday 05 February 2024 04:36:53 +0000 (0:00:00.016) 0:01:15.955 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:71 Monday 05 February 2024 04:36:53 +0000 (0:00:00.016) 0:01:15.971 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if group is in subgid file] ***** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:78 Monday 05 February 2024 04:36:53 +0000 (0:00:00.015) 0:01:15.987 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] ***** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:86 Monday 05 February 2024 04:36:53 +0000 (0:00:00.015) 0:01:16.003 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set config file paths] **************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:61 Monday 05 February 2024 04:36:53 +0000 (0:00:00.015) 0:01:16.019 ******* ok: [sut] => { "ansible_facts": { "__podman_container_conf_file": "/etc/containers/containers.conf.d/50-systemroles.conf", "__podman_policy_json_file": "/etc/containers/policy.json", "__podman_registries_conf_file": "/etc/containers/registries.conf.d/50-systemroles.conf", "__podman_storage_conf_file": "/etc/containers/storage.conf" }, "changed": false } TASK [fedora.linux_system_roles.podman : Handle container.conf.d] ************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:70 Monday 05 February 2024 04:36:53 +0000 (0:00:00.035) 0:01:16.054 ******* included: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_container_conf_d.yml for sut TASK [fedora.linux_system_roles.podman : Ensure containers.d exists] *********** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_container_conf_d.yml:5 Monday 05 February 2024 04:36:53 +0000 (0:00:00.032) 0:01:16.087 ******* skipping: [sut] => { "changed": false, "false_condition": "podman_containers_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Update container config file] ********* task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_container_conf_d.yml:13 Monday 05 February 2024 04:36:54 +0000 (0:00:00.017) 0:01:16.104 ******* skipping: [sut] => { "changed": false, "false_condition": "podman_containers_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Handle registries.conf.d] ************* task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:73 Monday 05 February 2024 04:36:54 +0000 (0:00:00.016) 0:01:16.121 ******* included: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_registries_conf_d.yml for sut TASK [fedora.linux_system_roles.podman : Ensure registries.d exists] *********** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_registries_conf_d.yml:5 Monday 05 February 2024 04:36:54 +0000 (0:00:00.032) 0:01:16.153 ******* skipping: [sut] => { "changed": false, "false_condition": "podman_registries_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Update registries config file] ******** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_registries_conf_d.yml:13 Monday 05 February 2024 04:36:54 +0000 (0:00:00.016) 0:01:16.170 ******* skipping: [sut] => { "changed": false, "false_condition": "podman_registries_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Handle storage.conf] ****************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:76 Monday 05 February 2024 04:36:54 +0000 (0:00:00.016) 0:01:16.186 ******* included: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_storage_conf.yml for sut TASK [fedora.linux_system_roles.podman : Ensure storage.conf parent dir exists] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_storage_conf.yml:5 Monday 05 February 2024 04:36:54 +0000 (0:00:00.053) 0:01:16.240 ******* skipping: [sut] => { "changed": false, "false_condition": "podman_storage_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Update storage config file] *********** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_storage_conf.yml:13 Monday 05 February 2024 04:36:54 +0000 (0:00:00.018) 0:01:16.258 ******* skipping: [sut] => { "changed": false, "false_condition": "podman_storage_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Handle policy.json] ******************* task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:79 Monday 05 February 2024 04:36:54 +0000 (0:00:00.016) 0:01:16.274 ******* included: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml for sut TASK [fedora.linux_system_roles.podman : Ensure policy.json parent dir exists] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:6 Monday 05 February 2024 04:36:54 +0000 (0:00:00.034) 0:01:16.309 ******* skipping: [sut] => { "changed": false, "false_condition": "podman_policy_json | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Stat the policy.json file] ************ task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:14 Monday 05 February 2024 04:36:54 +0000 (0:00:00.016) 0:01:16.326 ******* skipping: [sut] => { "changed": false, "false_condition": "podman_policy_json | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get the existing policy.json] ********* task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:19 Monday 05 February 2024 04:36:54 +0000 (0:00:00.016) 0:01:16.343 ******* skipping: [sut] => { "changed": false, "false_condition": "podman_policy_json | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Write new policy.json file] *********** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:25 Monday 05 February 2024 04:36:54 +0000 (0:00:00.017) 0:01:16.360 ******* skipping: [sut] => { "changed": false, "false_condition": "podman_policy_json | length > 0", "skip_reason": "Conditional result was False" } TASK [Manage firewall for specified ports] ************************************* task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:85 Monday 05 February 2024 04:36:54 +0000 (0:00:00.016) 0:01:16.376 ******* TASK [fedora.linux_system_roles.firewall : Setup firewalld] ******************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:2 Monday 05 February 2024 04:36:54 +0000 (0:00:00.058) 0:01:16.434 ******* included: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml for sut TASK [fedora.linux_system_roles.firewall : Ensure ansible_facts used by role] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:2 Monday 05 February 2024 04:36:54 +0000 (0:00:00.028) 0:01:16.463 ******* skipping: [sut] => { "changed": false, "false_condition": "__firewall_required_facts | difference(ansible_facts.keys() | list) | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Check if system is ostree] ********** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:10 Monday 05 February 2024 04:36:54 +0000 (0:00:00.021) 0:01:16.484 ******* ok: [sut] => { "changed": false, "stat": { "exists": false } } TASK [fedora.linux_system_roles.firewall : Set flag to indicate system is ostree] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:15 Monday 05 February 2024 04:36:54 +0000 (0:00:00.192) 0:01:16.676 ******* ok: [sut] => { "ansible_facts": { "__firewall_is_ostree": false }, "changed": false } TASK [fedora.linux_system_roles.firewall : Install firewalld] ****************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:19 Monday 05 February 2024 04:36:54 +0000 (0:00:00.020) 0:01:16.697 ******* ok: [sut] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [fedora.linux_system_roles.firewall : Collect service facts] ************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:5 Monday 05 February 2024 04:36:56 +0000 (0:00:02.399) 0:01:19.096 ******* skipping: [sut] => { "changed": false, "false_condition": "firewall_disable_conflicting_services | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Attempt to stop and disable conflicting services] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:9 Monday 05 February 2024 04:36:57 +0000 (0:00:00.017) 0:01:19.114 ******* skipping: [sut] => (item=nftables) => { "ansible_loop_var": "item", "changed": false, "false_condition": "firewall_disable_conflicting_services | bool", "item": "nftables", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=iptables) => { "ansible_loop_var": "item", "changed": false, "false_condition": "firewall_disable_conflicting_services | bool", "item": "iptables", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=ufw) => { "ansible_loop_var": "item", "changed": false, "false_condition": "firewall_disable_conflicting_services | bool", "item": "ufw", "skip_reason": "Conditional result was False" } skipping: [sut] => { "changed": false } MSG: All items skipped TASK [fedora.linux_system_roles.firewall : Unmask firewalld service] *********** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:22 Monday 05 February 2024 04:36:57 +0000 (0:00:00.019) 0:01:19.133 ******* ok: [sut] => { "changed": false, "name": "firewalld", "status": { "ActiveEnterTimestampMonotonic": "0", "ActiveExitTimestampMonotonic": "0", "ActiveState": "inactive", "After": "dbus.service sysinit.target system.slice polkit.service dbus.socket basic.target", "AllowIsolate": "no", "AllowedCPUs": "", "AllowedMemoryNodes": "", "AmbientCapabilities": "", "AssertResult": "no", "AssertTimestampMonotonic": "0", "Before": "network-pre.target shutdown.target", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "BusName": "org.fedoraproject.FirewallD1", "CPUAccounting": "no", "CPUAffinity": "", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "[not set]", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "yes", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf", "CollectMode": "inactive", "ConditionResult": "no", "ConditionTimestampMonotonic": "0", "ConfigurationDirectoryMode": "0755", "Conflicts": "ebtables.service nftables.service iptables.service shutdown.target ipset.service ip6tables.service", "ControlPID": "0", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "Delegate": "no", "Description": "firewalld - dynamic firewall daemon", "DevicePolicy": "auto", "Documentation": "man:firewalld(1)", "DynamicUser": "no", "EffectiveCPUs": "", "EffectiveMemoryNodes": "", "EnvironmentFiles": "/etc/sysconfig/firewalld (ignore_errors=yes)", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "0", "ExecMainStartTimestampMonotonic": "0", "ExecMainStatus": "0", "ExecReload": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStart": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FragmentPath": "/usr/lib/systemd/system/firewalld.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOSchedulingClass": "0", "IOSchedulingPriority": "0", "IOWeight": "[not set]", "IPAccounting": "no", "IPEgressBytes": "18446744073709551615", "IPEgressPackets": "18446744073709551615", "IPIngressBytes": "18446744073709551615", "IPIngressPackets": "18446744073709551615", "Id": "firewalld.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestampMonotonic": "0", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "mixed", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "0", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "65536", "LimitMEMLOCKSoft": "65536", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "262144", "LimitNOFILESoft": "1024", "LimitNPROC": "14004", "LimitNPROCSoft": "14004", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "14004", "LimitSIGPENDINGSoft": "14004", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "0", "MemoryAccounting": "yes", "MemoryCurrent": "[not set]", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemorySwapMax": "infinity", "MountAPIVFS": "no", "MountFlags": "", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAMask": "", "NUMAPolicy": "n/a", "Names": "firewalld.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "PermissionsStartOnly": "no", "Perpetual": "no", "PrivateDevices": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "RemainAfterExit": "no", "RemoveIPC": "no", "Requires": "system.slice sysinit.target dbus.socket", "Restart": "no", "RestartUSec": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "Slice": "system.slice", "StandardError": "null", "StandardInput": "null", "StandardInputData": "", "StandardOutput": "null", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StateChangeTimestampMonotonic": "0", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "dead", "SuccessAction": "none", "SyslogFacility": "3", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "0", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "[not set]", "TasksMax": "22406", "TimeoutStartUSec": "1min 30s", "TimeoutStopUSec": "1min 30s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "dbus", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "enabled", "UnitFileState": "disabled", "UtmpMode": "init", "Wants": "network-pre.target", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "0" } } TASK [fedora.linux_system_roles.firewall : Enable and start firewalld service] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:28 Monday 05 February 2024 04:36:57 +0000 (0:00:00.369) 0:01:19.503 ******* changed: [sut] => { "changed": true, "enabled": true, "name": "firewalld", "state": "started", "status": { "ActiveEnterTimestampMonotonic": "0", "ActiveExitTimestampMonotonic": "0", "ActiveState": "inactive", "After": "sysinit.target basic.target polkit.service system.slice dbus.socket dbus.service", "AllowIsolate": "no", "AllowedCPUs": "", "AllowedMemoryNodes": "", "AmbientCapabilities": "", "AssertResult": "no", "AssertTimestampMonotonic": "0", "Before": "network-pre.target shutdown.target", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "BusName": "org.fedoraproject.FirewallD1", "CPUAccounting": "no", "CPUAffinity": "", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "[not set]", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "yes", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf", "CollectMode": "inactive", "ConditionResult": "no", "ConditionTimestampMonotonic": "0", "ConfigurationDirectoryMode": "0755", "Conflicts": "ipset.service ebtables.service shutdown.target ip6tables.service iptables.service nftables.service", "ControlPID": "0", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "Delegate": "no", "Description": "firewalld - dynamic firewall daemon", "DevicePolicy": "auto", "Documentation": "man:firewalld(1)", "DynamicUser": "no", "EffectiveCPUs": "", "EffectiveMemoryNodes": "", "EnvironmentFiles": "/etc/sysconfig/firewalld (ignore_errors=yes)", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "0", "ExecMainStartTimestampMonotonic": "0", "ExecMainStatus": "0", "ExecReload": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStart": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FragmentPath": "/usr/lib/systemd/system/firewalld.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOSchedulingClass": "0", "IOSchedulingPriority": "0", "IOWeight": "[not set]", "IPAccounting": "no", "IPEgressBytes": "18446744073709551615", "IPEgressPackets": "18446744073709551615", "IPIngressBytes": "18446744073709551615", "IPIngressPackets": "18446744073709551615", "Id": "firewalld.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestampMonotonic": "0", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "mixed", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "0", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "65536", "LimitMEMLOCKSoft": "65536", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "262144", "LimitNOFILESoft": "1024", "LimitNPROC": "14004", "LimitNPROCSoft": "14004", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "14004", "LimitSIGPENDINGSoft": "14004", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "0", "MemoryAccounting": "yes", "MemoryCurrent": "[not set]", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemorySwapMax": "infinity", "MountAPIVFS": "no", "MountFlags": "", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAMask": "", "NUMAPolicy": "n/a", "Names": "firewalld.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "PermissionsStartOnly": "no", "Perpetual": "no", "PrivateDevices": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "RemainAfterExit": "no", "RemoveIPC": "no", "Requires": "sysinit.target system.slice dbus.socket", "Restart": "no", "RestartUSec": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "Slice": "system.slice", "StandardError": "null", "StandardInput": "null", "StandardInputData": "", "StandardOutput": "null", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StateChangeTimestampMonotonic": "0", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "dead", "SuccessAction": "none", "SyslogFacility": "3", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "0", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "[not set]", "TasksMax": "22406", "TimeoutStartUSec": "1min 30s", "TimeoutStopUSec": "1min 30s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "dbus", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "enabled", "UnitFileState": "disabled", "UtmpMode": "init", "Wants": "network-pre.target", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "0" } } TASK [fedora.linux_system_roles.firewall : Check if previous replaced is defined] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:34 Monday 05 February 2024 04:36:58 +0000 (0:00:01.008) 0:01:20.511 ******* ok: [sut] => { "ansible_facts": { "__firewall_previous_replaced": false, "__firewall_python_cmd": "/usr/bin/python3", "__firewall_report_changed": true }, "changed": false } TASK [fedora.linux_system_roles.firewall : Get config files, checksums before and remove] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:43 Monday 05 February 2024 04:36:58 +0000 (0:00:00.025) 0:01:20.537 ******* skipping: [sut] => { "changed": false, "false_condition": "__firewall_previous_replaced | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Tell firewall module it is able to report changed] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:55 Monday 05 February 2024 04:36:58 +0000 (0:00:00.016) 0:01:20.553 ******* skipping: [sut] => { "changed": false, "false_condition": "__firewall_previous_replaced | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Configure firewall] ***************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:71 Monday 05 February 2024 04:36:58 +0000 (0:00:00.016) 0:01:20.569 ******* changed: [sut] => (item={'port': '8000/tcp', 'state': 'enabled'}) => { "__firewall_changed": true, "ansible_loop_var": "item", "changed": true, "item": { "port": "8000/tcp", "state": "enabled" } } changed: [sut] => (item={'port': '9000/tcp', 'state': 'enabled'}) => { "__firewall_changed": true, "ansible_loop_var": "item", "changed": true, "item": { "port": "9000/tcp", "state": "enabled" } } TASK [fedora.linux_system_roles.firewall : Gather firewall config information] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:120 Monday 05 February 2024 04:36:59 +0000 (0:00:01.298) 0:01:21.868 ******* skipping: [sut] => (item={'port': '8000/tcp', 'state': 'enabled'}) => { "ansible_loop_var": "item", "changed": false, "false_condition": "firewall | length == 1", "item": { "port": "8000/tcp", "state": "enabled" }, "skip_reason": "Conditional result was False" } skipping: [sut] => (item={'port': '9000/tcp', 'state': 'enabled'}) => { "ansible_loop_var": "item", "changed": false, "false_condition": "firewall | length == 1", "item": { "port": "9000/tcp", "state": "enabled" }, "skip_reason": "Conditional result was False" } skipping: [sut] => { "changed": false } MSG: All items skipped TASK [fedora.linux_system_roles.firewall : Update firewalld_config fact] ******* task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:130 Monday 05 February 2024 04:36:59 +0000 (0:00:00.028) 0:01:21.896 ******* skipping: [sut] => { "changed": false, "false_condition": "firewall | length == 1", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Gather firewall config if no arguments] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:139 Monday 05 February 2024 04:36:59 +0000 (0:00:00.020) 0:01:21.917 ******* skipping: [sut] => { "changed": false, "false_condition": "firewall == None or firewall | length == 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Update firewalld_config fact] ******* task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:144 Monday 05 February 2024 04:36:59 +0000 (0:00:00.018) 0:01:21.935 ******* skipping: [sut] => { "changed": false, "false_condition": "firewall == None or firewall | length == 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Get config files, checksums after] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:153 Monday 05 February 2024 04:36:59 +0000 (0:00:00.019) 0:01:21.954 ******* skipping: [sut] => { "changed": false, "false_condition": "__firewall_previous_replaced | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Calculate what has changed] ********* task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:163 Monday 05 February 2024 04:36:59 +0000 (0:00:00.017) 0:01:21.971 ******* skipping: [sut] => { "changed": false, "false_condition": "__firewall_previous_replaced | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Show diffs] ************************* task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:169 Monday 05 February 2024 04:36:59 +0000 (0:00:00.017) 0:01:21.989 ******* skipping: [sut] => { "false_condition": "__firewall_previous_replaced | bool" } TASK [Manage selinux for specified ports] ************************************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:92 Monday 05 February 2024 04:36:59 +0000 (0:00:00.026) 0:01:22.016 ******* skipping: [sut] => { "changed": false, "false_condition": "podman_selinux_ports | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Keep track of users that need to cancel linger] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:99 Monday 05 February 2024 04:36:59 +0000 (0:00:00.018) 0:01:22.034 ******* ok: [sut] => { "ansible_facts": { "__podman_cancel_user_linger": [] }, "changed": false } TASK [fedora.linux_system_roles.podman : Handle secrets] *********************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:103 Monday 05 February 2024 04:36:59 +0000 (0:00:00.019) 0:01:22.053 ******* included: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml for sut => (item=(censored due to no_log)) included: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml for sut => (item=(censored due to no_log)) included: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml for sut => (item=(censored due to no_log)) TASK [fedora.linux_system_roles.podman : Set variables part 0] ***************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:3 Monday 05 February 2024 04:37:00 +0000 (0:00:00.066) 0:01:22.120 ******* ok: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Set variables part 1] ***************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:10 Monday 05 February 2024 04:37:00 +0000 (0:00:00.021) 0:01:22.141 ******* ok: [sut] => { "ansible_facts": { "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set variables part 2] ***************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:14 Monday 05 February 2024 04:37:00 +0000 (0:00:00.020) 0:01:22.162 ******* ok: [sut] => { "ansible_facts": { "__podman_rootless": false, "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:20 Monday 05 February 2024 04:37:00 +0000 (0:00:00.022) 0:01:22.185 ******* included: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for sut TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Monday 05 February 2024 04:37:00 +0000 (0:00:00.053) 0:01:22.239 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Monday 05 February 2024 04:37:00 +0000 (0:00:00.017) 0:01:22.256 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Monday 05 February 2024 04:37:00 +0000 (0:00:00.020) 0:01:22.276 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Manage each secret] ******************* task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:25 Monday 05 February 2024 04:37:00 +0000 (0:00:00.017) 0:01:22.294 ******* ok: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Set variables part 0] ***************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:3 Monday 05 February 2024 04:37:00 +0000 (0:00:00.302) 0:01:22.596 ******* ok: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Set variables part 1] ***************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:10 Monday 05 February 2024 04:37:00 +0000 (0:00:00.023) 0:01:22.619 ******* ok: [sut] => { "ansible_facts": { "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set variables part 2] ***************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:14 Monday 05 February 2024 04:37:00 +0000 (0:00:00.020) 0:01:22.640 ******* ok: [sut] => { "ansible_facts": { "__podman_rootless": false, "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:20 Monday 05 February 2024 04:37:00 +0000 (0:00:00.024) 0:01:22.664 ******* included: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for sut TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Monday 05 February 2024 04:37:00 +0000 (0:00:00.027) 0:01:22.692 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Monday 05 February 2024 04:37:00 +0000 (0:00:00.016) 0:01:22.709 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Monday 05 February 2024 04:37:00 +0000 (0:00:00.016) 0:01:22.725 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Manage each secret] ******************* task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:25 Monday 05 February 2024 04:37:00 +0000 (0:00:00.016) 0:01:22.741 ******* ok: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Set variables part 0] ***************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:3 Monday 05 February 2024 04:37:00 +0000 (0:00:00.239) 0:01:22.980 ******* ok: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Set variables part 1] ***************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:10 Monday 05 February 2024 04:37:00 +0000 (0:00:00.021) 0:01:23.002 ******* ok: [sut] => { "ansible_facts": { "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set variables part 2] ***************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:14 Monday 05 February 2024 04:37:00 +0000 (0:00:00.020) 0:01:23.023 ******* ok: [sut] => { "ansible_facts": { "__podman_rootless": false, "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:20 Monday 05 February 2024 04:37:00 +0000 (0:00:00.023) 0:01:23.046 ******* included: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for sut TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Monday 05 February 2024 04:37:00 +0000 (0:00:00.028) 0:01:23.075 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Monday 05 February 2024 04:37:00 +0000 (0:00:00.017) 0:01:23.092 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Monday 05 February 2024 04:37:01 +0000 (0:00:00.017) 0:01:23.110 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Manage each secret] ******************* task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:25 Monday 05 February 2024 04:37:01 +0000 (0:00:00.040) 0:01:23.150 ******* ok: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Handle Kubernetes specifications] ***** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:110 Monday 05 February 2024 04:37:01 +0000 (0:00:00.236) 0:01:23.387 ******* skipping: [sut] => { "changed": false, "skipped_reason": "No items in the list" } TASK [fedora.linux_system_roles.podman : Handle Quadlet specifications] ******** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:116 Monday 05 February 2024 04:37:01 +0000 (0:00:00.015) 0:01:23.403 ******* included: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for sut => (item={'file_src': 'quadlet-demo.network'}) included: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for sut => (item={'file_src': 'quadlet-demo-mysql.volume'}) included: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for sut => (item={'template_src': 'quadlet-demo-mysql.container.j2'}) included: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for sut => (item={'file_src': 'envoy-proxy-configmap.yml'}) included: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for sut => (item={'template_src': 'quadlet-demo.yml.j2'}) included: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for sut => (item={'file_src': 'quadlet-demo.kube'}) TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14 Monday 05 February 2024 04:37:01 +0000 (0:00:00.081) 0:01:23.484 ******* ok: [sut] => { "ansible_facts": { "__podman_quadlet_file_src": "quadlet-demo.network", "__podman_quadlet_spec": {}, "__podman_quadlet_str": "[Network]\nSubnet=192.168.30.0/24\nGateway=192.168.30.1\nLabel=app=wordpress", "__podman_quadlet_template_src": "" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25 Monday 05 February 2024 04:37:01 +0000 (0:00:00.027) 0:01:23.511 ******* ok: [sut] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_pull_image": true, "__podman_state": "created", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] ***** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35 Monday 05 February 2024 04:37:01 +0000 (0:00:00.024) 0:01:23.536 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_quadlet_file_src", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48 Monday 05 February 2024 04:37:01 +0000 (0:00:00.019) 0:01:23.555 ******* ok: [sut] => { "ansible_facts": { "__podman_quadlet_name": "quadlet-demo", "__podman_quadlet_type": "network", "__podman_rootless": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57 Monday 05 February 2024 04:37:01 +0000 (0:00:00.028) 0:01:23.584 ******* included: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for sut TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Monday 05 February 2024 04:37:01 +0000 (0:00:00.033) 0:01:23.617 ******* skipping: [sut] => { "changed": false, "false_condition": "'getent_passwd' not in ansible_facts or __podman_user not in ansible_facts['getent_passwd']", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Monday 05 February 2024 04:37:01 +0000 (0:00:00.021) 0:01:23.639 ******* skipping: [sut] => { "changed": false, "false_condition": "not ansible_facts[\"getent_passwd\"][__podman_user]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Monday 05 February 2024 04:37:01 +0000 (0:00:00.020) 0:01:23.659 ******* ok: [sut] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get group information] **************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:27 Monday 05 February 2024 04:37:01 +0000 (0:00:00.025) 0:01:23.684 ******* ok: [sut] => { "ansible_facts": { "getent_group": { "root": [ "x", "0", "" ] } }, "changed": false } TASK [fedora.linux_system_roles.podman : Set group name] *********************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:34 Monday 05 February 2024 04:37:01 +0000 (0:00:00.199) 0:01:23.884 ******* ok: [sut] => { "ansible_facts": { "__podman_group_name": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:38 Monday 05 February 2024 04:37:01 +0000 (0:00:00.025) 0:01:23.909 ******* ok: [sut] => { "changed": false, "stat": { "atime": 1707107813.7932265, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "bb5b46ffbafcaa8c4021f3c8b3cb8594f48ef34b", "ctime": 1707107779.304328, "dev": 51713, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 6792906, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-sharedlib", "mode": "0755", "mtime": 1700557386.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 12640, "uid": 0, "version": "3297496774", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************ task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:49 Monday 05 February 2024 04:37:02 +0000 (0:00:00.200) 0:01:24.110 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check group with getsubids] *********** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:53 Monday 05 February 2024 04:37:02 +0000 (0:00:00.018) 0:01:24.128 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if user is in subuid file] ****** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:63 Monday 05 February 2024 04:37:02 +0000 (0:00:00.017) 0:01:24.146 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:71 Monday 05 February 2024 04:37:02 +0000 (0:00:00.044) 0:01:24.190 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if group is in subgid file] ***** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:78 Monday 05 February 2024 04:37:02 +0000 (0:00:00.017) 0:01:24.208 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] ***** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:86 Monday 05 February 2024 04:37:02 +0000 (0:00:00.018) 0:01:24.226 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62 Monday 05 February 2024 04:37:02 +0000 (0:00:00.017) 0:01:24.243 ******* ok: [sut] => { "ansible_facts": { "__podman_activate_systemd_unit": true, "__podman_images_found": [], "__podman_kube_yamls_raw": "", "__podman_service_name": "", "__podman_systemd_scope": "system", "__podman_user_home_dir": "/root", "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73 Monday 05 February 2024 04:37:02 +0000 (0:00:00.033) 0:01:24.276 ******* ok: [sut] => { "ansible_facts": { "__podman_quadlet_path": "/etc/containers/systemd" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get kube yaml contents] *************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77 Monday 05 February 2024 04:37:02 +0000 (0:00:00.019) 0:01:24.296 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_kube_yamls_raw | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87 Monday 05 February 2024 04:37:02 +0000 (0:00:00.019) 0:01:24.316 ******* ok: [sut] => { "ansible_facts": { "__podman_images": [], "__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo.network", "__podman_volumes": [] }, "changed": false } TASK [fedora.linux_system_roles.podman : Cleanup quadlets] ********************* task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:103 Monday 05 February 2024 04:37:02 +0000 (0:00:00.045) 0:01:24.361 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_state == \"absent\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create and update quadlets] *********** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:107 Monday 05 February 2024 04:37:02 +0000 (0:00:00.018) 0:01:24.379 ******* included: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml for sut TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:2 Monday 05 February 2024 04:37:02 +0000 (0:00:00.039) 0:01:24.419 ******* included: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for sut TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Monday 05 February 2024 04:37:02 +0000 (0:00:00.030) 0:01:24.449 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Monday 05 February 2024 04:37:02 +0000 (0:00:00.017) 0:01:24.467 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Monday 05 February 2024 04:37:02 +0000 (0:00:00.018) 0:01:24.485 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create host directories] ************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:7 Monday 05 February 2024 04:37:02 +0000 (0:00:00.017) 0:01:24.502 ******* skipping: [sut] => { "changed": false, "skipped_reason": "No items in the list" } TASK [fedora.linux_system_roles.podman : Ensure container images are present] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:20 Monday 05 February 2024 04:37:02 +0000 (0:00:00.017) 0:01:24.519 ******* skipping: [sut] => { "changed": false, "skipped_reason": "No items in the list" } TASK [fedora.linux_system_roles.podman : Ensure the quadlet directory is present] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:37 Monday 05 February 2024 04:37:02 +0000 (0:00:00.017) 0:01:24.536 ******* ok: [sut] => { "changed": false, "gid": 0, "group": "root", "mode": "0755", "owner": "root", "path": "/etc/containers/systemd", "secontext": "system_u:object_r:etc_t:s0", "size": 6, "state": "directory", "uid": 0 } TASK [fedora.linux_system_roles.podman : Ensure quadlet file is copied] ******** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:46 Monday 05 February 2024 04:37:02 +0000 (0:00:00.210) 0:01:24.747 ******* changed: [sut] => { "changed": true, "checksum": "e57c08d49aff4bae8daab138d913aeddaa8682a0", "dest": "/etc/containers/systemd/quadlet-demo.network", "gid": 0, "group": "root", "md5sum": "061f3cf318cbd8ab5794bb1173831fb8", "mode": "0644", "owner": "root", "secontext": "system_u:object_r:etc_t:s0", "size": 74, "src": "/root/.ansible/tmp/ansible-tmp-1707107822.6875324-8017-246495456961851/source", "state": "file", "uid": 0 } TASK [fedora.linux_system_roles.podman : Ensure quadlet file content is present] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:56 Monday 05 February 2024 04:37:03 +0000 (0:00:00.622) 0:01:25.369 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_quadlet_file_src", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Ensure quadlet file is present] ******* task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:68 Monday 05 February 2024 04:37:03 +0000 (0:00:00.019) 0:01:25.388 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_copy_file is skipped", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Reload systemctl] ********************* task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:80 Monday 05 February 2024 04:37:03 +0000 (0:00:00.018) 0:01:25.406 ******* ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.podman : Start service] ************************ task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:108 Monday 05 February 2024 04:37:03 +0000 (0:00:00.497) 0:01:25.904 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_service_name | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Restart service] ********************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:123 Monday 05 February 2024 04:37:03 +0000 (0:00:00.020) 0:01:25.925 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_service_name | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14 Monday 05 February 2024 04:37:03 +0000 (0:00:00.021) 0:01:25.947 ******* ok: [sut] => { "ansible_facts": { "__podman_quadlet_file_src": "quadlet-demo-mysql.volume", "__podman_quadlet_spec": {}, "__podman_quadlet_str": "[Volume]", "__podman_quadlet_template_src": "" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25 Monday 05 February 2024 04:37:03 +0000 (0:00:00.027) 0:01:25.974 ******* ok: [sut] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_pull_image": true, "__podman_state": "created", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] ***** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35 Monday 05 February 2024 04:37:03 +0000 (0:00:00.024) 0:01:25.998 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_quadlet_file_src", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48 Monday 05 February 2024 04:37:03 +0000 (0:00:00.018) 0:01:26.017 ******* ok: [sut] => { "ansible_facts": { "__podman_quadlet_name": "quadlet-demo-mysql", "__podman_quadlet_type": "volume", "__podman_rootless": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57 Monday 05 February 2024 04:37:03 +0000 (0:00:00.028) 0:01:26.046 ******* included: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for sut TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Monday 05 February 2024 04:37:03 +0000 (0:00:00.034) 0:01:26.080 ******* skipping: [sut] => { "changed": false, "false_condition": "'getent_passwd' not in ansible_facts or __podman_user not in ansible_facts['getent_passwd']", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Monday 05 February 2024 04:37:04 +0000 (0:00:00.020) 0:01:26.101 ******* skipping: [sut] => { "changed": false, "false_condition": "not ansible_facts[\"getent_passwd\"][__podman_user]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Monday 05 February 2024 04:37:04 +0000 (0:00:00.022) 0:01:26.123 ******* ok: [sut] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get group information] **************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:27 Monday 05 February 2024 04:37:04 +0000 (0:00:00.028) 0:01:26.152 ******* ok: [sut] => { "ansible_facts": { "getent_group": { "root": [ "x", "0", "" ] } }, "changed": false } TASK [fedora.linux_system_roles.podman : Set group name] *********************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:34 Monday 05 February 2024 04:37:04 +0000 (0:00:00.202) 0:01:26.355 ******* ok: [sut] => { "ansible_facts": { "__podman_group_name": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:38 Monday 05 February 2024 04:37:04 +0000 (0:00:00.025) 0:01:26.380 ******* ok: [sut] => { "changed": false, "stat": { "atime": 1707107813.7932265, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "bb5b46ffbafcaa8c4021f3c8b3cb8594f48ef34b", "ctime": 1707107779.304328, "dev": 51713, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 6792906, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-sharedlib", "mode": "0755", "mtime": 1700557386.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 12640, "uid": 0, "version": "3297496774", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************ task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:49 Monday 05 February 2024 04:37:04 +0000 (0:00:00.198) 0:01:26.579 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check group with getsubids] *********** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:53 Monday 05 February 2024 04:37:04 +0000 (0:00:00.018) 0:01:26.597 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if user is in subuid file] ****** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:63 Monday 05 February 2024 04:37:04 +0000 (0:00:00.017) 0:01:26.614 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:71 Monday 05 February 2024 04:37:04 +0000 (0:00:00.018) 0:01:26.633 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if group is in subgid file] ***** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:78 Monday 05 February 2024 04:37:04 +0000 (0:00:00.017) 0:01:26.651 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] ***** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:86 Monday 05 February 2024 04:37:04 +0000 (0:00:00.018) 0:01:26.669 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62 Monday 05 February 2024 04:37:04 +0000 (0:00:00.016) 0:01:26.686 ******* ok: [sut] => { "ansible_facts": { "__podman_activate_systemd_unit": true, "__podman_images_found": [], "__podman_kube_yamls_raw": "", "__podman_service_name": "quadlet-demo-mysql-volume.service", "__podman_systemd_scope": "system", "__podman_user_home_dir": "/root", "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73 Monday 05 February 2024 04:37:04 +0000 (0:00:00.032) 0:01:26.719 ******* ok: [sut] => { "ansible_facts": { "__podman_quadlet_path": "/etc/containers/systemd" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get kube yaml contents] *************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77 Monday 05 February 2024 04:37:04 +0000 (0:00:00.048) 0:01:26.767 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_kube_yamls_raw | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87 Monday 05 February 2024 04:37:04 +0000 (0:00:00.019) 0:01:26.787 ******* ok: [sut] => { "ansible_facts": { "__podman_images": [], "__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo-mysql.volume", "__podman_volumes": [] }, "changed": false } TASK [fedora.linux_system_roles.podman : Cleanup quadlets] ********************* task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:103 Monday 05 February 2024 04:37:04 +0000 (0:00:00.045) 0:01:26.833 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_state == \"absent\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create and update quadlets] *********** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:107 Monday 05 February 2024 04:37:04 +0000 (0:00:00.018) 0:01:26.851 ******* included: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml for sut TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:2 Monday 05 February 2024 04:37:04 +0000 (0:00:00.037) 0:01:26.889 ******* included: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for sut TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Monday 05 February 2024 04:37:04 +0000 (0:00:00.029) 0:01:26.919 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Monday 05 February 2024 04:37:04 +0000 (0:00:00.017) 0:01:26.936 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Monday 05 February 2024 04:37:04 +0000 (0:00:00.018) 0:01:26.955 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create host directories] ************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:7 Monday 05 February 2024 04:37:04 +0000 (0:00:00.017) 0:01:26.973 ******* skipping: [sut] => { "changed": false, "skipped_reason": "No items in the list" } TASK [fedora.linux_system_roles.podman : Ensure container images are present] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:20 Monday 05 February 2024 04:37:04 +0000 (0:00:00.015) 0:01:26.989 ******* skipping: [sut] => { "changed": false, "skipped_reason": "No items in the list" } TASK [fedora.linux_system_roles.podman : Ensure the quadlet directory is present] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:37 Monday 05 February 2024 04:37:04 +0000 (0:00:00.017) 0:01:27.006 ******* ok: [sut] => { "changed": false, "gid": 0, "group": "root", "mode": "0755", "owner": "root", "path": "/etc/containers/systemd", "secontext": "system_u:object_r:etc_t:s0", "size": 34, "state": "directory", "uid": 0 } TASK [fedora.linux_system_roles.podman : Ensure quadlet file is copied] ******** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:46 Monday 05 February 2024 04:37:05 +0000 (0:00:00.207) 0:01:27.214 ******* changed: [sut] => { "changed": true, "checksum": "585f8cbdf0ec73000f9227dcffbef71e9552ea4a", "dest": "/etc/containers/systemd/quadlet-demo-mysql.volume", "gid": 0, "group": "root", "md5sum": "5ddd03a022aeb4502d9bc8ce436b4233", "mode": "0644", "owner": "root", "secontext": "system_u:object_r:etc_t:s0", "size": 9, "src": "/root/.ansible/tmp/ansible-tmp-1707107825.1552148-8065-231360590828526/source", "state": "file", "uid": 0 } TASK [fedora.linux_system_roles.podman : Ensure quadlet file content is present] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:56 Monday 05 February 2024 04:37:05 +0000 (0:00:00.523) 0:01:27.737 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_quadlet_file_src", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Ensure quadlet file is present] ******* task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:68 Monday 05 February 2024 04:37:05 +0000 (0:00:00.019) 0:01:27.756 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_copy_file is skipped", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Reload systemctl] ********************* task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:80 Monday 05 February 2024 04:37:05 +0000 (0:00:00.017) 0:01:27.774 ******* ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.podman : Start service] ************************ task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:108 Monday 05 February 2024 04:37:06 +0000 (0:00:00.466) 0:01:28.241 ******* changed: [sut] => { "changed": true, "name": "quadlet-demo-mysql-volume.service", "state": "started", "status": { "ActiveEnterTimestampMonotonic": "0", "ActiveExitTimestampMonotonic": "0", "ActiveState": "inactive", "After": "sysinit.target -.mount system.slice basic.target systemd-journald.socket", "AllowIsolate": "no", "AllowedCPUs": "", "AllowedMemoryNodes": "", "AmbientCapabilities": "", "AssertResult": "no", "AssertTimestampMonotonic": "0", "Before": "shutdown.target", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "CPUAccounting": "no", "CPUAffinity": "", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "[not set]", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "no", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf", "CollectMode": "inactive", "ConditionResult": "no", "ConditionTimestampMonotonic": "0", "ConfigurationDirectoryMode": "0755", "Conflicts": "shutdown.target", "ControlPID": "0", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "Delegate": "no", "Description": "quadlet-demo-mysql-volume.service", "DevicePolicy": "auto", "DynamicUser": "no", "EffectiveCPUs": "", "EffectiveMemoryNodes": "", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "0", "ExecMainStartTimestampMonotonic": "0", "ExecMainStatus": "0", "ExecStart": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman volume create --ignore systemd-quadlet-demo-mysql ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FragmentPath": "/run/systemd/generator/quadlet-demo-mysql-volume.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOSchedulingClass": "0", "IOSchedulingPriority": "0", "IOWeight": "[not set]", "IPAccounting": "no", "IPEgressBytes": "18446744073709551615", "IPEgressPackets": "18446744073709551615", "IPIngressBytes": "18446744073709551615", "IPIngressPackets": "18446744073709551615", "Id": "quadlet-demo-mysql-volume.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestampMonotonic": "0", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "control-group", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "0", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "65536", "LimitMEMLOCKSoft": "65536", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "262144", "LimitNOFILESoft": "1024", "LimitNPROC": "14004", "LimitNPROCSoft": "14004", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "14004", "LimitSIGPENDINGSoft": "14004", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "0", "MemoryAccounting": "yes", "MemoryCurrent": "[not set]", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemorySwapMax": "infinity", "MountAPIVFS": "no", "MountFlags": "", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAMask": "", "NUMAPolicy": "n/a", "Names": "quadlet-demo-mysql-volume.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "PermissionsStartOnly": "no", "Perpetual": "no", "PrivateDevices": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "RemainAfterExit": "yes", "RemoveIPC": "no", "Requires": "sysinit.target -.mount system.slice", "RequiresMountsFor": "/run/containers", "Restart": "no", "RestartUSec": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "Slice": "system.slice", "StandardError": "inherit", "StandardInput": "null", "StandardInputData": "", "StandardOutput": "journal", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StateChangeTimestampMonotonic": "0", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "dead", "SuccessAction": "none", "SyslogFacility": "3", "SyslogIdentifier": "quadlet-demo-mysql-volume", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "0", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "[not set]", "TasksMax": "22406", "TimeoutStartUSec": "infinity", "TimeoutStopUSec": "1min 30s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "oneshot", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "disabled", "UnitFileState": "generated", "UtmpMode": "init", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "0" } } TASK [fedora.linux_system_roles.podman : Restart service] ********************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:123 Monday 05 February 2024 04:37:06 +0000 (0:00:00.462) 0:01:28.703 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_service_started is changed", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14 Monday 05 February 2024 04:37:06 +0000 (0:00:00.019) 0:01:28.722 ******* ok: [sut] => { "ansible_facts": { "__podman_quadlet_file_src": "", "__podman_quadlet_spec": {}, "__podman_quadlet_str": "[Install]\nWantedBy=default.target\n\n[Container]\nImage=quay.io/linux-system-roles/mysql:5.6\nContainerName=quadlet-demo-mysql\nVolume=quadlet-demo-mysql.volume:/var/lib/mysql\nVolume=/tmp/quadlet_demo:/var/lib/quadlet_demo:Z\nNetwork=quadlet-demo.network\nSecret=mysql-root-password-container,type=env,target=MYSQL_ROOT_PASSWORD\nHealthCmd=/usr/bin/true\nHealthOnFailure=kill\n", "__podman_quadlet_template_src": "quadlet-demo-mysql.container.j2" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25 Monday 05 February 2024 04:37:06 +0000 (0:00:00.086) 0:01:28.808 ******* ok: [sut] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_pull_image": true, "__podman_state": "created", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] ***** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35 Monday 05 February 2024 04:37:06 +0000 (0:00:00.054) 0:01:28.863 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_quadlet_str", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48 Monday 05 February 2024 04:37:06 +0000 (0:00:00.019) 0:01:28.883 ******* ok: [sut] => { "ansible_facts": { "__podman_quadlet_name": "quadlet-demo-mysql", "__podman_quadlet_type": "container", "__podman_rootless": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57 Monday 05 February 2024 04:37:06 +0000 (0:00:00.029) 0:01:28.912 ******* included: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for sut TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Monday 05 February 2024 04:37:06 +0000 (0:00:00.034) 0:01:28.947 ******* skipping: [sut] => { "changed": false, "false_condition": "'getent_passwd' not in ansible_facts or __podman_user not in ansible_facts['getent_passwd']", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Monday 05 February 2024 04:37:06 +0000 (0:00:00.022) 0:01:28.970 ******* skipping: [sut] => { "changed": false, "false_condition": "not ansible_facts[\"getent_passwd\"][__podman_user]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Monday 05 February 2024 04:37:06 +0000 (0:00:00.021) 0:01:28.991 ******* ok: [sut] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get group information] **************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:27 Monday 05 February 2024 04:37:06 +0000 (0:00:00.026) 0:01:29.017 ******* ok: [sut] => { "ansible_facts": { "getent_group": { "root": [ "x", "0", "" ] } }, "changed": false } TASK [fedora.linux_system_roles.podman : Set group name] *********************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:34 Monday 05 February 2024 04:37:07 +0000 (0:00:00.201) 0:01:29.219 ******* ok: [sut] => { "ansible_facts": { "__podman_group_name": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:38 Monday 05 February 2024 04:37:07 +0000 (0:00:00.026) 0:01:29.246 ******* ok: [sut] => { "changed": false, "stat": { "atime": 1707107813.7932265, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "bb5b46ffbafcaa8c4021f3c8b3cb8594f48ef34b", "ctime": 1707107779.304328, "dev": 51713, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 6792906, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-sharedlib", "mode": "0755", "mtime": 1700557386.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 12640, "uid": 0, "version": "3297496774", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************ task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:49 Monday 05 February 2024 04:37:07 +0000 (0:00:00.199) 0:01:29.445 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check group with getsubids] *********** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:53 Monday 05 February 2024 04:37:07 +0000 (0:00:00.018) 0:01:29.464 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if user is in subuid file] ****** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:63 Monday 05 February 2024 04:37:07 +0000 (0:00:00.017) 0:01:29.481 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:71 Monday 05 February 2024 04:37:07 +0000 (0:00:00.019) 0:01:29.501 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if group is in subgid file] ***** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:78 Monday 05 February 2024 04:37:07 +0000 (0:00:00.017) 0:01:29.518 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] ***** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:86 Monday 05 February 2024 04:37:07 +0000 (0:00:00.018) 0:01:29.537 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62 Monday 05 February 2024 04:37:07 +0000 (0:00:00.017) 0:01:29.555 ******* ok: [sut] => { "ansible_facts": { "__podman_activate_systemd_unit": true, "__podman_images_found": [ "quay.io/linux-system-roles/mysql:5.6" ], "__podman_kube_yamls_raw": "", "__podman_service_name": "quadlet-demo-mysql.service", "__podman_systemd_scope": "system", "__podman_user_home_dir": "/root", "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73 Monday 05 February 2024 04:37:07 +0000 (0:00:00.033) 0:01:29.588 ******* ok: [sut] => { "ansible_facts": { "__podman_quadlet_path": "/etc/containers/systemd" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get kube yaml contents] *************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77 Monday 05 February 2024 04:37:07 +0000 (0:00:00.019) 0:01:29.608 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_kube_yamls_raw | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87 Monday 05 February 2024 04:37:07 +0000 (0:00:00.024) 0:01:29.633 ******* ok: [sut] => { "ansible_facts": { "__podman_images": [ "quay.io/linux-system-roles/mysql:5.6" ], "__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo-mysql.container", "__podman_volumes": [ "/tmp/quadlet_demo" ] }, "changed": false } TASK [fedora.linux_system_roles.podman : Cleanup quadlets] ********************* task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:103 Monday 05 February 2024 04:37:07 +0000 (0:00:00.045) 0:01:29.678 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_state == \"absent\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create and update quadlets] *********** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:107 Monday 05 February 2024 04:37:07 +0000 (0:00:00.047) 0:01:29.725 ******* included: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml for sut TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:2 Monday 05 February 2024 04:37:07 +0000 (0:00:00.040) 0:01:29.765 ******* included: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for sut TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Monday 05 February 2024 04:37:07 +0000 (0:00:00.030) 0:01:29.795 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Monday 05 February 2024 04:37:07 +0000 (0:00:00.017) 0:01:29.813 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Monday 05 February 2024 04:37:07 +0000 (0:00:00.018) 0:01:29.831 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create host directories] ************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:7 Monday 05 February 2024 04:37:07 +0000 (0:00:00.017) 0:01:29.848 ******* changed: [sut] => (item=/tmp/quadlet_demo) => { "ansible_loop_var": "item", "changed": true, "gid": 0, "group": "root", "item": "/tmp/quadlet_demo", "mode": "0777", "owner": "root", "path": "/tmp/quadlet_demo", "secontext": "unconfined_u:object_r:user_tmp_t:s0", "size": 6, "state": "directory", "uid": 0 } TASK [fedora.linux_system_roles.podman : Ensure container images are present] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:20 Monday 05 February 2024 04:37:07 +0000 (0:00:00.212) 0:01:30.061 ******* changed: [sut] => (item=quay.io/linux-system-roles/mysql:5.6) => { "actions": [ "Pulled image quay.io/linux-system-roles/mysql:5.6" ], "ansible_loop_var": "item", "changed": true, "failed_when_result": false, "image": [ { "Annotations": {}, "Architecture": "amd64", "Author": "", "Comment": "", "Config": { "Cmd": [ "mysqld" ], "Entrypoint": [ "docker-entrypoint.sh" ], "Env": [ "PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin", "GOSU_VERSION=1.12", "MYSQL_MAJOR=5.6", "MYSQL_VERSION=5.6.51-1debian9" ], "ExposedPorts": { "3306/tcp": {} }, "Volumes": { "/var/lib/mysql": {} } }, "Created": "2021-12-21T02:57:59.587337844Z", "Digest": "sha256:2d0419de75076a55e0d0c0680b4f86ecdbb8a88de92054698a0028bcb218a3a8", "GraphDriver": { "Data": { "LowerDir": "/var/lib/containers/storage/overlay/63c8cde32bc583ca7971591c23194c1fd072197949f28fcad7f10780ad1af022/diff:/var/lib/containers/storage/overlay/4904a0ac4db0b6060c944fc7c882cba99b79b50f90151785350c0e4ba7f94390/diff:/var/lib/containers/storage/overlay/1822e40336d4b2b31d888f669f443965d4476dae35fd0a67e58eefd9c5ecb9e5/diff:/var/lib/containers/storage/overlay/988a3979591d296974121b0a893ec621ea5e9939ab3cc5e7c194dbe27f7b5677/diff:/var/lib/containers/storage/overlay/ba0315059f870fcb66ebd6517f455f714fb4eeeb4fdee9cefb3d7f59a7b0cdb7/diff:/var/lib/containers/storage/overlay/0dbdd9024cc38ecc226b2c4772718d32ff0abc288b41669d58d506274c16c67a/diff:/var/lib/containers/storage/overlay/f8ddb99f39386eac409b19f8b347a7a10ee91771cfb838c88793ad8947a233f5/diff:/var/lib/containers/storage/overlay/2ecb8e9889676e4b9b129bab56314644612905ce6e6ce1c92d5a86ac8f08f03b/diff:/var/lib/containers/storage/overlay/07e715cdf99b679f1e067f4819d1b7e1ba9ea24229ede4e6b45b9240fe5865f8/diff:/var/lib/containers/storage/overlay/2b83e5699838047f936a3875bcce7fe1b169983bf86785ae7519c5bc488558ae/diff", "UpperDir": "/var/lib/containers/storage/overlay/bb762056c5cde2e3213fe6ccf595176708249829a607b0afc5799075d4a852ec/diff", "WorkDir": "/var/lib/containers/storage/overlay/bb762056c5cde2e3213fe6ccf595176708249829a607b0afc5799075d4a852ec/work" }, "Name": "overlay" }, "History": [ { "created": "2021-12-21T01:24:41.626621944Z", "created_by": "/bin/sh -c #(nop) ADD file:b1650c19c1dd1f4cff71553f2bb0bc949944d0bc24b54c318b2880c14538648a in / " }, { "created": "2021-12-21T01:24:42.056427889Z", "created_by": "/bin/sh -c #(nop) CMD [\"bash\"]", "empty_layer": true }, { "created": "2021-12-21T02:56:59.289452902Z", "created_by": "/bin/sh -c groupadd -r mysql && useradd -r -g mysql mysql" }, { "created": "2021-12-21T02:57:05.692613282Z", "created_by": "/bin/sh -c apt-get update && apt-get install -y --no-install-recommends gnupg dirmngr && rm -rf /var/lib/apt/lists/*" }, { "created": "2021-12-21T02:57:05.921738618Z", "created_by": "/bin/sh -c #(nop) ENV GOSU_VERSION=1.12", "empty_layer": true }, { "created": "2021-12-21T02:57:18.981216995Z", "created_by": "/bin/sh -c set -eux; \tsavedAptMark=\"$(apt-mark showmanual)\"; \tapt-get update; \tapt-get install -y --no-install-recommends ca-certificates wget; \trm -rf /var/lib/apt/lists/*; \tdpkgArch=\"$(dpkg --print-architecture | awk -F- '{ print $NF }')\"; \twget -O /usr/local/bin/gosu \"https://github.com/tianon/gosu/releases/download/$GOSU_VERSION/gosu-$dpkgArch\"; \twget -O /usr/local/bin/gosu.asc \"https://github.com/tianon/gosu/releases/download/$GOSU_VERSION/gosu-$dpkgArch.asc\"; \texport GNUPGHOME=\"$(mktemp -d)\"; \tgpg --batch --keyserver hkps://keys.openpgp.org --recv-keys B42F6819007F00F88E364FD4036A9C25BF357DD4; \tgpg --batch --verify /usr/local/bin/gosu.asc /usr/local/bin/gosu; \tgpgconf --kill all; \trm -rf \"$GNUPGHOME\" /usr/local/bin/gosu.asc; \tapt-mark auto '.*' > /dev/null; \t[ -z \"$savedAptMark\" ] || apt-mark manual $savedAptMark > /dev/null; \tapt-get purge -y --auto-remove -o APT::AutoRemove::RecommendsImportant=false; \tchmod +x /usr/local/bin/gosu; \tgosu --version; \tgosu nobody true" }, { "created": "2021-12-21T02:57:19.781116502Z", "created_by": "/bin/sh -c mkdir /docker-entrypoint-initdb.d" }, { "created": "2021-12-21T02:57:25.555452451Z", "created_by": "/bin/sh -c apt-get update && apt-get install -y --no-install-recommends \t\tpwgen \t\tperl \t\txz-utils \t&& rm -rf /var/lib/apt/lists/*" }, { "created": "2021-12-21T02:57:37.684730941Z", "created_by": "/bin/sh -c set -ex; \tkey='A4A9406876FCBD3C456770C88C718D3B5072E1F5'; \texport GNUPGHOME=\"$(mktemp -d)\"; \tgpg --batch --keyserver keyserver.ubuntu.com --recv-keys \"$key\"; \tgpg --batch --export \"$key\" > /etc/apt/trusted.gpg.d/mysql.gpg; \tgpgconf --kill all; \trm -rf \"$GNUPGHOME\"; \tapt-key list > /dev/null" }, { "created": "2021-12-21T02:57:37.903115972Z", "created_by": "/bin/sh -c #(nop) ENV MYSQL_MAJOR=5.6", "empty_layer": true }, { "created": "2021-12-21T02:57:38.155528708Z", "created_by": "/bin/sh -c #(nop) ENV MYSQL_VERSION=5.6.51-1debian9", "empty_layer": true }, { "created": "2021-12-21T02:57:39.106901365Z", "created_by": "/bin/sh -c echo 'deb http://repo.mysql.com/apt/debian/ stretch mysql-5.6' > /etc/apt/sources.list.d/mysql.list" }, { "created": "2021-12-21T02:57:56.693791892Z", "created_by": "/bin/sh -c { \t\techo mysql-community-server mysql-community-server/data-dir select ''; \t\techo mysql-community-server mysql-community-server/root-pass password ''; \t\techo mysql-community-server mysql-community-server/re-root-pass password ''; \t\techo mysql-community-server mysql-community-server/remove-test-db select false; \t} | debconf-set-selections \t&& apt-get update \t&& apt-get install -y \t\tmysql-server=\"${MYSQL_VERSION}\" \t&& find /etc/mysql/ -name '*.cnf' -print0 \t\t| xargs -0 grep -lZE '^(bind-address|log)' \t\t| xargs -rt -0 sed -Ei 's/^(bind-address|log)/#&/' \t&& echo '[mysqld]\\nskip-host-cache\\nskip-name-resolve' > /etc/mysql/conf.d/docker.cnf \t&& rm -rf /var/lib/apt/lists/* \t&& rm -rf /var/lib/mysql && mkdir -p /var/lib/mysql /var/run/mysqld \t&& chown -R mysql:mysql /var/lib/mysql /var/run/mysqld \t&& chmod 1777 /var/run/mysqld /var/lib/mysql" }, { "created": "2021-12-21T02:57:57.29721744Z", "created_by": "/bin/sh -c #(nop) VOLUME [/var/lib/mysql]", "empty_layer": true }, { "created": "2021-12-21T02:57:57.591077869Z", "created_by": "/bin/sh -c #(nop) COPY file:345a22fe55d3e6783a17075612415413487e7dba27fbf1000a67c7870364b739 in /usr/local/bin/ " }, { "created": "2021-12-21T02:57:58.761392735Z", "created_by": "/bin/sh -c ln -s usr/local/bin/docker-entrypoint.sh /entrypoint.sh # backwards compat" }, { "created": "2021-12-21T02:57:59.04809683Z", "created_by": "/bin/sh -c #(nop) ENTRYPOINT [\"docker-entrypoint.sh\"]", "empty_layer": true }, { "created": "2021-12-21T02:57:59.31392792Z", "created_by": "/bin/sh -c #(nop) EXPOSE 3306", "empty_layer": true }, { "created": "2021-12-21T02:57:59.587337844Z", "created_by": "/bin/sh -c #(nop) CMD [\"mysqld\"]", "empty_layer": true } ], "Id": "dd3b2a5dcb48ff61113592ed5ddd762581be4387c7bc552375a2159422aa6bf5", "Labels": null, "ManifestType": "application/vnd.docker.distribution.manifest.v2+json", "NamesHistory": [ "quay.io/linux-system-roles/mysql:5.6" ], "Os": "linux", "Parent": "", "RepoDigests": [ "quay.io/linux-system-roles/mysql@sha256:2d0419de75076a55e0d0c0680b4f86ecdbb8a88de92054698a0028bcb218a3a8", "quay.io/linux-system-roles/mysql@sha256:4faa0a30c0055742c72daff1482976814f340d4bfb647f0fa2c3562fd1a7abda" ], "RepoTags": [ "quay.io/linux-system-roles/mysql:5.6" ], "RootFS": { "Layers": [ "sha256:2b83e5699838047f936a3875bcce7fe1b169983bf86785ae7519c5bc488558ae", "sha256:2e10295573911964a8acc37c738aed958d0bc385bc51df6f184b196fb592c975", "sha256:d414fdead0b9c934781e5611c016d7c237062c9470db810ef28b321a492dbb9c", "sha256:4085e588967d10dafb76ad52f66f661b41b9008cc16568ce02853de7cbbf7433", "sha256:7ea96a4e341b4d139f5e076e71d19229537292226c3b24ec2d588567667a1191", "sha256:e3dce1c82d4e4142e8e1bf2fb9c6fd2555196d40088167270ff95c9af37b88af", "sha256:2612088e90f60ecb4a13e68adbfb2b1366f82fc8fbf1f0970f8769cf3cd753ee", "sha256:eba393347f896c4028340bf1a40cc5d6b8f21eeb7abf3e79799d842a6babb93f", "sha256:7c5a5c1986b16b0d8ad4eaa8066627910ec7794a657467e8f18d223f1c0034fd", "sha256:49a1ca1cd2b80f84d6616df2b5123458f4bcb065bca28c43d9c3fd5a0480b907", "sha256:7137327a7221ddb123883146beb68003ee31d652ec46ab8bea3ba12c2df8df8f" ], "Type": "layers" }, "Size": 307607194, "User": "", "Version": "20.10.7", "VirtualSize": 307607194 } ], "item": "quay.io/linux-system-roles/mysql:5.6", "podman_actions": [ "/usr/bin/podman image ls quay.io/linux-system-roles/mysql:5.6 --format json", "/usr/bin/podman image exists quay.io/linux-system-roles/mysql:5.6", "/usr/bin/podman pull quay.io/linux-system-roles/mysql:5.6 -q", "/usr/bin/podman inspect dd3b2a5dcb48ff61113592ed5ddd762581be4387c7bc552375a2159422aa6bf5 --format json" ] } TASK [fedora.linux_system_roles.podman : Ensure the quadlet directory is present] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:37 Monday 05 February 2024 04:37:15 +0000 (0:00:07.965) 0:01:38.027 ******* ok: [sut] => { "changed": false, "gid": 0, "group": "root", "mode": "0755", "owner": "root", "path": "/etc/containers/systemd", "secontext": "system_u:object_r:etc_t:s0", "size": 67, "state": "directory", "uid": 0 } TASK [fedora.linux_system_roles.podman : Ensure quadlet file is copied] ******** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:46 Monday 05 February 2024 04:37:16 +0000 (0:00:00.207) 0:01:38.234 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_quadlet_file_src | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Ensure quadlet file content is present] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:56 Monday 05 February 2024 04:37:16 +0000 (0:00:00.018) 0:01:38.253 ******* changed: [sut] => { "changed": true, "checksum": "7bf9312d57274d4dabbd222b01717e2cc0cfa0ed", "dest": "/etc/containers/systemd/quadlet-demo-mysql.container", "gid": 0, "group": "root", "md5sum": "f1cd3875deb45b806389825bb4b5ad4a", "mode": "0644", "owner": "root", "secontext": "system_u:object_r:etc_t:s0", "size": 367, "src": "/root/.ansible/tmp/ansible-tmp-1707107836.1947784-8117-161871808305394/source", "state": "file", "uid": 0 } TASK [fedora.linux_system_roles.podman : Ensure quadlet file is present] ******* task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:68 Monday 05 February 2024 04:37:16 +0000 (0:00:00.522) 0:01:38.776 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_copy_content is skipped", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Reload systemctl] ********************* task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:80 Monday 05 February 2024 04:37:16 +0000 (0:00:00.017) 0:01:38.793 ******* ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.podman : Start service] ************************ task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:108 Monday 05 February 2024 04:37:17 +0000 (0:00:00.469) 0:01:39.263 ******* fatal: [sut]: FAILED! => { "changed": false } MSG: Unable to start service quadlet-demo-mysql.service: Job for quadlet-demo-mysql.service failed because the control process exited with error code. See "systemctl status quadlet-demo-mysql.service" and "journalctl -xe" for details. TASK [Dump journal] ************************************************************ task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/tests/tests_quadlet_demo.yml:131 Monday 05 February 2024 04:37:17 +0000 (0:00:00.525) 0:01:39.789 ******* ok: [sut] => { "changed": false, "cmd": [ "journalctl", "-ex" ], "delta": "0:00:00.034325", "end": "2024-02-05 04:37:17.899278", "rc": 0, "start": "2024-02-05 04:37:17.864953" } STDOUT: -- Logs begin at Mon 2024-02-05 04:23:03 UTC, end at Mon 2024-02-05 04:37:17 UTC. -- Feb 05 04:23:05 localhost.localdomain systemd[1]: Stopped target Initrd Root Device. -- Subject: Unit initrd-root-device.target has finished shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit initrd-root-device.target has finished shutting down. Feb 05 04:23:05 localhost.localdomain systemd[1]: Stopped target Basic System. -- Subject: Unit basic.target has finished shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit basic.target has finished shutting down. Feb 05 04:23:05 localhost.localdomain systemd[1]: Stopped target Sockets. -- Subject: Unit sockets.target has finished shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit sockets.target has finished shutting down. Feb 05 04:23:05 localhost.localdomain systemd[1]: Stopped target System Initialization. -- Subject: Unit sysinit.target has finished shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit sysinit.target has finished shutting down. Feb 05 04:23:05 localhost.localdomain systemd[1]: systemd-tmpfiles-setup.service: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit systemd-tmpfiles-setup.service has successfully entered the 'dead' state. Feb 05 04:23:05 localhost.localdomain systemd[1]: Stopped Create Volatile Files and Directories. -- Subject: Unit systemd-tmpfiles-setup.service has finished shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit systemd-tmpfiles-setup.service has finished shutting down. Feb 05 04:23:05 localhost.localdomain systemd[1]: Stopped target Local File Systems. -- Subject: Unit local-fs.target has finished shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit local-fs.target has finished shutting down. Feb 05 04:23:05 localhost.localdomain systemd[1]: Stopped target Paths. -- Subject: Unit paths.target has finished shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit paths.target has finished shutting down. Feb 05 04:23:05 localhost.localdomain systemd[1]: Stopped target Slices. -- Subject: Unit slices.target has finished shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit slices.target has finished shutting down. Feb 05 04:23:05 localhost.localdomain systemd[1]: Stopped target Remote File Systems. -- Subject: Unit remote-fs.target has finished shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit remote-fs.target has finished shutting down. Feb 05 04:23:05 localhost.localdomain systemd[1]: Stopped target Remote File Systems (Pre). -- Subject: Unit remote-fs-pre.target has finished shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit remote-fs-pre.target has finished shutting down. Feb 05 04:23:05 localhost.localdomain systemd[1]: dracut-initqueue.service: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit dracut-initqueue.service has successfully entered the 'dead' state. Feb 05 04:23:05 localhost.localdomain systemd[1]: Stopped dracut initqueue hook. -- Subject: Unit dracut-initqueue.service has finished shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit dracut-initqueue.service has finished shutting down. Feb 05 04:23:05 localhost.localdomain systemd[1]: systemd-udev-trigger.service: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit systemd-udev-trigger.service has successfully entered the 'dead' state. Feb 05 04:23:05 localhost.localdomain systemd[1]: Stopped udev Coldplug all Devices. -- Subject: Unit systemd-udev-trigger.service has finished shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit systemd-udev-trigger.service has finished shutting down. Feb 05 04:23:05 localhost.localdomain systemd[1]: Stopped target Timers. -- Subject: Unit timers.target has finished shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit timers.target has finished shutting down. Feb 05 04:23:05 localhost.localdomain systemd[1]: Stopping udev Kernel Device Manager... -- Subject: Unit systemd-udevd.service has begun shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit systemd-udevd.service has begun shutting down. Feb 05 04:23:05 localhost.localdomain systemd[1]: Starting Plymouth switch root service... -- Subject: Unit plymouth-switch-root.service has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit plymouth-switch-root.service has begun starting up. Feb 05 04:23:05 localhost.localdomain systemd[1]: systemd-sysctl.service: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit systemd-sysctl.service has successfully entered the 'dead' state. Feb 05 04:23:05 localhost.localdomain systemd[1]: Stopped Apply Kernel Variables. -- Subject: Unit systemd-sysctl.service has finished shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit systemd-sysctl.service has finished shutting down. Feb 05 04:23:05 localhost.localdomain systemd[1]: Stopped target Swap. -- Subject: Unit swap.target has finished shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit swap.target has finished shutting down. Feb 05 04:23:05 localhost.localdomain systemd[1]: initrd-cleanup.service: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit initrd-cleanup.service has successfully entered the 'dead' state. Feb 05 04:23:05 localhost.localdomain systemd[1]: Started Cleaning Up and Shutting Down Daemons. -- Subject: Unit initrd-cleanup.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit initrd-cleanup.service has finished starting up. -- -- The start-up result is done. Feb 05 04:23:05 localhost.localdomain systemd[1]: systemd-udevd.service: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit systemd-udevd.service has successfully entered the 'dead' state. Feb 05 04:23:05 localhost.localdomain systemd[1]: Stopped udev Kernel Device Manager. -- Subject: Unit systemd-udevd.service has finished shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit systemd-udevd.service has finished shutting down. Feb 05 04:23:05 localhost.localdomain systemd[1]: systemd-tmpfiles-setup-dev.service: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit systemd-tmpfiles-setup-dev.service has successfully entered the 'dead' state. Feb 05 04:23:05 localhost.localdomain systemd[1]: Stopped Create Static Device Nodes in /dev. -- Subject: Unit systemd-tmpfiles-setup-dev.service has finished shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit systemd-tmpfiles-setup-dev.service has finished shutting down. Feb 05 04:23:05 localhost.localdomain systemd[1]: kmod-static-nodes.service: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit kmod-static-nodes.service has successfully entered the 'dead' state. Feb 05 04:23:05 localhost.localdomain systemd[1]: Stopped Create list of required static device nodes for the current kernel. -- Subject: Unit kmod-static-nodes.service has finished shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit kmod-static-nodes.service has finished shutting down. Feb 05 04:23:05 localhost.localdomain systemd[1]: dracut-pre-udev.service: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit dracut-pre-udev.service has successfully entered the 'dead' state. Feb 05 04:23:05 localhost.localdomain systemd[1]: Stopped dracut pre-udev hook. -- Subject: Unit dracut-pre-udev.service has finished shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit dracut-pre-udev.service has finished shutting down. Feb 05 04:23:05 localhost.localdomain systemd[1]: dracut-cmdline.service: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit dracut-cmdline.service has successfully entered the 'dead' state. Feb 05 04:23:05 localhost.localdomain systemd[1]: Stopped dracut cmdline hook. -- Subject: Unit dracut-cmdline.service has finished shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit dracut-cmdline.service has finished shutting down. Feb 05 04:23:05 localhost.localdomain systemd[1]: systemd-udevd-control.socket: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit systemd-udevd-control.socket has successfully entered the 'dead' state. Feb 05 04:23:05 localhost.localdomain systemd[1]: Closed udev Control Socket. -- Subject: Unit systemd-udevd-control.socket has finished shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit systemd-udevd-control.socket has finished shutting down. Feb 05 04:23:05 localhost.localdomain systemd[1]: systemd-udevd-kernel.socket: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit systemd-udevd-kernel.socket has successfully entered the 'dead' state. Feb 05 04:23:05 localhost.localdomain systemd[1]: Closed udev Kernel Socket. -- Subject: Unit systemd-udevd-kernel.socket has finished shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit systemd-udevd-kernel.socket has finished shutting down. Feb 05 04:23:05 localhost.localdomain systemd[1]: Starting Cleanup udevd DB... -- Subject: Unit initrd-udevadm-cleanup-db.service has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit initrd-udevadm-cleanup-db.service has begun starting up. Feb 05 04:23:05 localhost.localdomain systemd[1]: initrd-udevadm-cleanup-db.service: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit initrd-udevadm-cleanup-db.service has successfully entered the 'dead' state. Feb 05 04:23:05 localhost.localdomain systemd[1]: Started Cleanup udevd DB. -- Subject: Unit initrd-udevadm-cleanup-db.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit initrd-udevadm-cleanup-db.service has finished starting up. -- -- The start-up result is done. Feb 05 04:23:06 localhost.localdomain systemd[1]: systemd-vconsole-setup.service: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit systemd-vconsole-setup.service has successfully entered the 'dead' state. Feb 05 04:23:06 localhost.localdomain systemd[1]: Started Setup Virtual Console. -- Subject: Unit systemd-vconsole-setup.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit systemd-vconsole-setup.service has finished starting up. -- -- The start-up result is done. Feb 05 04:23:06 localhost.localdomain systemd[1]: Reached target Switch Root. -- Subject: Unit initrd-switch-root.target has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit initrd-switch-root.target has finished starting up. -- -- The start-up result is done. Feb 05 04:23:06 localhost.localdomain systemd[1]: Started Plymouth switch root service. -- Subject: Unit plymouth-switch-root.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit plymouth-switch-root.service has finished starting up. -- -- The start-up result is done. Feb 05 04:23:06 localhost.localdomain systemd[1]: Starting Switch Root... -- Subject: Unit initrd-switch-root.service has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit initrd-switch-root.service has begun starting up. Feb 05 04:23:06 localhost.localdomain systemd[1]: Switching root. Feb 05 04:23:06 localhost.localdomain systemd-journald[220]: Journal stopped -- Subject: The journal has been stopped -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The system journal process has shut down and closed all currently -- active journal files. Feb 05 04:23:12 localhost.localdomain kernel: printk: systemd: 16 output lines suppressed due to ratelimiting Feb 05 04:23:12 localhost.localdomain kernel: audit: type=1404 audit(1707106988.635:2): enforcing=1 old_enforcing=0 auid=4294967295 ses=4294967295 enabled=1 old-enabled=1 lsm=selinux res=1 Feb 05 04:23:12 localhost.localdomain kernel: SELinux: policy capability network_peer_controls=1 Feb 05 04:23:12 localhost.localdomain kernel: SELinux: policy capability open_perms=1 Feb 05 04:23:12 localhost.localdomain kernel: SELinux: policy capability extended_socket_class=1 Feb 05 04:23:12 localhost.localdomain kernel: SELinux: policy capability always_check_network=0 Feb 05 04:23:12 localhost.localdomain kernel: SELinux: policy capability cgroup_seclabel=1 Feb 05 04:23:12 localhost.localdomain kernel: SELinux: policy capability nnp_nosuid_transition=1 Feb 05 04:23:12 localhost.localdomain kernel: audit: type=1403 audit(1707106989.503:3): auid=4294967295 ses=4294967295 lsm=selinux res=1 Feb 05 04:23:12 localhost.localdomain systemd[1]: Successfully loaded SELinux policy in 870.092ms. Feb 05 04:23:12 localhost.localdomain systemd[1]: Relabelled /dev, /run and /sys/fs/cgroup in 18.668ms. Feb 05 04:23:12 localhost.localdomain systemd[1]: systemd 239 (239-79.el8) running in system mode. (+PAM +AUDIT +SELINUX +IMA -APPARMOR +SMACK +SYSVINIT +UTMP +LIBCRYPTSETUP +GCRYPT +GNUTLS +ACL +XZ +LZ4 +SECCOMP +BLKID +ELFUTILS +KMOD +IDN2 -IDN +PCRE2 default-hierarchy=legacy) Feb 05 04:23:12 localhost.localdomain systemd[1]: Detected virtualization xen. Feb 05 04:23:12 localhost.localdomain systemd[1]: Detected architecture x86-64. Feb 05 04:23:12 localhost.localdomain systemd[1]: Set hostname to . Feb 05 04:23:12 localhost.localdomain systemd[1]: Initializing machine ID from random generator. Feb 05 04:23:12 localhost.localdomain systemd[1]: Installed transient /etc/machine-id file. Feb 05 04:23:12 localhost.localdomain systemd[1]: /usr/lib/systemd/system/cloud-init.target:15: Unknown lvalue 'ConditionEnvironment' in section 'Unit' Feb 05 04:23:12 localhost.localdomain systemd[1]: /usr/lib/systemd/system/cloud-init.service:18: Unknown lvalue 'ConditionEnvironment' in section 'Unit' Feb 05 04:23:12 localhost.localdomain systemd[1]: /usr/lib/systemd/system/cloud-init-local.service:12: Unknown lvalue 'ConditionEnvironment' in section 'Unit' Feb 05 04:23:12 localhost.localdomain systemd[1]: /usr/lib/systemd/system/cloud-final.service:7: Unknown lvalue 'ConditionEnvironment' in section 'Unit' Feb 05 04:23:12 localhost.localdomain systemd-journald[504]: Journal started -- Subject: The journal has been started -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The system journal process has started up, opened the journal -- files for writing and is now ready to process requests. Feb 05 04:23:12 localhost.localdomain systemd-journald[504]: Runtime journal (/run/log/journal/8b94584e0ddd46018c56198f0cd9d615) is 8.0M, max 176.9M, 168.9M free. -- Subject: Disk space used by the journal -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Runtime journal (/run/log/journal/8b94584e0ddd46018c56198f0cd9d615) is currently using 8.0M. -- Maximum allowed usage is set to 176.9M. -- Leaving at least 265.4M free (of currently available 1.7G of disk space). -- Enforced usage limit is thus 176.9M, of which 168.9M are still available. -- -- The limits controlling how much disk space is used by the journal may -- be configured with SystemMaxUse=, SystemKeepFree=, SystemMaxFileSize=, -- RuntimeMaxUse=, RuntimeKeepFree=, RuntimeMaxFileSize= settings in -- /etc/systemd/journald.conf. See journald.conf(5) for details. Feb 05 04:23:11 localhost.localdomain systemd[1]: systemd-journald.service: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit systemd-journald.service has successfully entered the 'dead' state. Feb 05 04:23:12 localhost.localdomain systemd[1]: Mounted POSIX Message Queue File System. -- Subject: Unit dev-mqueue.mount has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit dev-mqueue.mount has finished starting up. -- -- The start-up result is done. Feb 05 04:23:12 localhost.localdomain systemd[1]: Mounted Kernel Debug File System. -- Subject: Unit sys-kernel-debug.mount has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit sys-kernel-debug.mount has finished starting up. -- -- The start-up result is done. Feb 05 04:23:12 localhost.localdomain systemd[1]: Mounted Huge Pages File System. -- Subject: Unit dev-hugepages.mount has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit dev-hugepages.mount has finished starting up. -- -- The start-up result is done. Feb 05 04:23:12 localhost.localdomain systemd[1]: Started Read and set NIS domainname from /etc/sysconfig/network. -- Subject: Unit nis-domainname.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit nis-domainname.service has finished starting up. -- -- The start-up result is done. Feb 05 04:23:12 localhost.localdomain systemd[1]: Started Create list of required static device nodes for the current kernel. -- Subject: Unit kmod-static-nodes.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit kmod-static-nodes.service has finished starting up. -- -- The start-up result is done. Feb 05 04:23:12 localhost.localdomain systemd[1]: Started Remount Root and Kernel File Systems. -- Subject: Unit systemd-remount-fs.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit systemd-remount-fs.service has finished starting up. -- -- The start-up result is done. Feb 05 04:23:12 localhost.localdomain systemd[1]: Starting Flush Journal to Persistent Storage... -- Subject: Unit systemd-journal-flush.service has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit systemd-journal-flush.service has begun starting up. Feb 05 04:23:12 localhost.localdomain systemd[1]: Starting Load/Save Random Seed... -- Subject: Unit systemd-random-seed.service has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit systemd-random-seed.service has begun starting up. Feb 05 04:23:12 localhost.localdomain systemd[1]: Starting Create Static Device Nodes in /dev... -- Subject: Unit systemd-tmpfiles-setup-dev.service has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit systemd-tmpfiles-setup-dev.service has begun starting up. Feb 05 04:23:12 localhost.localdomain systemd[1]: Started udev Coldplug all Devices. -- Subject: Unit systemd-udev-trigger.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit systemd-udev-trigger.service has finished starting up. -- -- The start-up result is done. Feb 05 04:23:12 localhost.localdomain systemd[1]: Started Load/Save Random Seed. -- Subject: Unit systemd-random-seed.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit systemd-random-seed.service has finished starting up. -- -- The start-up result is done. Feb 05 04:23:12 localhost.localdomain systemd-journald[504]: Runtime journal (/run/log/journal/8b94584e0ddd46018c56198f0cd9d615) is 8.0M, max 176.9M, 168.9M free. -- Subject: Disk space used by the journal -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Runtime journal (/run/log/journal/8b94584e0ddd46018c56198f0cd9d615) is currently using 8.0M. -- Maximum allowed usage is set to 176.9M. -- Leaving at least 265.4M free (of currently available 1.7G of disk space). -- Enforced usage limit is thus 176.9M, of which 168.9M are still available. -- -- The limits controlling how much disk space is used by the journal may -- be configured with SystemMaxUse=, SystemKeepFree=, SystemMaxFileSize=, -- RuntimeMaxUse=, RuntimeKeepFree=, RuntimeMaxFileSize= settings in -- /etc/systemd/journald.conf. See journald.conf(5) for details. Feb 05 04:23:12 localhost.localdomain systemd[1]: Started Flush Journal to Persistent Storage. -- Subject: Unit systemd-journal-flush.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit systemd-journal-flush.service has finished starting up. -- -- The start-up result is done. Feb 05 04:23:12 localhost.localdomain systemd[1]: Started Apply Kernel Variables. -- Subject: Unit systemd-sysctl.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit systemd-sysctl.service has finished starting up. -- -- The start-up result is done. Feb 05 04:23:12 localhost.localdomain systemd[1]: Started Create Static Device Nodes in /dev. -- Subject: Unit systemd-tmpfiles-setup-dev.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit systemd-tmpfiles-setup-dev.service has finished starting up. -- -- The start-up result is done. Feb 05 04:23:12 localhost.localdomain systemd[1]: Starting udev Kernel Device Manager... -- Subject: Unit systemd-udevd.service has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit systemd-udevd.service has begun starting up. Feb 05 04:23:12 localhost.localdomain systemd[1]: Reached target Local File Systems (Pre). -- Subject: Unit local-fs-pre.target has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit local-fs-pre.target has finished starting up. -- -- The start-up result is done. Feb 05 04:23:12 localhost.localdomain systemd[1]: Reached target Local File Systems. -- Subject: Unit local-fs.target has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit local-fs.target has finished starting up. -- -- The start-up result is done. Feb 05 04:23:12 localhost.localdomain systemd[1]: Starting Tell Plymouth To Write Out Runtime Data... -- Subject: Unit plymouth-read-write.service has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit plymouth-read-write.service has begun starting up. Feb 05 04:23:12 localhost.localdomain systemd[1]: Starting Import network configuration from initramfs... -- Subject: Unit import-state.service has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit import-state.service has begun starting up. Feb 05 04:23:12 localhost.localdomain systemd[1]: Starting Commit a transient machine-id on disk... -- Subject: Unit systemd-machine-id-commit.service has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit systemd-machine-id-commit.service has begun starting up. Feb 05 04:23:12 localhost.localdomain systemd[1]: Started Commit a transient machine-id on disk. -- Subject: Unit systemd-machine-id-commit.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit systemd-machine-id-commit.service has finished starting up. -- -- The start-up result is done. Feb 05 04:23:12 localhost.localdomain systemd[1]: Received SIGRTMIN+20 from PID 339 (plymouthd). Feb 05 04:23:12 localhost.localdomain systemd[1]: Started Tell Plymouth To Write Out Runtime Data. -- Subject: Unit plymouth-read-write.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit plymouth-read-write.service has finished starting up. -- -- The start-up result is done. Feb 05 04:23:12 localhost.localdomain systemd[1]: systemd-vconsole-setup.service: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit systemd-vconsole-setup.service has successfully entered the 'dead' state. Feb 05 04:23:12 localhost.localdomain systemd[1]: Started Setup Virtual Console. -- Subject: Unit systemd-vconsole-setup.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit systemd-vconsole-setup.service has finished starting up. -- -- The start-up result is done. Feb 05 04:23:12 localhost.localdomain systemd[1]: Started Import network configuration from initramfs. -- Subject: Unit import-state.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit import-state.service has finished starting up. -- -- The start-up result is done. Feb 05 04:23:12 localhost.localdomain systemd[1]: Starting Create Volatile Files and Directories... -- Subject: Unit systemd-tmpfiles-setup.service has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit systemd-tmpfiles-setup.service has begun starting up. Feb 05 04:23:12 localhost.localdomain systemd-udevd[535]: Network interface NamePolicy= disabled on kernel command line, ignoring. Feb 05 04:23:12 localhost.localdomain systemd[1]: Started Create Volatile Files and Directories. -- Subject: Unit systemd-tmpfiles-setup.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit systemd-tmpfiles-setup.service has finished starting up. -- -- The start-up result is done. Feb 05 04:23:12 localhost.localdomain systemd[1]: Starting RPC Bind... -- Subject: Unit rpcbind.service has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit rpcbind.service has begun starting up. Feb 05 04:23:12 localhost.localdomain systemd[1]: Starting Security Auditing Service... -- Subject: Unit auditd.service has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit auditd.service has begun starting up. Feb 05 04:23:12 localhost.localdomain auditd[559]: No plugins found, not dispatching events Feb 05 04:23:12 localhost.localdomain auditd[559]: Init complete, auditd 3.1.2 listening for events (startup state enable) Feb 05 04:23:12 localhost.localdomain systemd[1]: Started RPC Bind. -- Subject: Unit rpcbind.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit rpcbind.service has finished starting up. -- -- The start-up result is done. Feb 05 04:23:12 localhost.localdomain systemd[1]: Started udev Kernel Device Manager. -- Subject: Unit systemd-udevd.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit systemd-udevd.service has finished starting up. -- -- The start-up result is done. Feb 05 04:23:13 localhost.localdomain systemd[1]: etc-machine\x2did.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit etc-machine\x2did.mount has successfully entered the 'dead' state. Feb 05 04:23:13 localhost.localdomain systemd[1]: Mounting RPC Pipe File System... -- Subject: Unit var-lib-nfs-rpc_pipefs.mount has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit var-lib-nfs-rpc_pipefs.mount has begun starting up. Feb 05 04:23:13 localhost.localdomain augenrules[562]: /sbin/augenrules: No change Feb 05 04:23:13 localhost.localdomain augenrules[593]: No rules Feb 05 04:23:13 localhost.localdomain augenrules[593]: enabled 1 Feb 05 04:23:13 localhost.localdomain augenrules[593]: failure 1 Feb 05 04:23:13 localhost.localdomain augenrules[593]: pid 559 Feb 05 04:23:13 localhost.localdomain augenrules[593]: rate_limit 0 Feb 05 04:23:13 localhost.localdomain augenrules[593]: backlog_limit 8192 Feb 05 04:23:13 localhost.localdomain augenrules[593]: lost 0 Feb 05 04:23:13 localhost.localdomain augenrules[593]: backlog 3 Feb 05 04:23:13 localhost.localdomain augenrules[593]: backlog_wait_time 60000 Feb 05 04:23:13 localhost.localdomain augenrules[593]: backlog_wait_time_actual 0 Feb 05 04:23:13 localhost.localdomain augenrules[593]: enabled 1 Feb 05 04:23:13 localhost.localdomain augenrules[593]: failure 1 Feb 05 04:23:13 localhost.localdomain augenrules[593]: pid 559 Feb 05 04:23:13 localhost.localdomain augenrules[593]: rate_limit 0 Feb 05 04:23:13 localhost.localdomain augenrules[593]: backlog_limit 8192 Feb 05 04:23:13 localhost.localdomain augenrules[593]: lost 0 Feb 05 04:23:13 localhost.localdomain augenrules[593]: backlog 0 Feb 05 04:23:13 localhost.localdomain augenrules[593]: backlog_wait_time 60000 Feb 05 04:23:13 localhost.localdomain augenrules[593]: backlog_wait_time_actual 0 Feb 05 04:23:13 localhost.localdomain augenrules[593]: enabled 1 Feb 05 04:23:13 localhost.localdomain augenrules[593]: failure 1 Feb 05 04:23:13 localhost.localdomain augenrules[593]: pid 559 Feb 05 04:23:13 localhost.localdomain augenrules[593]: rate_limit 0 Feb 05 04:23:13 localhost.localdomain augenrules[593]: backlog_limit 8192 Feb 05 04:23:13 localhost.localdomain augenrules[593]: lost 0 Feb 05 04:23:13 localhost.localdomain augenrules[593]: backlog 0 Feb 05 04:23:13 localhost.localdomain augenrules[593]: backlog_wait_time 60000 Feb 05 04:23:13 localhost.localdomain augenrules[593]: backlog_wait_time_actual 0 Feb 05 04:23:13 localhost.localdomain systemd[1]: Started Security Auditing Service. -- Subject: Unit auditd.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit auditd.service has finished starting up. -- -- The start-up result is done. Feb 05 04:23:13 localhost.localdomain systemd[1]: Starting Update UTMP about System Boot/Shutdown... -- Subject: Unit systemd-update-utmp.service has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit systemd-update-utmp.service has begun starting up. Feb 05 04:23:13 localhost.localdomain systemd[1]: Started Update UTMP about System Boot/Shutdown. -- Subject: Unit systemd-update-utmp.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit systemd-update-utmp.service has finished starting up. -- -- The start-up result is done. Feb 05 04:23:13 localhost.localdomain systemd[1]: Reached target System Initialization. -- Subject: Unit sysinit.target has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit sysinit.target has finished starting up. -- -- The start-up result is done. Feb 05 04:23:13 localhost.localdomain systemd[1]: Listening on D-Bus System Message Bus Socket. -- Subject: Unit dbus.socket has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit dbus.socket has finished starting up. -- -- The start-up result is done. Feb 05 04:23:13 localhost.localdomain systemd[1]: Started daily update of the root trust anchor for DNSSEC. -- Subject: Unit unbound-anchor.timer has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit unbound-anchor.timer has finished starting up. -- -- The start-up result is done. Feb 05 04:23:13 localhost.localdomain systemd[1]: Started Daily Cleanup of Temporary Directories. -- Subject: Unit systemd-tmpfiles-clean.timer has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit systemd-tmpfiles-clean.timer has finished starting up. -- -- The start-up result is done. Feb 05 04:23:13 localhost.localdomain systemd[1]: Started dnf makecache --timer. -- Subject: Unit dnf-makecache.timer has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit dnf-makecache.timer has finished starting up. -- -- The start-up result is done. Feb 05 04:23:13 localhost.localdomain systemd[1]: Reached target Timers. -- Subject: Unit timers.target has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit timers.target has finished starting up. -- -- The start-up result is done. Feb 05 04:23:13 localhost.localdomain kernel: input: PC Speaker as /devices/platform/pcspkr/input/input5 Feb 05 04:23:13 localhost.localdomain kernel: piix4_smbus 0000:00:01.3: SMBus base address uninitialized - upgrade BIOS or use force_addr=0xaddr Feb 05 04:23:13 localhost.localdomain systemd[1]: Listening on SSSD Kerberos Cache Manager responder socket. -- Subject: Unit sssd-kcm.socket has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit sssd-kcm.socket has finished starting up. -- -- The start-up result is done. Feb 05 04:23:13 localhost.localdomain systemd[1]: Reached target Sockets. -- Subject: Unit sockets.target has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit sockets.target has finished starting up. -- -- The start-up result is done. Feb 05 04:23:13 localhost.localdomain systemd[1]: Reached target Basic System. -- Subject: Unit basic.target has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit basic.target has finished starting up. -- -- The start-up result is done. Feb 05 04:23:13 localhost.localdomain systemd-udevd[587]: link_config: autonegotiation is unset or enabled, the speed and duplex are not writable. Feb 05 04:23:13 localhost.localdomain kernel: RAPL PMU: API unit is 2^-32 Joules, 0 fixed counters, 655360 ms ovfl timer Feb 05 04:23:13 localhost.localdomain kernel: RPC: Registered named UNIX socket transport module. Feb 05 04:23:13 localhost.localdomain kernel: RPC: Registered udp transport module. Feb 05 04:23:13 localhost.localdomain kernel: RPC: Registered tcp transport module. Feb 05 04:23:13 localhost.localdomain kernel: RPC: Registered tcp NFSv4.1 backchannel transport module. Feb 05 04:23:13 localhost.localdomain systemd[1]: Starting NTP client/server... -- Subject: Unit chronyd.service has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit chronyd.service has begun starting up. Feb 05 04:23:13 localhost.localdomain systemd[1]: Starting Restore /run/initramfs on shutdown... -- Subject: Unit dracut-shutdown.service has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit dracut-shutdown.service has begun starting up. Feb 05 04:23:13 localhost.localdomain systemd[1]: Started Hardware RNG Entropy Gatherer Daemon. -- Subject: Unit rngd.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit rngd.service has finished starting up. -- -- The start-up result is done. Feb 05 04:23:13 localhost.localdomain systemd[1]: Started irqbalance daemon. -- Subject: Unit irqbalance.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit irqbalance.service has finished starting up. -- -- The start-up result is done. Feb 05 04:23:13 localhost.localdomain systemd[1]: Reached target User and Group Name Lookups. -- Subject: Unit nss-user-lookup.target has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit nss-user-lookup.target has finished starting up. -- -- The start-up result is done. Feb 05 04:23:13 localhost.localdomain systemd[1]: Started D-Bus System Message Bus. -- Subject: Unit dbus.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit dbus.service has finished starting up. -- -- The start-up result is done. Feb 05 04:23:13 localhost.localdomain systemd[1]: Reached target sshd-keygen.target. -- Subject: Unit sshd-keygen.target has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit sshd-keygen.target has finished starting up. -- -- The start-up result is done. Feb 05 04:23:13 localhost.localdomain systemd[1]: Starting Login Service... -- Subject: Unit systemd-logind.service has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit systemd-logind.service has begun starting up. Feb 05 04:23:13 localhost.localdomain systemd[1]: Starting Initial cloud-init job (pre-networking)... -- Subject: Unit cloud-init-local.service has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit cloud-init-local.service has begun starting up. Feb 05 04:23:13 localhost.localdomain systemd[1]: Mounted RPC Pipe File System. -- Subject: Unit var-lib-nfs-rpc_pipefs.mount has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit var-lib-nfs-rpc_pipefs.mount has finished starting up. -- -- The start-up result is done. Feb 05 04:23:13 localhost.localdomain systemd[1]: Started Restore /run/initramfs on shutdown. -- Subject: Unit dracut-shutdown.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit dracut-shutdown.service has finished starting up. -- -- The start-up result is done. Feb 05 04:23:13 localhost.localdomain systemd[1]: Starting update of the root trust anchor for DNSSEC validation in unbound... -- Subject: Unit unbound-anchor.service has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit unbound-anchor.service has begun starting up. Feb 05 04:23:13 localhost.localdomain systemd[1]: Reached target rpc_pipefs.target. -- Subject: Unit rpc_pipefs.target has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit rpc_pipefs.target has finished starting up. -- -- The start-up result is done. Feb 05 04:23:13 localhost.localdomain kernel: EDAC sbridge: Seeking for: PCI ID 8086:2fa0 Feb 05 04:23:13 localhost.localdomain kernel: EDAC sbridge: Ver: 1.1.2 Feb 05 04:23:13 localhost.localdomain systemd-logind[629]: New seat seat0. -- Subject: A new seat seat0 is now available -- Defined-By: systemd -- Support: https://access.redhat.com/support -- Documentation: https://www.freedesktop.org/wiki/Software/systemd/multiseat -- -- A new seat seat0 has been configured and is now available. Feb 05 04:23:13 localhost.localdomain systemd-logind[629]: Watching system buttons on /dev/input/event0 (Power Button) Feb 05 04:23:13 localhost.localdomain systemd-logind[629]: Watching system buttons on /dev/input/event1 (Sleep Button) Feb 05 04:23:13 localhost.localdomain systemd-logind[629]: Watching system buttons on /dev/input/event2 (AT Translated Set 2 keyboard) Feb 05 04:23:13 localhost.localdomain systemd[1]: Started Login Service. -- Subject: Unit systemd-logind.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit systemd-logind.service has finished starting up. -- -- The start-up result is done. Feb 05 04:23:13 localhost.localdomain rngd[625]: Disabling 7: PKCS11 Entropy generator (pkcs11) Feb 05 04:23:13 localhost.localdomain rngd[625]: Disabling 5: NIST Network Entropy Beacon (nist) Feb 05 04:23:13 localhost.localdomain rngd[625]: Disabling 9: Qrypt quantum entropy beacon (qrypt) Feb 05 04:23:13 localhost.localdomain rngd[625]: Initializing available sources Feb 05 04:23:13 localhost.localdomain rngd[625]: [hwrng ]: Initialization Failed Feb 05 04:23:13 localhost.localdomain rngd[625]: [rdrand]: Enabling RDRAND rng support Feb 05 04:23:13 localhost.localdomain rngd[625]: [rdrand]: Initialized Feb 05 04:23:13 localhost.localdomain rngd[625]: [jitter]: JITTER timeout set to 5 sec Feb 05 04:23:13 localhost.localdomain rngd[625]: [jitter]: Initializing AES buffer Feb 05 04:23:14 localhost.localdomain systemd-udevd[579]: Using default interface naming scheme 'rhel-8.0'. Feb 05 04:23:14 localhost.localdomain chronyd[641]: chronyd version 4.2 starting (+CMDMON +NTP +REFCLOCK +RTC +PRIVDROP +SCFILTER +SIGND +ASYNCDNS +NTS +SECHASH +IPV6 +DEBUG) Feb 05 04:23:14 localhost.localdomain chronyd[641]: Frequency 0.000 +/- 1000000.000 ppm read from /var/lib/chrony/drift Feb 05 04:23:14 localhost.localdomain systemd-udevd[579]: link_config: autonegotiation is unset or enabled, the speed and duplex are not writable. Feb 05 04:23:14 localhost.localdomain chronyd[641]: Using right/UTC timezone to obtain leap second data Feb 05 04:23:14 localhost.localdomain systemd[1]: unbound-anchor.service: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit unbound-anchor.service has successfully entered the 'dead' state. Feb 05 04:23:14 localhost.localdomain systemd[1]: Started update of the root trust anchor for DNSSEC validation in unbound. -- Subject: Unit unbound-anchor.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit unbound-anchor.service has finished starting up. -- -- The start-up result is done. Feb 05 04:23:14 localhost.localdomain systemd[1]: Started NTP client/server. -- Subject: Unit chronyd.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit chronyd.service has finished starting up. -- -- The start-up result is done. Feb 05 04:23:18 localhost.localdomain rngd[625]: [jitter]: Unable to obtain AES key, disabling JITTER source Feb 05 04:23:18 localhost.localdomain rngd[625]: [jitter]: Initialization Failed Feb 05 04:23:18 localhost.localdomain rngd[625]: Process privileges have been dropped to 2:2 Feb 05 04:23:19 localhost.localdomain dhclient[654]: Internet Systems Consortium DHCP Client 4.3.6 Feb 05 04:23:19 localhost.localdomain dhclient[654]: Copyright 2004-2017 Internet Systems Consortium. Feb 05 04:23:19 localhost.localdomain dhclient[654]: All rights reserved. Feb 05 04:23:19 localhost.localdomain dhclient[654]: For info, please visit https://www.isc.org/software/dhcp/ Feb 05 04:23:19 localhost.localdomain dhclient[654]: Feb 05 04:23:19 localhost.localdomain dhclient[654]: Listening on LPF/eth0/12:86:38:a3:1b:25 Feb 05 04:23:19 localhost.localdomain dhclient[654]: Sending on LPF/eth0/12:86:38:a3:1b:25 Feb 05 04:23:19 localhost.localdomain dhclient[654]: Sending on Socket/fallback Feb 05 04:23:19 localhost.localdomain dhclient[654]: Created duid "\000\004\213\224XN\015\335F\001\214V\031\217\014\331\326\025". Feb 05 04:23:19 localhost.localdomain dhclient[654]: DHCPDISCOVER on eth0 to 255.255.255.255 port 67 interval 8 (xid=0xb6964822) Feb 05 04:23:19 localhost.localdomain dhclient[654]: DHCPREQUEST on eth0 to 255.255.255.255 port 67 (xid=0xb6964822) Feb 05 04:23:19 localhost.localdomain dhclient[654]: DHCPOFFER from 10.31.8.1 Feb 05 04:23:19 localhost.localdomain dhclient[654]: DHCPACK from 10.31.8.1 (xid=0xb6964822) Feb 05 04:23:19 localhost.localdomain dhclient[654]: bound to 10.31.9.239 -- renewal in 1430 seconds. Feb 05 04:23:19 localhost.localdomain dbus-daemon[628]: [system] Activating via systemd: service name='org.freedesktop.hostname1' unit='dbus-org.freedesktop.hostname1.service' requested by ':1.3' (uid=0 pid=669 comm="hostnamectl set-hostname ip-10-31-9-239.us-east-1." label="system_u:system_r:cloud_init_t:s0") Feb 05 04:23:19 localhost.localdomain systemd[1]: Starting Hostname Service... -- Subject: Unit systemd-hostnamed.service has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit systemd-hostnamed.service has begun starting up. Feb 05 04:23:19 localhost.localdomain dbus-daemon[628]: [system] Successfully activated service 'org.freedesktop.hostname1' Feb 05 04:23:19 localhost.localdomain systemd[1]: Started Hostname Service. -- Subject: Unit systemd-hostnamed.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit systemd-hostnamed.service has finished starting up. -- -- The start-up result is done. Feb 05 04:23:19 ip-10-31-9-239.us-east-1.aws.redhat.com systemd-hostnamed[670]: Changed static host name to 'ip-10-31-9-239.us-east-1.aws.redhat.com' Feb 05 04:23:19 ip-10-31-9-239.us-east-1.aws.redhat.com systemd-hostnamed[670]: Changed host name to 'ip-10-31-9-239.us-east-1.aws.redhat.com' Feb 05 04:23:19 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[652]: Cloud-init v. 23.4-2.el8 running 'init-local' at Mon, 05 Feb 2024 04:23:18 +0000. Up 15.66 seconds. Feb 05 04:23:19 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: Started Initial cloud-init job (pre-networking). -- Subject: Unit cloud-init-local.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit cloud-init-local.service has finished starting up. -- -- The start-up result is done. Feb 05 04:23:19 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: Reached target Network (Pre). -- Subject: Unit network-pre.target has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit network-pre.target has finished starting up. -- -- The start-up result is done. Feb 05 04:23:19 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: Starting Network Manager... -- Subject: Unit NetworkManager.service has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit NetworkManager.service has begun starting up. Feb 05 04:23:19 ip-10-31-9-239.us-east-1.aws.redhat.com NetworkManager[675]: [1707106999.9039] NetworkManager (version 1.40.16-14.el8) is starting... (boot:4c834d63-595f-489f-a326-ce9c733910bd) Feb 05 04:23:19 ip-10-31-9-239.us-east-1.aws.redhat.com NetworkManager[675]: [1707106999.9059] Read config: /etc/NetworkManager/NetworkManager.conf (run: 15-carrier-timeout.conf) Feb 05 04:23:19 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: Started Network Manager. -- Subject: Unit NetworkManager.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit NetworkManager.service has finished starting up. -- -- The start-up result is done. Feb 05 04:23:19 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: Reached target Network. -- Subject: Unit network.target has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit network.target has finished starting up. -- -- The start-up result is done. Feb 05 04:23:19 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: Starting Dynamic System Tuning Daemon... -- Subject: Unit tuned.service has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit tuned.service has begun starting up. Feb 05 04:23:19 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: Starting GSSAPI Proxy Daemon... -- Subject: Unit gssproxy.service has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit gssproxy.service has begun starting up. Feb 05 04:23:19 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: Starting Network Manager Wait Online... -- Subject: Unit NetworkManager-wait-online.service has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit NetworkManager-wait-online.service has begun starting up. Feb 05 04:23:19 ip-10-31-9-239.us-east-1.aws.redhat.com NetworkManager[675]: [1707106999.9158] bus-manager: acquired D-Bus service "org.freedesktop.NetworkManager" Feb 05 04:23:19 ip-10-31-9-239.us-east-1.aws.redhat.com NetworkManager[675]: [1707106999.9488] manager[0x560da7ae2080]: monitoring kernel firmware directory '/lib/firmware'. Feb 05 04:23:19 ip-10-31-9-239.us-east-1.aws.redhat.com NetworkManager[675]: [1707106999.9519] hostname: hostname: using hostnamed Feb 05 04:23:19 ip-10-31-9-239.us-east-1.aws.redhat.com NetworkManager[675]: [1707106999.9519] hostname: static hostname changed from (none) to "ip-10-31-9-239.us-east-1.aws.redhat.com" Feb 05 04:23:19 ip-10-31-9-239.us-east-1.aws.redhat.com NetworkManager[675]: [1707106999.9523] dns-mgr: init: dns=default,systemd-resolved rc-manager=symlink Feb 05 04:23:19 ip-10-31-9-239.us-east-1.aws.redhat.com NetworkManager[675]: [1707106999.9614] Loaded device plugin: NMTeamFactory (/usr/lib64/NetworkManager/1.40.16-14.el8/libnm-device-plugin-team.so) Feb 05 04:23:19 ip-10-31-9-239.us-east-1.aws.redhat.com NetworkManager[675]: [1707106999.9614] manager: rfkill: Wi-Fi enabled by radio killswitch; enabled by state file Feb 05 04:23:19 ip-10-31-9-239.us-east-1.aws.redhat.com NetworkManager[675]: [1707106999.9616] manager: rfkill: WWAN enabled by radio killswitch; enabled by state file Feb 05 04:23:19 ip-10-31-9-239.us-east-1.aws.redhat.com NetworkManager[675]: [1707106999.9618] manager: Networking is enabled by state file Feb 05 04:23:19 ip-10-31-9-239.us-east-1.aws.redhat.com dbus-daemon[628]: [system] Activating via systemd: service name='org.freedesktop.nm_dispatcher' unit='dbus-org.freedesktop.nm-dispatcher.service' requested by ':1.5' (uid=0 pid=675 comm="/usr/sbin/NetworkManager --no-daemon " label="system_u:system_r:NetworkManager_t:s0") Feb 05 04:23:19 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: Starting Network Manager Script Dispatcher Service... -- Subject: Unit NetworkManager-dispatcher.service has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit NetworkManager-dispatcher.service has begun starting up. Feb 05 04:23:19 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: Started GSSAPI Proxy Daemon. -- Subject: Unit gssproxy.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit gssproxy.service has finished starting up. -- -- The start-up result is done. Feb 05 04:23:19 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: Reached target NFS client services. -- Subject: Unit nfs-client.target has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit nfs-client.target has finished starting up. -- -- The start-up result is done. Feb 05 04:23:19 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: Reached target Remote File Systems (Pre). -- Subject: Unit remote-fs-pre.target has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit remote-fs-pre.target has finished starting up. -- -- The start-up result is done. Feb 05 04:23:19 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: Reached target Remote File Systems. -- Subject: Unit remote-fs.target has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit remote-fs.target has finished starting up. -- -- The start-up result is done. Feb 05 04:23:19 ip-10-31-9-239.us-east-1.aws.redhat.com NetworkManager[675]: [1707106999.9715] settings: Loaded settings plugin: ifcfg-rh ("/usr/lib64/NetworkManager/1.40.16-14.el8/libnm-settings-plugin-ifcfg-rh.so") Feb 05 04:23:19 ip-10-31-9-239.us-east-1.aws.redhat.com NetworkManager[675]: [1707106999.9717] settings: Loaded settings plugin: keyfile (internal) Feb 05 04:23:19 ip-10-31-9-239.us-east-1.aws.redhat.com NetworkManager[675]: [1707106999.9770] dhcp: init: Using DHCP client 'internal' Feb 05 04:23:19 ip-10-31-9-239.us-east-1.aws.redhat.com NetworkManager[675]: [1707106999.9771] device (lo): carrier: link connected Feb 05 04:23:19 ip-10-31-9-239.us-east-1.aws.redhat.com NetworkManager[675]: [1707106999.9774] manager: (lo): new Generic device (/org/freedesktop/NetworkManager/Devices/1) Feb 05 04:23:19 ip-10-31-9-239.us-east-1.aws.redhat.com NetworkManager[675]: [1707106999.9786] manager: (eth0): new Ethernet device (/org/freedesktop/NetworkManager/Devices/2) Feb 05 04:23:19 ip-10-31-9-239.us-east-1.aws.redhat.com NetworkManager[675]: [1707106999.9791] device (eth0): state change: unmanaged -> unavailable (reason 'managed', sys-iface-state: 'external') Feb 05 04:23:19 ip-10-31-9-239.us-east-1.aws.redhat.com kernel: IPv6: ADDRCONF(NETDEV_UP): eth0: link is not ready Feb 05 04:23:19 ip-10-31-9-239.us-east-1.aws.redhat.com NetworkManager[675]: [1707106999.9809] device (eth0): carrier: link connected Feb 05 04:23:19 ip-10-31-9-239.us-east-1.aws.redhat.com NetworkManager[675]: [1707106999.9811] device (eth0): state change: unavailable -> disconnected (reason 'carrier-changed', sys-iface-state: 'managed') Feb 05 04:23:19 ip-10-31-9-239.us-east-1.aws.redhat.com NetworkManager[675]: [1707106999.9816] policy: auto-activating connection 'System eth0' (5fb06bd0-0bb0-7ffb-45f1-d6edd65f3e03) Feb 05 04:23:19 ip-10-31-9-239.us-east-1.aws.redhat.com NetworkManager[675]: [1707106999.9822] device (eth0): Activation: starting connection 'System eth0' (5fb06bd0-0bb0-7ffb-45f1-d6edd65f3e03) Feb 05 04:23:19 ip-10-31-9-239.us-east-1.aws.redhat.com NetworkManager[675]: [1707106999.9824] device (eth0): state change: disconnected -> prepare (reason 'none', sys-iface-state: 'managed') Feb 05 04:23:19 ip-10-31-9-239.us-east-1.aws.redhat.com NetworkManager[675]: [1707106999.9827] manager: NetworkManager state is now CONNECTING Feb 05 04:23:19 ip-10-31-9-239.us-east-1.aws.redhat.com NetworkManager[675]: [1707106999.9829] device (eth0): state change: prepare -> config (reason 'none', sys-iface-state: 'managed') Feb 05 04:23:19 ip-10-31-9-239.us-east-1.aws.redhat.com NetworkManager[675]: [1707106999.9834] device (eth0): state change: config -> ip-config (reason 'none', sys-iface-state: 'managed') Feb 05 04:23:19 ip-10-31-9-239.us-east-1.aws.redhat.com NetworkManager[675]: [1707106999.9844] dhcp4 (eth0): activation: beginning transaction (timeout in 45 seconds) Feb 05 04:23:19 ip-10-31-9-239.us-east-1.aws.redhat.com NetworkManager[675]: [1707106999.9864] dhcp4 (eth0): state changed new lease, address=10.31.9.239 Feb 05 04:23:19 ip-10-31-9-239.us-east-1.aws.redhat.com NetworkManager[675]: [1707106999.9868] policy: set 'System eth0' (eth0) as default for IPv4 routing and DNS Feb 05 04:23:19 ip-10-31-9-239.us-east-1.aws.redhat.com dbus-daemon[628]: [system] Activating via systemd: service name='org.freedesktop.resolve1' unit='dbus-org.freedesktop.resolve1.service' requested by ':1.5' (uid=0 pid=675 comm="/usr/sbin/NetworkManager --no-daemon " label="system_u:system_r:NetworkManager_t:s0") Feb 05 04:23:19 ip-10-31-9-239.us-east-1.aws.redhat.com dbus-daemon[628]: [system] Activation via systemd failed for unit 'dbus-org.freedesktop.resolve1.service': Unit dbus-org.freedesktop.resolve1.service not found. Feb 05 04:23:19 ip-10-31-9-239.us-east-1.aws.redhat.com NetworkManager[675]: [1707106999.9937] device (eth0): state change: ip-config -> ip-check (reason 'none', sys-iface-state: 'managed') Feb 05 04:23:20 ip-10-31-9-239.us-east-1.aws.redhat.com dbus-daemon[628]: [system] Successfully activated service 'org.freedesktop.nm_dispatcher' Feb 05 04:23:20 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: Started Network Manager Script Dispatcher Service. -- Subject: Unit NetworkManager-dispatcher.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit NetworkManager-dispatcher.service has finished starting up. -- -- The start-up result is done. Feb 05 04:23:20 ip-10-31-9-239.us-east-1.aws.redhat.com NetworkManager[675]: [1707107000.0827] device (eth0): state change: ip-check -> secondaries (reason 'none', sys-iface-state: 'managed') Feb 05 04:23:20 ip-10-31-9-239.us-east-1.aws.redhat.com NetworkManager[675]: [1707107000.0829] device (eth0): state change: secondaries -> activated (reason 'none', sys-iface-state: 'managed') Feb 05 04:23:20 ip-10-31-9-239.us-east-1.aws.redhat.com NetworkManager[675]: [1707107000.0833] manager: NetworkManager state is now CONNECTED_SITE Feb 05 04:23:20 ip-10-31-9-239.us-east-1.aws.redhat.com NetworkManager[675]: [1707107000.0836] device (eth0): Activation: successful, device activated. Feb 05 04:23:20 ip-10-31-9-239.us-east-1.aws.redhat.com NetworkManager[675]: [1707107000.0841] manager: NetworkManager state is now CONNECTED_GLOBAL Feb 05 04:23:20 ip-10-31-9-239.us-east-1.aws.redhat.com NetworkManager[675]: [1707107000.0845] manager: startup complete Feb 05 04:23:20 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: Started Network Manager Wait Online. -- Subject: Unit NetworkManager-wait-online.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit NetworkManager-wait-online.service has finished starting up. -- -- The start-up result is done. Feb 05 04:23:20 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: Starting Initial cloud-init job (metadata service crawler)... -- Subject: Unit cloud-init.service has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit cloud-init.service has begun starting up. Feb 05 04:23:20 ip-10-31-9-239.us-east-1.aws.redhat.com dbus-daemon[628]: [system] Activating via systemd: service name='org.freedesktop.PolicyKit1' unit='polkit.service' requested by ':1.9' (uid=0 pid=678 comm="/usr/libexec/platform-python -Es /usr/sbin/tuned -" label="system_u:system_r:tuned_t:s0") Feb 05 04:23:20 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: Starting Authorization Manager... -- Subject: Unit polkit.service has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit polkit.service has begun starting up. Feb 05 04:23:20 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: Started Dynamic System Tuning Daemon. -- Subject: Unit tuned.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit tuned.service has finished starting up. -- -- The start-up result is done. Feb 05 04:23:20 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: Cloud-init v. 23.4-2.el8 running 'init' at Mon, 05 Feb 2024 04:23:20 +0000. Up 17.79 seconds. Feb 05 04:23:20 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: ci-info: +++++++++++++++++++++++++++++++++++++++Net device info+++++++++++++++++++++++++++++++++++++++ Feb 05 04:23:20 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: ci-info: +--------+------+------------------------------+---------------+--------+-------------------+ Feb 05 04:23:20 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: ci-info: | Device | Up | Address | Mask | Scope | Hw-Address | Feb 05 04:23:20 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: ci-info: +--------+------+------------------------------+---------------+--------+-------------------+ Feb 05 04:23:20 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: ci-info: | eth0 | True | 10.31.9.239 | 255.255.252.0 | global | 12:86:38:a3:1b:25 | Feb 05 04:23:20 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: ci-info: | eth0 | True | fe80::1086:38ff:fea3:1b25/64 | . | link | 12:86:38:a3:1b:25 | Feb 05 04:23:20 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: ci-info: | lo | True | 127.0.0.1 | 255.0.0.0 | host | . | Feb 05 04:23:20 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: ci-info: | lo | True | ::1/128 | . | host | . | Feb 05 04:23:20 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: ci-info: +--------+------+------------------------------+---------------+--------+-------------------+ Feb 05 04:23:20 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: ci-info: ++++++++++++++++++++++++++++Route IPv4 info++++++++++++++++++++++++++++ Feb 05 04:23:20 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: ci-info: +-------+-------------+-----------+---------------+-----------+-------+ Feb 05 04:23:20 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: ci-info: | Route | Destination | Gateway | Genmask | Interface | Flags | Feb 05 04:23:20 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: ci-info: +-------+-------------+-----------+---------------+-----------+-------+ Feb 05 04:23:20 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: ci-info: | 0 | 0.0.0.0 | 10.31.8.1 | 0.0.0.0 | eth0 | UG | Feb 05 04:23:20 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: ci-info: | 1 | 10.31.8.0 | 0.0.0.0 | 255.255.252.0 | eth0 | U | Feb 05 04:23:20 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: ci-info: +-------+-------------+-----------+---------------+-----------+-------+ Feb 05 04:23:20 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: ci-info: +++++++++++++++++++Route IPv6 info+++++++++++++++++++ Feb 05 04:23:20 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: ci-info: +-------+-------------+---------+-----------+-------+ Feb 05 04:23:20 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: ci-info: | Route | Destination | Gateway | Interface | Flags | Feb 05 04:23:20 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: ci-info: +-------+-------------+---------+-----------+-------+ Feb 05 04:23:20 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: ci-info: | 1 | fe80::/64 | :: | eth0 | U | Feb 05 04:23:20 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: ci-info: | 3 | multicast | :: | eth0 | U | Feb 05 04:23:20 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: ci-info: +-------+-------------+---------+-----------+-------+ Feb 05 04:23:20 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: 2024-02-05 04:23:20,794 - __init__.py[WARNING]: Unhandled non-multipart (text/x-not-multipart) userdata: 'b'~/.citool.d/post-install'...' Feb 05 04:23:21 ip-10-31-9-239.us-east-1.aws.redhat.com polkitd[939]: Started polkitd version 0.115 Feb 05 04:23:21 ip-10-31-9-239.us-east-1.aws.redhat.com polkitd[939]: Loading rules from directory /etc/polkit-1/rules.d Feb 05 04:23:21 ip-10-31-9-239.us-east-1.aws.redhat.com polkitd[939]: Loading rules from directory /usr/share/polkit-1/rules.d Feb 05 04:23:21 ip-10-31-9-239.us-east-1.aws.redhat.com polkitd[939]: Finished loading, compiling and executing 2 rules Feb 05 04:23:21 ip-10-31-9-239.us-east-1.aws.redhat.com dbus-daemon[628]: [system] Successfully activated service 'org.freedesktop.PolicyKit1' Feb 05 04:23:21 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: Started Authorization Manager. -- Subject: Unit polkit.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit polkit.service has finished starting up. -- -- The start-up result is done. Feb 05 04:23:21 ip-10-31-9-239.us-east-1.aws.redhat.com polkitd[939]: Acquired the name org.freedesktop.PolicyKit1 on the system bus Feb 05 04:23:26 ip-10-31-9-239.us-east-1.aws.redhat.com chronyd[641]: Selected source 50.205.57.38 (2.centos.pool.ntp.org) Feb 05 04:23:26 ip-10-31-9-239.us-east-1.aws.redhat.com chronyd[641]: System clock TAI offset set to 37 seconds Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: Generating public/private rsa key pair. Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: Your identification has been saved in /etc/ssh/ssh_host_rsa_key. Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: Your public key has been saved in /etc/ssh/ssh_host_rsa_key.pub. Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: The key fingerprint is: Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: SHA256:NigHwxW+wpVyv/FN3fQM7G/IWrnvU5fOTDsvSrMphfM root@ip-10-31-9-239.us-east-1.aws.redhat.com Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: The key's randomart image is: Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: +---[RSA 3072]----+ Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: | o. | Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: | . o . . | Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: | = = o .| Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: | . * + o =.| Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: | + + S .. o =| Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: | + . =oo.. *o| Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: | . .++ X *| Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: | ..E*.X | Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: | .=..+B| Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: +----[SHA256]-----+ Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: Generating public/private dsa key pair. Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: Your identification has been saved in /etc/ssh/ssh_host_dsa_key. Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: Your public key has been saved in /etc/ssh/ssh_host_dsa_key.pub. Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: The key fingerprint is: Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: SHA256:ovhbVQckJv/322sEnQXJBUXj1cll1psXsdaoBUuGK6Q root@ip-10-31-9-239.us-east-1.aws.redhat.com Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: The key's randomart image is: Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: +---[DSA 1024]----+ Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: | . o.o.+.+X#| Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: | +...+ o+BO| Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: | o. ..o ++B| Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: | E .o.. +.=.| Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: | . S.. o . .| Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: | . . o . . . | Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: | . . . o | Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: | . . + | Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: | o. o.o| Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: +----[SHA256]-----+ Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: Generating public/private ecdsa key pair. Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: Your identification has been saved in /etc/ssh/ssh_host_ecdsa_key. Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: Your public key has been saved in /etc/ssh/ssh_host_ecdsa_key.pub. Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: The key fingerprint is: Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: SHA256:6gn0Rw+QHRNlOHEFCt3mzhUg+tUBJlK5aligFsQpnkA root@ip-10-31-9-239.us-east-1.aws.redhat.com Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: The key's randomart image is: Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: +---[ECDSA 256]---+ Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: |.E.. ooOBO=o | Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: |o + . *=Xo... | Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: |o..o .+ o=. .. | Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: | oo .o... . | Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: | . .o .So . | Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: | ...oo oo | Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: | ..o . . | Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: | o o | Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: | o | Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: +----[SHA256]-----+ Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: Generating public/private ed25519 key pair. Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: Your identification has been saved in /etc/ssh/ssh_host_ed25519_key. Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: Your public key has been saved in /etc/ssh/ssh_host_ed25519_key.pub. Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: The key fingerprint is: Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: SHA256:7htWOIJADz3r1Aq5CEvZLyRUcICvrWLJ47sjSSqVxz0 root@ip-10-31-9-239.us-east-1.aws.redhat.com Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: The key's randomart image is: Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: +--[ED25519 256]--+ Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: |.o*+ | Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: |.o.oo | Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: |..+..+ | Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: |.+++o.. . | Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: |o=oBoo. S . | Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: |+.=.=.Eo o | Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: |++... .+ | Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: |B* o . | Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: |*+= o. | Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[934]: +----[SHA256]-----+ Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: Started Initial cloud-init job (metadata service crawler). -- Subject: Unit cloud-init.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit cloud-init.service has finished starting up. -- -- The start-up result is done. Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: Reached target Cloud-config availability. -- Subject: Unit cloud-config.target has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit cloud-config.target has finished starting up. -- -- The start-up result is done. Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: Starting OpenSSH server daemon... -- Subject: Unit sshd.service has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit sshd.service has begun starting up. Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: Reached target Network is Online. -- Subject: Unit network-online.target has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit network-online.target has finished starting up. -- -- The start-up result is done. Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: Starting The restraint harness.... -- Subject: Unit restraintd.service has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit restraintd.service has begun starting up. Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: Starting Apply the settings specified in cloud-config... -- Subject: Unit cloud-config.service has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit cloud-config.service has begun starting up. Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: Starting Notify NFS peers of a restart... -- Subject: Unit rpc-statd-notify.service has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit rpc-statd-notify.service has begun starting up. Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: Starting System Logging Service... -- Subject: Unit rsyslog.service has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit rsyslog.service has begun starting up. Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: Starting Crash recovery kernel arming... -- Subject: Unit kdump.service has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit kdump.service has begun starting up. Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com sshd[1030]: Server listening on 0.0.0.0 port 22. Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com sshd[1030]: Server listening on :: port 22. Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: Started OpenSSH server daemon. -- Subject: Unit sshd.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit sshd.service has finished starting up. -- -- The start-up result is done. Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: Started The restraint harness.. -- Subject: Unit restraintd.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit restraintd.service has finished starting up. -- -- The start-up result is done. Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com sm-notify[1033]: Version 2.3.3 starting Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: Started Notify NFS peers of a restart. -- Subject: Unit rpc-statd-notify.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit rpc-statd-notify.service has finished starting up. -- -- The start-up result is done. Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com rsyslogd[1034]: [origin software="rsyslogd" swVersion="8.2102.0-15.el8" x-pid="1034" x-info="https://www.rsyslog.com"] start Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: Started System Logging Service. -- Subject: Unit rsyslog.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit rsyslog.service has finished starting up. -- -- The start-up result is done. Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com restraintd[1043]: Listening on http://localhost:8081 Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com rsyslogd[1034]: imjournal: journal files changed, reloading... [v8.2102.0-15.el8 try https://www.rsyslog.com/e/0 ] Feb 05 04:23:28 ip-10-31-9-239.us-east-1.aws.redhat.com kdumpctl[1037]: kdump: Detected change(s) in the following file(s): /etc/fstab Feb 05 04:23:29 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[1275]: Cloud-init v. 23.4-2.el8 running 'modules:config' at Mon, 05 Feb 2024 04:23:29 +0000. Up 26.20 seconds. Feb 05 04:23:29 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: Started Apply the settings specified in cloud-config. -- Subject: Unit cloud-config.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit cloud-config.service has finished starting up. -- -- The start-up result is done. Feb 05 04:23:29 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: Starting Permit User Sessions... -- Subject: Unit systemd-user-sessions.service has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit systemd-user-sessions.service has begun starting up. Feb 05 04:23:29 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: Starting Execute cloud user/final scripts... -- Subject: Unit cloud-final.service has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit cloud-final.service has begun starting up. Feb 05 04:23:29 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: Started Permit User Sessions. -- Subject: Unit systemd-user-sessions.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit systemd-user-sessions.service has finished starting up. -- -- The start-up result is done. Feb 05 04:23:29 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: Starting Terminate Plymouth Boot Screen... -- Subject: Unit plymouth-quit.service has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit plymouth-quit.service has begun starting up. Feb 05 04:23:29 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: Starting Hold until boot process finishes up... -- Subject: Unit plymouth-quit-wait.service has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit plymouth-quit-wait.service has begun starting up. Feb 05 04:23:29 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: Started Command Scheduler. -- Subject: Unit crond.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit crond.service has finished starting up. -- -- The start-up result is done. Feb 05 04:23:29 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: Received SIGRTMIN+21 from PID 339 (plymouthd). Feb 05 04:23:29 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: Started Terminate Plymouth Boot Screen. -- Subject: Unit plymouth-quit.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit plymouth-quit.service has finished starting up. -- -- The start-up result is done. Feb 05 04:23:29 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: Started Hold until boot process finishes up. -- Subject: Unit plymouth-quit-wait.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit plymouth-quit-wait.service has finished starting up. -- -- The start-up result is done. Feb 05 04:23:29 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: Received SIGRTMIN+21 from PID 339 (n/a). Feb 05 04:23:29 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: Started Getty on tty1. -- Subject: Unit getty@tty1.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit getty@tty1.service has finished starting up. -- -- The start-up result is done. Feb 05 04:23:29 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: Reached target Login Prompts. -- Subject: Unit getty.target has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit getty.target has finished starting up. -- -- The start-up result is done. Feb 05 04:23:29 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: Reached target Multi-User System. -- Subject: Unit multi-user.target has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit multi-user.target has finished starting up. -- -- The start-up result is done. Feb 05 04:23:29 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: Starting Update UTMP about System Runlevel Changes... -- Subject: Unit systemd-update-utmp-runlevel.service has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit systemd-update-utmp-runlevel.service has begun starting up. Feb 05 04:23:29 ip-10-31-9-239.us-east-1.aws.redhat.com crond[1286]: (CRON) STARTUP (1.5.2) Feb 05 04:23:29 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: systemd-update-utmp-runlevel.service: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit systemd-update-utmp-runlevel.service has successfully entered the 'dead' state. Feb 05 04:23:29 ip-10-31-9-239.us-east-1.aws.redhat.com crond[1286]: (CRON) INFO (Syslog will be used instead of sendmail.) Feb 05 04:23:29 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: Started Update UTMP about System Runlevel Changes. -- Subject: Unit systemd-update-utmp-runlevel.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit systemd-update-utmp-runlevel.service has finished starting up. -- -- The start-up result is done. Feb 05 04:23:29 ip-10-31-9-239.us-east-1.aws.redhat.com crond[1286]: (CRON) INFO (RANDOM_DELAY will be scaled with factor 17% if used.) Feb 05 04:23:29 ip-10-31-9-239.us-east-1.aws.redhat.com crond[1286]: (CRON) INFO (running with inotify support) Feb 05 04:23:29 ip-10-31-9-239.us-east-1.aws.redhat.com kdumpctl[1037]: kdump: Rebuilding /boot/initramfs-4.18.0-536.el8.x86_64kdump.img Feb 05 04:23:29 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[1369]: ############################################################# Feb 05 04:23:29 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[1370]: -----BEGIN SSH HOST KEY FINGERPRINTS----- Feb 05 04:23:29 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[1372]: 1024 SHA256:ovhbVQckJv/322sEnQXJBUXj1cll1psXsdaoBUuGK6Q root@ip-10-31-9-239.us-east-1.aws.redhat.com (DSA) Feb 05 04:23:29 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[1374]: 256 SHA256:6gn0Rw+QHRNlOHEFCt3mzhUg+tUBJlK5aligFsQpnkA root@ip-10-31-9-239.us-east-1.aws.redhat.com (ECDSA) Feb 05 04:23:29 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[1376]: 256 SHA256:7htWOIJADz3r1Aq5CEvZLyRUcICvrWLJ47sjSSqVxz0 root@ip-10-31-9-239.us-east-1.aws.redhat.com (ED25519) Feb 05 04:23:29 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[1378]: 3072 SHA256:NigHwxW+wpVyv/FN3fQM7G/IWrnvU5fOTDsvSrMphfM root@ip-10-31-9-239.us-east-1.aws.redhat.com (RSA) Feb 05 04:23:29 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[1379]: -----END SSH HOST KEY FINGERPRINTS----- Feb 05 04:23:29 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[1380]: ############################################################# Feb 05 04:23:29 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[1359]: Cloud-init v. 23.4-2.el8 running 'modules:final' at Mon, 05 Feb 2024 04:23:29 +0000. Up 26.84 seconds. Feb 05 04:23:29 ip-10-31-9-239.us-east-1.aws.redhat.com cloud-init[1359]: Cloud-init v. 23.4-2.el8 finished at Mon, 05 Feb 2024 04:23:29 +0000. Datasource DataSourceEc2Local. Up 27.07 seconds Feb 05 04:23:29 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: Started Execute cloud user/final scripts. -- Subject: Unit cloud-final.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit cloud-final.service has finished starting up. -- -- The start-up result is done. Feb 05 04:23:29 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: Reached target Cloud-init target. -- Subject: Unit cloud-init.target has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit cloud-init.target has finished starting up. -- -- The start-up result is done. Feb 05 04:23:30 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: NetworkManager-dispatcher.service: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit NetworkManager-dispatcher.service has successfully entered the 'dead' state. Feb 05 04:23:30 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1591]: dracut-049-233.git20240115.el8 Feb 05 04:23:30 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: Executing: /usr/bin/dracut --add kdumpbase --quiet --hostonly --hostonly-cmdline --hostonly-i18n --hostonly-mode strict --hostonly-nics -o "plymouth dash resume ifcfg earlykdump" --compress=xz --mount "/dev/disk/by-uuid/929d1e8c-5aa9-4a2e-9c9b-ff48578ec2fe /sysroot xfs rw,relatime,seclabel,attr2,inode64,logbufs=8,logbsize=32k,noquota,nofail,x-systemd.before=initrd-fs.target" --no-hostonly-default-device -f /boot/initramfs-4.18.0-536.el8.x86_64kdump.img 4.18.0-536.el8.x86_64 Feb 05 04:23:30 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: dracut module 'busybox' will not be installed, because command 'busybox' could not be found! Feb 05 04:23:31 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: dracut module 'ifcfg' will not be installed, because it's in the list to be omitted! Feb 05 04:23:31 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: dracut module 'plymouth' will not be installed, because it's in the list to be omitted! Feb 05 04:23:31 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: dracut module 'lvmmerge' will not be installed, because command 'lvm' could not be found! Feb 05 04:23:31 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: dracut module 'lvmthinpool-monitor' will not be installed, because command 'lvm' could not be found! Feb 05 04:23:31 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: dracut module 'btrfs' will not be installed, because command 'btrfs' could not be found! Feb 05 04:23:31 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: dracut module 'dmraid' will not be installed, because command 'dmraid' could not be found! Feb 05 04:23:31 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: dracut module 'lvm' will not be installed, because command 'lvm' could not be found! Feb 05 04:23:31 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: dracut module 'mdraid' will not be installed, because command 'mdadm' could not be found! Feb 05 04:23:31 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: dracut module 'cifs' will not be installed, because command 'mount.cifs' could not be found! Feb 05 04:23:31 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: dracut module 'iscsi' will not be installed, because command 'iscsi-iname' could not be found! Feb 05 04:23:31 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: dracut module 'iscsi' will not be installed, because command 'iscsiadm' could not be found! Feb 05 04:23:31 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: dracut module 'iscsi' will not be installed, because command 'iscsid' could not be found! Feb 05 04:23:31 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: dracut module 'nvmf' will not be installed, because command 'nvme' could not be found! Feb 05 04:23:31 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: dracut module 'resume' will not be installed, because it's in the list to be omitted! Feb 05 04:23:31 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: dracut module 'earlykdump' will not be installed, because it's in the list to be omitted! Feb 05 04:23:31 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: memstrack is available Feb 05 04:23:31 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: dracut module 'busybox' will not be installed, because command 'busybox' could not be found! Feb 05 04:23:31 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: dracut module 'lvmmerge' will not be installed, because command 'lvm' could not be found! Feb 05 04:23:31 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: dracut module 'lvmthinpool-monitor' will not be installed, because command 'lvm' could not be found! Feb 05 04:23:31 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: dracut module 'btrfs' will not be installed, because command 'btrfs' could not be found! Feb 05 04:23:31 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: dracut module 'dmraid' will not be installed, because command 'dmraid' could not be found! Feb 05 04:23:31 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: dracut module 'lvm' will not be installed, because command 'lvm' could not be found! Feb 05 04:23:31 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: dracut module 'mdraid' will not be installed, because command 'mdadm' could not be found! Feb 05 04:23:31 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: dracut module 'cifs' will not be installed, because command 'mount.cifs' could not be found! Feb 05 04:23:31 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: dracut module 'iscsi' will not be installed, because command 'iscsi-iname' could not be found! Feb 05 04:23:31 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: dracut module 'iscsi' will not be installed, because command 'iscsiadm' could not be found! Feb 05 04:23:31 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: dracut module 'iscsi' will not be installed, because command 'iscsid' could not be found! Feb 05 04:23:31 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: dracut module 'nvmf' will not be installed, because command 'nvme' could not be found! Feb 05 04:23:32 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: *** Including module: bash *** Feb 05 04:23:32 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: *** Including module: systemd *** Feb 05 04:23:32 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: *** Including module: systemd-initrd *** Feb 05 04:23:32 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: *** Including module: watchdog-modules *** Feb 05 04:23:32 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: *** Including module: nss-softokn *** Feb 05 04:23:32 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: *** Including module: rngd *** Feb 05 04:23:32 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: *** Including module: i18n *** Feb 05 04:23:32 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: *** Including module: drm *** Feb 05 04:23:32 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: *** Including module: prefixdevname *** Feb 05 04:23:32 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: *** Including module: kernel-modules *** Feb 05 04:23:33 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: *** Including module: kernel-modules-extra *** Feb 05 04:23:33 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: kernel-modules-extra: configuration source "/run/depmod.d/" does not exist Feb 05 04:23:33 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: kernel-modules-extra: configuration source "/lib/depmod.d/" does not exist Feb 05 04:23:33 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: kernel-modules-extra: parsing configuration file "/etc/depmod.d//dist.conf" Feb 05 04:23:33 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: kernel-modules-extra: /etc/depmod.d//dist.conf: added "updates extra built-in weak-updates" to the list of search directories Feb 05 04:23:33 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: *** Including module: fstab-sys *** Feb 05 04:23:33 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: *** Including module: rootfs-block *** Feb 05 04:23:33 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: *** Including module: terminfo *** Feb 05 04:23:33 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: *** Including module: udev-rules *** Feb 05 04:23:33 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: Skipping udev rule: 91-permissions.rules Feb 05 04:23:33 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: Skipping udev rule: 80-drivers-modprobe.rules Feb 05 04:23:33 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: *** Including module: biosdevname *** Feb 05 04:23:33 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: *** Including module: dracut-systemd *** Feb 05 04:23:33 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: *** Including module: usrmount *** Feb 05 04:23:33 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: *** Including module: base *** Feb 05 04:23:33 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: *** Including module: fs-lib *** Feb 05 04:23:33 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: *** Including module: kdumpbase *** Feb 05 04:23:34 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: *** Including module: memstrack *** Feb 05 04:23:34 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: *** Including module: microcode_ctl-fw_dir_override *** Feb 05 04:23:34 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: microcode_ctl module: mangling fw_dir Feb 05 04:23:34 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: microcode_ctl: reset fw_dir to "/lib/firmware/updates /lib/firmware" Feb 05 04:23:34 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: microcode_ctl: processing data directory "/usr/share/microcode_ctl/ucode_with_caveats/intel"... Feb 05 04:23:34 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: microcode_ctl: intel: caveats check for kernel version "4.18.0-536.el8.x86_64" passed, adding "/usr/share/microcode_ctl/ucode_with_caveats/intel" to fw_dir variable Feb 05 04:23:34 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: microcode_ctl: processing data directory "/usr/share/microcode_ctl/ucode_with_caveats/intel-06-2d-07"... Feb 05 04:23:34 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: microcode_ctl: configuration "intel-06-2d-07" is ignored Feb 05 04:23:34 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: microcode_ctl: processing data directory "/usr/share/microcode_ctl/ucode_with_caveats/intel-06-4e-03"... Feb 05 04:23:34 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: microcode_ctl: configuration "intel-06-4e-03" is ignored Feb 05 04:23:34 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: microcode_ctl: processing data directory "/usr/share/microcode_ctl/ucode_with_caveats/intel-06-4f-01"... Feb 05 04:23:34 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: microcode_ctl: configuration "intel-06-4f-01" is ignored Feb 05 04:23:34 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: microcode_ctl: processing data directory "/usr/share/microcode_ctl/ucode_with_caveats/intel-06-55-04"... Feb 05 04:23:34 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: microcode_ctl: configuration "intel-06-55-04" is ignored Feb 05 04:23:34 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: microcode_ctl: processing data directory "/usr/share/microcode_ctl/ucode_with_caveats/intel-06-5e-03"... Feb 05 04:23:34 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: microcode_ctl: configuration "intel-06-5e-03" is ignored Feb 05 04:23:34 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: microcode_ctl: processing data directory "/usr/share/microcode_ctl/ucode_with_caveats/intel-06-8c-01"... Feb 05 04:23:34 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: microcode_ctl: configuration "intel-06-8c-01" is ignored Feb 05 04:23:34 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: microcode_ctl: processing data directory "/usr/share/microcode_ctl/ucode_with_caveats/intel-06-8e-9e-0x-0xca"... Feb 05 04:23:34 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: microcode_ctl: configuration "intel-06-8e-9e-0x-0xca" is ignored Feb 05 04:23:34 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: microcode_ctl: processing data directory "/usr/share/microcode_ctl/ucode_with_caveats/intel-06-8e-9e-0x-dell"... Feb 05 04:23:34 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: microcode_ctl: configuration "intel-06-8e-9e-0x-dell" is ignored Feb 05 04:23:34 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: microcode_ctl: final fw_dir: "/usr/share/microcode_ctl/ucode_with_caveats/intel /lib/firmware/updates /lib/firmware" Feb 05 04:23:34 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: *** Including module: shutdown *** Feb 05 04:23:34 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: *** Including module: squash *** Feb 05 04:23:34 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: *** Including modules done *** Feb 05 04:23:34 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: *** Installing kernel module dependencies *** Feb 05 04:23:35 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: *** Installing kernel module dependencies done *** Feb 05 04:23:35 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: *** Resolving executable dependencies *** Feb 05 04:23:36 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: *** Resolving executable dependencies done*** Feb 05 04:23:36 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: *** Hardlinking files *** Feb 05 04:23:36 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: *** Hardlinking files done *** Feb 05 04:23:36 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: *** Generating early-microcode cpio image *** Feb 05 04:23:36 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: *** Constructing GenuineIntel.bin *** Feb 05 04:23:36 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: *** Constructing GenuineIntel.bin *** Feb 05 04:23:36 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: *** Store current command line parameters *** Feb 05 04:23:36 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: Stored kernel commandline: Feb 05 04:23:36 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: No dracut internal kernel commandline stored in the initramfs Feb 05 04:23:36 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: *** Install squash loader *** Feb 05 04:23:37 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: *** Stripping files *** Feb 05 04:23:37 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: *** Stripping files done *** Feb 05 04:23:37 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: *** Squashing the files inside the initramfs *** Feb 05 04:23:47 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: *** Squashing the files inside the initramfs done *** Feb 05 04:23:47 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: *** Creating image file '/boot/initramfs-4.18.0-536.el8.x86_64kdump.img' *** Feb 05 04:23:47 ip-10-31-9-239.us-east-1.aws.redhat.com dracut[1594]: *** Creating initramfs image file '/boot/initramfs-4.18.0-536.el8.x86_64kdump.img' done *** Feb 05 04:23:48 ip-10-31-9-239.us-east-1.aws.redhat.com kdumpctl[1037]: kdump: kexec: loaded kdump kernel Feb 05 04:23:48 ip-10-31-9-239.us-east-1.aws.redhat.com kdumpctl[1037]: kdump: Starting kdump: [OK] Feb 05 04:23:48 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: Started Crash recovery kernel arming. -- Subject: Unit kdump.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit kdump.service has finished starting up. -- -- The start-up result is done. Feb 05 04:23:48 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: Startup finished in 678ms (kernel) + 5.096s (initrd) + 39.677s (userspace) = 45.452s. -- Subject: System start-up is now complete -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- All system services necessary queued for starting at boot have been -- started. Note that this does not mean that the machine is now idle as services -- might still be busy with completing start-up. -- -- Kernel start-up required 678098 microseconds. -- -- Initial RAM disk start-up required 5096470 microseconds. -- -- Userspace start-up required 39677622 microseconds. Feb 05 04:23:49 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: systemd-hostnamed.service: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit systemd-hostnamed.service has successfully entered the 'dead' state. Feb 05 04:24:31 ip-10-31-9-239.us-east-1.aws.redhat.com chronyd[641]: Selected source 10.2.32.38 Feb 05 04:24:40 ip-10-31-9-239.us-east-1.aws.redhat.com sshd[5077]: Accepted publickey for root from 10.30.32.10 port 35204 ssh2: RSA SHA256:W3cSdmPJK+d9RwU97ardijPXIZnxHswrpTHWW9oYtEU Feb 05 04:24:40 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: Created slice User Slice of UID 0. -- Subject: Unit user-0.slice has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit user-0.slice has finished starting up. -- -- The start-up result is done. Feb 05 04:24:40 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: Starting User runtime directory /run/user/0... -- Subject: Unit user-runtime-dir@0.service has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit user-runtime-dir@0.service has begun starting up. Feb 05 04:24:40 ip-10-31-9-239.us-east-1.aws.redhat.com systemd-logind[629]: New session 1 of user root. -- Subject: A new session 1 has been created for user root -- Defined-By: systemd -- Support: https://access.redhat.com/support -- Documentation: https://www.freedesktop.org/wiki/Software/systemd/multiseat -- -- A new session with the ID 1 has been created for the user root. -- -- The leading process of the session is 5077. Feb 05 04:24:40 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: Started User runtime directory /run/user/0. -- Subject: Unit user-runtime-dir@0.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit user-runtime-dir@0.service has finished starting up. -- -- The start-up result is done. Feb 05 04:24:40 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: Starting User Manager for UID 0... -- Subject: Unit user@0.service has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit user@0.service has begun starting up. Feb 05 04:24:40 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[5082]: pam_unix(systemd-user:session): session opened for user root by (uid=0) Feb 05 04:24:40 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[5082]: Reached target Paths. -- Subject: Unit UNIT has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished starting up. -- -- The start-up result is done. Feb 05 04:24:40 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[5082]: Starting D-Bus User Message Bus Socket. -- Subject: Unit UNIT has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has begun starting up. Feb 05 04:24:40 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[5082]: Reached target Timers. -- Subject: Unit UNIT has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished starting up. -- -- The start-up result is done. Feb 05 04:24:40 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[5082]: Listening on D-Bus User Message Bus Socket. -- Subject: Unit UNIT has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished starting up. -- -- The start-up result is done. Feb 05 04:24:40 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[5082]: Reached target Sockets. -- Subject: Unit UNIT has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished starting up. -- -- The start-up result is done. Feb 05 04:24:40 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[5082]: Reached target Basic System. -- Subject: Unit UNIT has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished starting up. -- -- The start-up result is done. Feb 05 04:24:40 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[5082]: Reached target Default. -- Subject: Unit UNIT has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished starting up. -- -- The start-up result is done. Feb 05 04:24:40 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[5082]: Startup finished in 107ms. -- Subject: User manager start-up is now complete -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The user manager instance for user 0 has been started. All services queued -- for starting have been started. Note that other services might still be starting -- up or be started at any later time. -- -- Startup of the manager took 107381 microseconds. Feb 05 04:24:40 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: Started User Manager for UID 0. -- Subject: Unit user@0.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit user@0.service has finished starting up. -- -- The start-up result is done. Feb 05 04:24:40 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: Started Session 1 of user root. -- Subject: Unit session-1.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit session-1.scope has finished starting up. -- -- The start-up result is done. Feb 05 04:24:40 ip-10-31-9-239.us-east-1.aws.redhat.com sshd[5077]: pam_unix(sshd:session): session opened for user root by (uid=0) Feb 05 04:24:41 ip-10-31-9-239.us-east-1.aws.redhat.com sshd[5091]: Received disconnect from 10.30.32.10 port 35204:11: disconnected by user Feb 05 04:24:41 ip-10-31-9-239.us-east-1.aws.redhat.com sshd[5091]: Disconnected from user root 10.30.32.10 port 35204 Feb 05 04:24:41 ip-10-31-9-239.us-east-1.aws.redhat.com sshd[5077]: pam_unix(sshd:session): session closed for user root Feb 05 04:24:41 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: session-1.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit session-1.scope has successfully entered the 'dead' state. Feb 05 04:24:41 ip-10-31-9-239.us-east-1.aws.redhat.com systemd-logind[629]: Session 1 logged out. Waiting for processes to exit. Feb 05 04:24:41 ip-10-31-9-239.us-east-1.aws.redhat.com systemd-logind[629]: Removed session 1. -- Subject: Session 1 has been terminated -- Defined-By: systemd -- Support: https://access.redhat.com/support -- Documentation: https://www.freedesktop.org/wiki/Software/systemd/multiseat -- -- A session with the ID 1 has been terminated. Feb 05 04:24:42 ip-10-31-9-239.us-east-1.aws.redhat.com sshd[5111]: error: kex_exchange_identification: Connection closed by remote host Feb 05 04:24:42 ip-10-31-9-239.us-east-1.aws.redhat.com sshd[5112]: Accepted publickey for root from 10.29.163.169 port 48922 ssh2: RSA SHA256:W3cSdmPJK+d9RwU97ardijPXIZnxHswrpTHWW9oYtEU Feb 05 04:24:42 ip-10-31-9-239.us-east-1.aws.redhat.com systemd-logind[629]: New session 3 of user root. -- Subject: A new session 3 has been created for user root -- Defined-By: systemd -- Support: https://access.redhat.com/support -- Documentation: https://www.freedesktop.org/wiki/Software/systemd/multiseat -- -- A new session with the ID 3 has been created for the user root. -- -- The leading process of the session is 5112. Feb 05 04:24:42 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: Started Session 3 of user root. -- Subject: Unit session-3.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit session-3.scope has finished starting up. -- -- The start-up result is done. Feb 05 04:24:42 ip-10-31-9-239.us-east-1.aws.redhat.com sshd[5112]: pam_unix(sshd:session): session opened for user root by (uid=0) Feb 05 04:24:42 ip-10-31-9-239.us-east-1.aws.redhat.com sshd[5115]: Received disconnect from 10.29.163.169 port 48922:11: disconnected by user Feb 05 04:24:42 ip-10-31-9-239.us-east-1.aws.redhat.com sshd[5115]: Disconnected from user root 10.29.163.169 port 48922 Feb 05 04:24:42 ip-10-31-9-239.us-east-1.aws.redhat.com sshd[5112]: pam_unix(sshd:session): session closed for user root Feb 05 04:24:42 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: session-3.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit session-3.scope has successfully entered the 'dead' state. Feb 05 04:24:42 ip-10-31-9-239.us-east-1.aws.redhat.com systemd-logind[629]: Session 3 logged out. Waiting for processes to exit. Feb 05 04:24:42 ip-10-31-9-239.us-east-1.aws.redhat.com systemd-logind[629]: Removed session 3. -- Subject: Session 3 has been terminated -- Defined-By: systemd -- Support: https://access.redhat.com/support -- Documentation: https://www.freedesktop.org/wiki/Software/systemd/multiseat -- -- A session with the ID 3 has been terminated. Feb 05 04:24:43 ip-10-31-9-239.us-east-1.aws.redhat.com sshd[5135]: Accepted publickey for root from 10.29.163.169 port 48926 ssh2: RSA SHA256:W3cSdmPJK+d9RwU97ardijPXIZnxHswrpTHWW9oYtEU Feb 05 04:24:43 ip-10-31-9-239.us-east-1.aws.redhat.com systemd-logind[629]: New session 4 of user root. -- Subject: A new session 4 has been created for user root -- Defined-By: systemd -- Support: https://access.redhat.com/support -- Documentation: https://www.freedesktop.org/wiki/Software/systemd/multiseat -- -- A new session with the ID 4 has been created for the user root. -- -- The leading process of the session is 5135. Feb 05 04:24:43 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: Started Session 4 of user root. -- Subject: Unit session-4.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit session-4.scope has finished starting up. -- -- The start-up result is done. Feb 05 04:24:43 ip-10-31-9-239.us-east-1.aws.redhat.com sshd[5135]: pam_unix(sshd:session): session opened for user root by (uid=0) Feb 05 04:24:43 ip-10-31-9-239.us-east-1.aws.redhat.com sshd[5138]: Received disconnect from 10.29.163.169 port 48926:11: disconnected by user Feb 05 04:24:43 ip-10-31-9-239.us-east-1.aws.redhat.com sshd[5138]: Disconnected from user root 10.29.163.169 port 48926 Feb 05 04:24:43 ip-10-31-9-239.us-east-1.aws.redhat.com sshd[5135]: pam_unix(sshd:session): session closed for user root Feb 05 04:24:43 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: session-4.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit session-4.scope has successfully entered the 'dead' state. Feb 05 04:24:43 ip-10-31-9-239.us-east-1.aws.redhat.com systemd-logind[629]: Session 4 logged out. Waiting for processes to exit. Feb 05 04:24:43 ip-10-31-9-239.us-east-1.aws.redhat.com systemd-logind[629]: Removed session 4. -- Subject: Session 4 has been terminated -- Defined-By: systemd -- Support: https://access.redhat.com/support -- Documentation: https://www.freedesktop.org/wiki/Software/systemd/multiseat -- -- A session with the ID 4 has been terminated. Feb 05 04:24:43 ip-10-31-9-239.us-east-1.aws.redhat.com sshd[5158]: Accepted publickey for root from 10.29.163.169 port 48930 ssh2: RSA SHA256:W3cSdmPJK+d9RwU97ardijPXIZnxHswrpTHWW9oYtEU Feb 05 04:24:43 ip-10-31-9-239.us-east-1.aws.redhat.com systemd-logind[629]: New session 5 of user root. -- Subject: A new session 5 has been created for user root -- Defined-By: systemd -- Support: https://access.redhat.com/support -- Documentation: https://www.freedesktop.org/wiki/Software/systemd/multiseat -- -- A new session with the ID 5 has been created for the user root. -- -- The leading process of the session is 5158. Feb 05 04:24:43 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: Started Session 5 of user root. -- Subject: Unit session-5.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit session-5.scope has finished starting up. -- -- The start-up result is done. Feb 05 04:24:43 ip-10-31-9-239.us-east-1.aws.redhat.com sshd[5158]: pam_unix(sshd:session): session opened for user root by (uid=0) Feb 05 04:24:43 ip-10-31-9-239.us-east-1.aws.redhat.com sshd[5161]: Received disconnect from 10.29.163.169 port 48930:11: disconnected by user Feb 05 04:24:43 ip-10-31-9-239.us-east-1.aws.redhat.com sshd[5161]: Disconnected from user root 10.29.163.169 port 48930 Feb 05 04:24:43 ip-10-31-9-239.us-east-1.aws.redhat.com sshd[5158]: pam_unix(sshd:session): session closed for user root Feb 05 04:24:43 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: session-5.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit session-5.scope has successfully entered the 'dead' state. Feb 05 04:24:43 ip-10-31-9-239.us-east-1.aws.redhat.com systemd-logind[629]: Session 5 logged out. Waiting for processes to exit. Feb 05 04:24:43 ip-10-31-9-239.us-east-1.aws.redhat.com systemd-logind[629]: Removed session 5. -- Subject: Session 5 has been terminated -- Defined-By: systemd -- Support: https://access.redhat.com/support -- Documentation: https://www.freedesktop.org/wiki/Software/systemd/multiseat -- -- A session with the ID 5 has been terminated. Feb 05 04:24:44 ip-10-31-9-239.us-east-1.aws.redhat.com sshd[5181]: Accepted publickey for root from 10.29.163.169 port 48940 ssh2: RSA SHA256:W3cSdmPJK+d9RwU97ardijPXIZnxHswrpTHWW9oYtEU Feb 05 04:24:44 ip-10-31-9-239.us-east-1.aws.redhat.com systemd-logind[629]: New session 6 of user root. -- Subject: A new session 6 has been created for user root -- Defined-By: systemd -- Support: https://access.redhat.com/support -- Documentation: https://www.freedesktop.org/wiki/Software/systemd/multiseat -- -- A new session with the ID 6 has been created for the user root. -- -- The leading process of the session is 5181. Feb 05 04:24:44 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: Started Session 6 of user root. -- Subject: Unit session-6.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit session-6.scope has finished starting up. -- -- The start-up result is done. Feb 05 04:24:44 ip-10-31-9-239.us-east-1.aws.redhat.com sshd[5181]: pam_unix(sshd:session): session opened for user root by (uid=0) Feb 05 04:25:08 ip-10-31-9-239.us-east-1.aws.redhat.com python3[5285]: ansible-ansible.legacy.setup Invoked with gather_subset=['all'] gather_timeout=10 filter=[] fact_path=/etc/ansible/facts.d Feb 05 04:25:09 ip-10-31-9-239.us-east-1.aws.redhat.com python3[5332]: ansible-service_facts Invoked Feb 05 04:25:11 ip-10-31-9-239.us-east-1.aws.redhat.com python3[5438]: ansible-ansible.legacy.command Invoked with _raw_params=rpm -qa _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 05 04:25:12 ip-10-31-9-239.us-east-1.aws.redhat.com python3[5459]: ansible-stat Invoked with path=/etc/yum.repos.d/qa-tools.repo follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Feb 05 04:25:12 ip-10-31-9-239.us-east-1.aws.redhat.com python3[5481]: ansible-ansible.legacy.command Invoked with _raw_params=curl -sLI --connect-timeout 5 -w '%{response_code}' http://liver2.lab.eng.brq.redhat.com/repo/ | grep ^200 _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 05 04:25:13 ip-10-31-9-239.us-east-1.aws.redhat.com python3[5505]: ansible-ansible.legacy.dnf Invoked with name=['ca-certificates'] state=latest allow_downgrade=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 allowerasing=False nobest=False conf_file=None disable_excludes=None download_dir=None list=None releasever=None Feb 05 04:25:16 ip-10-31-9-239.us-east-1.aws.redhat.com python3[5526]: ansible-ansible.legacy.dnf Invoked with name=['curl'] state=latest allow_downgrade=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 allowerasing=False nobest=False conf_file=None disable_excludes=None download_dir=None list=None releasever=None Feb 05 04:25:18 ip-10-31-9-239.us-east-1.aws.redhat.com python3[5547]: ansible-ansible.legacy.command Invoked with _raw_params=curl -skL -o /etc/pki/ca-trust/source/anchors/Current-IT-Root-CAs.pem https://certs.corp.redhat.com/certs/Current-IT-Root-CAs.pem _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 05 04:25:18 ip-10-31-9-239.us-east-1.aws.redhat.com python3[5569]: ansible-ansible.legacy.command Invoked with _raw_params=bash -c 'update-ca-trust' _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 05 04:25:19 ip-10-31-9-239.us-east-1.aws.redhat.com python3[5596]: ansible-ansible.legacy.command Invoked with _raw_params=bash -c 'update-ca-trust extract' _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 05 04:25:20 ip-10-31-9-239.us-east-1.aws.redhat.com python3[5623]: ansible-ansible.legacy.command Invoked with _raw_params=curl -v https://gitlab.cee.redhat.com _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 05 04:25:21 ip-10-31-9-239.us-east-1.aws.redhat.com python3[5645]: ansible-ansible.legacy.command Invoked with _raw_params=curl -v https://beaker.engineering.redhat.com/ _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 05 04:25:22 ip-10-31-9-239.us-east-1.aws.redhat.com python3[5667]: ansible-file Invoked with path=/etc/yum.repos.d/CentOS-Media.repo state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Feb 05 04:25:22 ip-10-31-9-239.us-east-1.aws.redhat.com python3[5687]: ansible-file Invoked with path=/etc/yum.repos.d/CentOS-Linux-Media.repo state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Feb 05 04:25:22 ip-10-31-9-239.us-east-1.aws.redhat.com python3[5707]: ansible-file Invoked with path=/etc/yum.repos.d/CentOS-Stream-Media.repo state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Feb 05 04:25:22 ip-10-31-9-239.us-east-1.aws.redhat.com python3[5727]: ansible-file Invoked with path=/etc/yum.repos.d/beaker-client-testing.repo state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Feb 05 04:25:22 ip-10-31-9-239.us-east-1.aws.redhat.com python3[5747]: ansible-stat Invoked with path=/etc/yum.repos.d/beaker-client.repo follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Feb 05 04:25:23 ip-10-31-9-239.us-east-1.aws.redhat.com python3[5769]: ansible-replace Invoked with path=/etc/yum.repos.d/beaker-client.repo regexp=7|\$releasever replace=8 backup=False encoding=utf-8 unsafe_writes=False after=None before=None validate=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Feb 05 04:25:23 ip-10-31-9-239.us-east-1.aws.redhat.com python3[5789]: ansible-stat Invoked with path=/etc/yum.repos.d/beaker-client.repo follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Feb 05 04:25:23 ip-10-31-9-239.us-east-1.aws.redhat.com python3[5811]: ansible-community.general.ini_file Invoked with path=/etc/yum.repos.d/beaker-client.repo section=beaker-client option=skip_if_unavailable value=1 backup=False state=present exclusive=True no_extra_spaces=False ignore_spaces=False allow_no_value=False modify_inactive_option=True create=True follow=False unsafe_writes=False values=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Feb 05 04:25:23 ip-10-31-9-239.us-east-1.aws.redhat.com python3[5831]: ansible-ansible.legacy.command Invoked with _raw_params=for repofile in $(ls -1 /etc/yum.repos.d/); do echo "# vvvvv $repofile vvvvv vvvvv"; cat "/etc/yum.repos.d/$repofile"; echo "# ^^^^^ ^^^^^ ^^^^^ ^^^^^"; echo ""; done _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 05 04:25:24 ip-10-31-9-239.us-east-1.aws.redhat.com python3[5874]: ansible-ansible.legacy.command Invoked with _raw_params=yum -q repolist _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 05 04:25:24 ip-10-31-9-239.us-east-1.aws.redhat.com python3[5896]: ansible-stat Invoked with path=/etc/yum.repos.d/baseos-ci.repo follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Feb 05 04:25:24 ip-10-31-9-239.us-east-1.aws.redhat.com python3[5916]: ansible-ansible.legacy.command Invoked with _raw_params=for repofile in $(ls -1 /etc/yum.repos.d/); do echo "# vvvvv $repofile vvvvv vvvvv"; cat "/etc/yum.repos.d/$repofile"; echo "# ^^^^^ ^^^^^ ^^^^^ ^^^^^"; echo ""; done _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 05 04:25:25 ip-10-31-9-239.us-east-1.aws.redhat.com python3[5959]: ansible-ansible.legacy.command Invoked with _raw_params=yum -q repolist _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 05 04:25:25 ip-10-31-9-239.us-east-1.aws.redhat.com python3[5981]: ansible-ansible.legacy.command Invoked with _raw_params=for repofile in $(ls -1 /etc/yum.repos.d/); do echo "# vvvvv $repofile vvvvv vvvvv"; cat "/etc/yum.repos.d/$repofile"; echo "# ^^^^^ ^^^^^ ^^^^^ ^^^^^"; echo ""; done _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 05 04:25:26 ip-10-31-9-239.us-east-1.aws.redhat.com python3[6024]: ansible-ansible.legacy.command Invoked with _raw_params=yum -q repolist _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 05 04:25:26 ip-10-31-9-239.us-east-1.aws.redhat.com python3[6046]: ansible-ansible.legacy.command Invoked with _raw_params=for repofile in $(ls -1 /etc/yum.repos.d/); do echo "# vvvvv $repofile vvvvv vvvvv"; cat "/etc/yum.repos.d/$repofile"; echo "# ^^^^^ ^^^^^ ^^^^^ ^^^^^"; echo ""; done _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 05 04:25:27 ip-10-31-9-239.us-east-1.aws.redhat.com python3[6089]: ansible-ansible.legacy.command Invoked with _raw_params=yum -q repolist _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 05 04:25:27 ip-10-31-9-239.us-east-1.aws.redhat.com python3[6111]: ansible-ansible.legacy.command Invoked with _raw_params=for repofile in $(ls -1 /etc/yum.repos.d/); do echo "# vvvvv $repofile vvvvv vvvvv"; cat "/etc/yum.repos.d/$repofile"; echo "# ^^^^^ ^^^^^ ^^^^^ ^^^^^"; echo ""; done _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 05 04:25:27 ip-10-31-9-239.us-east-1.aws.redhat.com python3[6154]: ansible-ansible.legacy.command Invoked with _raw_params=yum -q repolist _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 05 04:25:28 ip-10-31-9-239.us-east-1.aws.redhat.com python3[6176]: ansible-ansible.legacy.command Invoked with _raw_params=dnf clean all --enablerepo=* _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 05 04:25:28 ip-10-31-9-239.us-east-1.aws.redhat.com python3[6198]: ansible-ansible.legacy.command Invoked with _raw_params=dnf makecache --enablerepo=* _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 05 04:27:44 ip-10-31-9-239.us-east-1.aws.redhat.com python3[6385]: ansible-ansible.legacy.dnf Invoked with name=['createrepo'] state=latest allow_downgrade=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 allowerasing=False nobest=False conf_file=None disable_excludes=None download_dir=None list=None releasever=None Feb 05 04:27:49 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: Started /usr/bin/systemctl start man-db-cache-update. -- Subject: Unit run-r99d75392dae64bf9be4a5740d174241d.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit run-r99d75392dae64bf9be4a5740d174241d.service has finished starting up. -- -- The start-up result is done. Feb 05 04:27:49 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Feb 05 04:27:49 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: Starting man-db-cache-update.service... -- Subject: Unit man-db-cache-update.service has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit man-db-cache-update.service has begun starting up. Feb 05 04:27:49 ip-10-31-9-239.us-east-1.aws.redhat.com python3[6420]: ansible-ansible.legacy.dnf Invoked with name=['make'] state=latest allow_downgrade=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 allowerasing=False nobest=False conf_file=None disable_excludes=None download_dir=None list=None releasever=None Feb 05 04:27:50 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: man-db-cache-update.service: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit man-db-cache-update.service has successfully entered the 'dead' state. Feb 05 04:27:50 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: Started man-db-cache-update.service. -- Subject: Unit man-db-cache-update.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit man-db-cache-update.service has finished starting up. -- -- The start-up result is done. Feb 05 04:27:50 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: run-r99d75392dae64bf9be4a5740d174241d.service: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit run-r99d75392dae64bf9be4a5740d174241d.service has successfully entered the 'dead' state. Feb 05 04:27:52 ip-10-31-9-239.us-east-1.aws.redhat.com python3[6493]: ansible-ansible.legacy.dnf Invoked with name=['parted'] state=latest allow_downgrade=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 allowerasing=False nobest=False conf_file=None disable_excludes=None download_dir=None list=None releasever=None Feb 05 04:27:54 ip-10-31-9-239.us-east-1.aws.redhat.com python3[6514]: ansible-ansible.legacy.dnf Invoked with name=['psmisc'] state=latest allow_downgrade=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 allowerasing=False nobest=False conf_file=None disable_excludes=None download_dir=None list=None releasever=None Feb 05 04:27:56 ip-10-31-9-239.us-east-1.aws.redhat.com python3[6535]: ansible-ansible.legacy.dnf Invoked with name=['chrony'] state=latest allow_downgrade=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 allowerasing=False nobest=False conf_file=None disable_excludes=None download_dir=None list=None releasever=None Feb 05 04:27:59 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: Reloading. Feb 05 04:28:00 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: /usr/lib/systemd/system/cloud-init.target:15: Unknown lvalue 'ConditionEnvironment' in section 'Unit' Feb 05 04:28:00 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: /usr/lib/systemd/system/cloud-init.service:18: Unknown lvalue 'ConditionEnvironment' in section 'Unit' Feb 05 04:28:00 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: /usr/lib/systemd/system/cloud-init-local.service:12: Unknown lvalue 'ConditionEnvironment' in section 'Unit' Feb 05 04:28:00 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: /usr/lib/systemd/system/cloud-final.service:7: Unknown lvalue 'ConditionEnvironment' in section 'Unit' Feb 05 04:28:00 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: /usr/lib/systemd/system/cloud-config.service:9: Unknown lvalue 'ConditionEnvironment' in section 'Unit' Feb 05 04:28:00 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: Stopping NTP client/server... -- Subject: Unit chronyd.service has begun shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit chronyd.service has begun shutting down. Feb 05 04:28:00 ip-10-31-9-239.us-east-1.aws.redhat.com chronyd[641]: chronyd exiting Feb 05 04:28:00 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: chronyd.service: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit chronyd.service has successfully entered the 'dead' state. Feb 05 04:28:00 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: Stopped NTP client/server. -- Subject: Unit chronyd.service has finished shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit chronyd.service has finished shutting down. Feb 05 04:28:00 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: Starting NTP client/server... -- Subject: Unit chronyd.service has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit chronyd.service has begun starting up. Feb 05 04:28:00 ip-10-31-9-239.us-east-1.aws.redhat.com chronyd[6583]: chronyd version 4.5 starting (+CMDMON +NTP +REFCLOCK +RTC +PRIVDROP +SCFILTER +SIGND +ASYNCDNS +NTS +SECHASH +IPV6 +DEBUG) Feb 05 04:28:00 ip-10-31-9-239.us-east-1.aws.redhat.com chronyd[6583]: Loaded 0 symmetric keys Feb 05 04:28:00 ip-10-31-9-239.us-east-1.aws.redhat.com chronyd[6583]: Frequency 32.403 +/- 0.749 ppm read from /var/lib/chrony/drift Feb 05 04:28:00 ip-10-31-9-239.us-east-1.aws.redhat.com chronyd[6583]: Using right/UTC timezone to obtain leap second data Feb 05 04:28:00 ip-10-31-9-239.us-east-1.aws.redhat.com chronyd[6583]: Added source 10.2.32.37 Feb 05 04:28:00 ip-10-31-9-239.us-east-1.aws.redhat.com chronyd[6583]: Added source 10.2.32.38 Feb 05 04:28:00 ip-10-31-9-239.us-east-1.aws.redhat.com chronyd[6583]: Added source 10.11.160.238 Feb 05 04:28:00 ip-10-31-9-239.us-east-1.aws.redhat.com chronyd[6583]: Added source 10.18.100.10 Feb 05 04:28:00 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: Started NTP client/server. -- Subject: Unit chronyd.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit chronyd.service has finished starting up. -- -- The start-up result is done. Feb 05 04:28:00 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: Started /usr/bin/systemctl start man-db-cache-update. -- Subject: Unit run-r41794fa330854cdc8f7943d8fa885b8a.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit run-r41794fa330854cdc8f7943d8fa885b8a.service has finished starting up. -- -- The start-up result is done. Feb 05 04:28:00 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: Starting man-db-cache-update.service... -- Subject: Unit man-db-cache-update.service has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit man-db-cache-update.service has begun starting up. Feb 05 04:28:00 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: Reloading. Feb 05 04:28:00 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: /usr/lib/systemd/system/cloud-init.target:15: Unknown lvalue 'ConditionEnvironment' in section 'Unit' Feb 05 04:28:00 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: /usr/lib/systemd/system/cloud-init.service:18: Unknown lvalue 'ConditionEnvironment' in section 'Unit' Feb 05 04:28:00 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: /usr/lib/systemd/system/cloud-init-local.service:12: Unknown lvalue 'ConditionEnvironment' in section 'Unit' Feb 05 04:28:00 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: /usr/lib/systemd/system/cloud-final.service:7: Unknown lvalue 'ConditionEnvironment' in section 'Unit' Feb 05 04:28:00 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: /usr/lib/systemd/system/cloud-config.service:9: Unknown lvalue 'ConditionEnvironment' in section 'Unit' Feb 05 04:28:00 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: Started /usr/bin/systemctl start man-db-cache-update. -- Subject: Unit run-r7e4df4f93f374291b089a5080f1ddf5a.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit run-r7e4df4f93f374291b089a5080f1ddf5a.service has finished starting up. -- -- The start-up result is done. Feb 05 04:28:01 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: man-db-cache-update.service: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit man-db-cache-update.service has successfully entered the 'dead' state. Feb 05 04:28:01 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: Started man-db-cache-update.service. -- Subject: Unit man-db-cache-update.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit man-db-cache-update.service has finished starting up. -- -- The start-up result is done. Feb 05 04:28:01 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: run-r41794fa330854cdc8f7943d8fa885b8a.service: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit run-r41794fa330854cdc8f7943d8fa885b8a.service has successfully entered the 'dead' state. Feb 05 04:28:01 ip-10-31-9-239.us-east-1.aws.redhat.com systemd[1]: run-r7e4df4f93f374291b089a5080f1ddf5a.service: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit run-r7e4df4f93f374291b089a5080f1ddf5a.service has successfully entered the 'dead' state. Feb 05 04:28:01 ip-10-31-9-239.us-east-1.aws.redhat.com python3[7004]: ansible-ansible.legacy.command Invoked with _raw_params=bash -c "chronyc sources" _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 05 04:28:05 ip-10-31-9-239.us-east-1.aws.redhat.com chronyd[6583]: Selected source 162.159.200.123 (2.centos.pool.ntp.org) Feb 05 04:28:05 ip-10-31-9-239.us-east-1.aws.redhat.com chronyd[6583]: System clock TAI offset set to 37 seconds Feb 05 04:28:06 ip-10-31-9-239.us-east-1.aws.redhat.com python3[7028]: ansible-ansible.legacy.command Invoked with _raw_params=bash -c "chronyc sources | grep redhat | grep -v util.phx2.redhat || chronyc add server clock.corp.redhat.com iburst" _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 05 04:28:07 ip-10-31-9-239.us-east-1.aws.redhat.com python3[7052]: ansible-ansible.legacy.command Invoked with _raw_params=bash -c "chronyc sources" _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 05 04:28:07 ip-10-31-9-239.us-east-1.aws.redhat.com python3[7073]: ansible-ansible.legacy.command Invoked with _raw_params=chronyc waitsync 5 _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 05 04:28:07 ip-10-31-9-239.us-east-1.aws.redhat.com python3[7094]: ansible-ansible.legacy.command Invoked with _raw_params=setenforce 0; timedatectl set-timeow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Feb 05 04:28:22 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm python3[8035]: ansible-ansible.legacy.command Invoked with _raw_params=sed -e 's/rpm -q $package/rpm -q --whatprovides $package/' -i /usr/bin/rstrnt-package _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 05 04:28:22 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm python3[8056]: ansible-stat Invoked with path=/var/lib/restraint follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Feb 05 04:28:23 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm python3[8078]: ansible-file Invoked with path=/var/lib/restraint/avc_since state=touch recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Feb 05 04:28:23 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm python3[8098]: ansible-stat Invoked with path=/usr/share/beakerlib/beakerlib.sh follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Feb 05 04:28:23 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm python3[8120]: ansible-file Invoked with dest=/usr/lib/beakerlib state=directory path=/usr/lib/beakerlib recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Feb 05 04:28:23 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm python3[8140]: ansible-file Invoked with dest=/usr/share/rhts-library state=directory path=/usr/share/rhts-library recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Feb 05 04:28:23 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm python3[8160]: ansible-file Invoked with src=/usr/share/beakerlib/beakerlib.sh dest=/usr/lib/beakerlib/beakerlib.sh state=link path=/usr/lib/beakerlib/beakerlib.sh recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Feb 05 04:28:24 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm python3[8180]: ansible-file Invoked with src=/usr/share/beakerlib/beakerlib.sh dest=/usr/share/rhts-library/rhtslib.sh state=link path=/usr/share/rhts-library/rhtslib.sh recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Feb 05 04:28:24 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm python3[8200]: ansible-ansible.legacy.command Invoked with _raw_params=mv /var/log/audit/audit.log /var/log/audit/audit.log.bak _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 05 04:28:24 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm python3[8221]: ansible-ansible.legacy.command Invoked with _raw_params=/usr/sbin/service auditd restart _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 05 04:28:24 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm auditd[560]: The audit daemon is exiting. Feb 05 04:28:24 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm kernel: audit: type=1305 audit(1707107304.683:710): op=set audit_pid=0 old=560 auid=4294967295 ses=4294967295 subj=system_u:system_r:auditd_t:s0 res=1 Feb 05 04:28:24 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm systemd[1]: auditd.service: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit auditd.service has successfully entered the 'dead' state. Feb 05 04:28:24 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm kernel: audit: type=1130 audit(1707107304.689:711): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:init_t:s0 msg='unit=auditd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 05 04:28:24 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm kernel: audit: type=1131 audit(1707107304.689:712): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:init_t:s0 msg='unit=auditd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 05 04:28:26 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm systemd[1]: Starting Security Auditing Service... -- Subject: Unit auditd.service has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit auditd.service has begun starting up. Feb 05 04:28:26 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm auditd[8243]: No plugins found, not dispatching events Feb 05 04:28:26 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm auditd[8243]: Init complete, auditd 3.1.2 listening for events (startup state enable) Feb 05 04:28:26 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm augenrules[8246]: /sbin/augenrules: No change Feb 05 04:28:26 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm augenrules[8256]: No rules Feb 05 04:28:26 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm augenrules[8256]: enabled 1 Feb 05 04:28:26 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm augenrules[8256]: failure 1 Feb 05 04:28:26 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm augenrules[8256]: pid 8243 Feb 05 04:28:26 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm augenrules[8256]: rate_limit 0 Feb 05 04:28:26 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm augenrules[8256]: backlog_limit 8192 Feb 05 04:28:26 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm augenrules[8256]: lost 0 Feb 05 04:28:26 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm augenrules[8256]: backlog 4 Feb 05 04:28:26 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm augenrules[8256]: backlog_wait_time 60000 Feb 05 04:28:26 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm augenrules[8256]: backlog_wait_time_actual 0 Feb 05 04:28:26 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm augenrules[8256]: enabled 1 Feb 05 04:28:26 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm augenrules[8256]: failure 1 Feb 05 04:28:26 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm augenrules[8256]: pid 8243 Feb 05 04:28:26 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm augenrules[8256]: rate_limit 0 Feb 05 04:28:26 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm augenrules[8256]: backlog_limit 8192 Feb 05 04:28:26 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm augenrules[8256]: lost 0 Feb 05 04:28:26 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm augenrules[8256]: backlog 1 Feb 05 04:28:26 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm augenrules[8256]: backlog_wait_time 60000 Feb 05 04:28:26 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm augenrules[8256]: backlog_wait_time_actual 0 Feb 05 04:28:26 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm augenrules[8256]: enabled 1 Feb 05 04:28:26 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm augenrules[8256]: failure 1 Feb 05 04:28:26 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm augenrules[8256]: pid 8243 Feb 05 04:28:26 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm augenrules[8256]: rate_limit 0 Feb 05 04:28:26 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm augenrules[8256]: backlog_limit 8192 Feb 05 04:28:26 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm augenrules[8256]: lost 0 Feb 05 04:28:26 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm augenrules[8256]: backlog 0 Feb 05 04:28:26 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm augenrules[8256]: backlog_wait_time 60000 Feb 05 04:28:26 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm augenrules[8256]: backlog_wait_time_actual 0 Feb 05 04:28:26 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm systemd[1]: Started Security Auditing Service. -- Subject: Unit auditd.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit auditd.service has finished starting up. -- -- The start-up result is done. Feb 05 04:28:26 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm python3[8281]: ansible-stat Invoked with path=/etc/NetworkManager/conf.d follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Feb 05 04:28:27 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm python3[8345]: ansible-ansible.legacy.stat Invoked with path=/etc/NetworkManager/conf.d/99-zzz-tft-workaround-dns-default.conf follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Feb 05 04:28:27 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm python3[8402]: ansible-ansible.legacy.copy Invoked with dest=/etc/NetworkManager/conf.d/99-zzz-tft-workaround-dns-default.conf src=/root/.ansible/tmp/ansible-tmp-1707107306.9737978-7401-242002235029711/source _original_basename=tmpa0svqt5s follow=False checksum=0145f3ae57eef5aa08bbb678fedbb3edd001cd2d backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Feb 05 04:28:27 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm python3[8443]: ansible-ansible.legacy.command Invoked with _raw_params=for repofile in $(ls -1 /etc/yum.repos.d/); do echo "# vvvvv $repofile vvvvv vvvvv"; cat "/etc/yum.repos.d/$repofile"; echo "# ^^^^^ ^^^^^ ^^^^^ ^^^^^"; echo ""; done _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 05 04:28:28 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm python3[8486]: ansible-ansible.legacy.command Invoked with _raw_params=yum -q repolist _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 05 04:28:28 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm python3[8508]: ansible-ansible.legacy.command Invoked with _raw_params=dnf config-manager --set-enabled ha _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 05 04:28:29 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm python3[8530]: ansible-ansible.legacy.command Invoked with _raw_params=dnf clean all --enablerepo=* _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 05 04:28:29 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm python3[8552]: ansible-ansible.legacy.command Invoked with _raw_params=dnf makecache --enablerepo=* _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 05 04:28:31 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm systemd[1]: NetworkManager-dispatcher.service: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit NetworkManager-dispatcher.service has successfully entered the 'dead' state. Feb 05 04:28:44 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm systemd[1]: timedatex.service: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit timedatex.service has successfully entered the 'dead' state. Feb 05 04:28:51 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm systemd[1]: systemd-hostnamed.service: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit systemd-hostnamed.service has successfully entered the 'dead' state. Feb 05 04:29:11 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm chronyd[6574]: Selected source 10.2.32.38 Feb 05 04:30:35 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm python3[8769]: ansible-ansible.legacy.setup Invoked with gather_subset=['all'] gather_timeout=10 filter=[] fact_path=/etc/ansible/facts.d Feb 05 04:30:36 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm python3[8816]: ansible-service_facts Invoked Feb 05 04:30:38 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm python3[8922]: ansible-ansible.legacy.command Invoked with _raw_params=rpm -qa _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 05 04:30:39 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm python3[8943]: ansible-ansible.legacy.command Invoked with _raw_params=if [ $(find /etc/yum.repos.d/ -name 'epel.repo' | wc -l ) -gt 0 ]; then dnf config-manager --set-disabled epel; fi _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 05 04:30:40 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm python3[8969]: ansible-ansible.legacy.command Invoked with _raw_params=dnf clean all --enablerepo=* _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 05 04:30:41 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm python3[8991]: ansible-ansible.legacy.command Invoked with _raw_params=dnf makecache --enablerepo=* _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 05 04:32:57 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm python3[9200]: ansible-ansible.legacy.setup Invoked with gather_subset=['all'] gather_timeout=10 filter=[] fact_path=/etc/ansible/facts.d Feb 05 04:32:58 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm python3[9247]: ansible-service_facts Invoked Feb 05 04:33:00 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm python3[9353]: ansible-ansible.legacy.command Invoked with _raw_params=rpm -qa _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 05 04:33:02 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm python3[9397]: ansible-ansible.legacy.setup Invoked with gather_subset=['all'] gather_timeout=10 filter=[] fact_path=/etc/ansible/facts.d Feb 05 04:33:03 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm python3[9444]: ansible-ansible.legacy.dnf Invoked with name=['ca-certificates'] state=latest allow_downgrade=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 allowerasing=False nobest=False conf_file=None disable_excludes=None download_dir=None list=None releasever=None Feb 05 04:33:06 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm python3[9465]: ansible-ansible.legacy.dnf Invoked with name=['curl'] state=latest allow_downgrade=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 allowerasing=False nobest=False conf_file=None disable_excludes=None download_dir=None list=None releasever=None Feb 05 04:33:08 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm python3[9486]: ansible-ansible.legacy.command Invoked with _raw_params=curl -skL -o /etc/pki/ca-trust/source/anchors/Current-IT-Root-CAs.pem https://certs.corp.redhat.com/certs/Current-IT-Root-CAs.pem _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 05 04:33:08 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm python3[9508]: ansible-ansible.legacy.command Invoked with _raw_params=bash -c 'update-ca-trust' _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 05 04:33:09 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm python3[9535]: ansible-ansible.legacy.command Invoked with _raw_params=bash -c 'update-ca-trust extract' _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 05 04:33:11 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm python3[9562]: ansible-ansible.legacy.command Invoked with _raw_params=curl -v https://gitlab.cee.redhat.com _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 05 04:33:11 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm python3[9584]: ansible-ansible.legacy.command Invoked with _raw_params=curl -v https://beaker.engineering.redhat.com/ _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 05 04:33:12 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm python3[9606]: ansible-service_facts Invoked Feb 05 04:33:14 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm python3[9712]: ansible-ansible.legacy.command Invoked with _raw_params=rpm -qa _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 05 04:33:15 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm python3[9733]: ansible-service_facts Invoked Feb 05 04:33:17 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm python3[9839]: ansible-ansible.legacy.command Invoked with _raw_params=rpm -qa _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 05 04:33:18 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm python3[9860]: ansible-ansible.legacy.command Invoked with _raw_params=for repofile in $(ls -1 /etc/yum.repos.d/); do echo "# vvvvv $repofile vvvvv vvvvv"; cat "/etc/yum.repos.d/$repofile"; echo "# ^^^^^ ^^^^^ ^^^^^ ^^^^^"; echo ""; done _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 05 04:33:18 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm python3[9903]: ansible-ansible.legacy.command Invoked with _raw_params=yum -q repolist _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 05 04:33:19 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm python3[9925]: ansible-ansible.legacy.command Invoked with _raw_params=dnf clean all --enablerepo=* _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 05 04:33:19 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm python3[9947]: ansible-ansible.legacy.command Invoked with _raw_params=dnf makecache --enablerepo=* _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 05 04:35:38 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm python3[10158]: ansible-setup Invoked with gather_subset=['all'] gather_timeout=10 filter=* fact_path=/etc/ansible/facts.d Feb 05 04:35:39 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm python3[10203]: ansible-stat Invoked with path=/run/ostree-booted follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Feb 05 04:35:39 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm python3[10223]: ansible-dnf Invoked with name=['python3-pyasn1', 'python3-cryptography', 'python3-dbus'] state=present allow_downgrade=False autoremove=False bugfix=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True lock_timeout=30 conf_file=None disable_excludes=None download_dir=None list=None releasever=None Feb 05 04:35:43 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm python3[10248]: ansible-dnf Invoked with name=['certmonger'] state=present allow_downgrade=False autoremove=False bugfix=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True lock_timeout=30 conf_file=None disable_excludes=None download_dir=None list=None releasever=None Feb 05 04:35:45 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm dbus-daemon[622]: [system] Reloaded configuration Feb 05 04:35:45 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm dbus-daemon[622]: [system] Reloaded configuration Feb 05 04:35:45 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm dbus-daemon[622]: [system] Reloaded configuration Feb 05 04:35:45 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm dbus-daemon[622]: [system] Reloaded configuration Feb 05 04:35:45 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm dbus-daemon[622]: [system] Reloaded configuration Feb 05 04:35:45 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm systemd[1]: Reloading. Feb 05 04:35:46 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm systemd[1]: /usr/lib/systemd/system/cloud-init.target:15: Unknown lvalue 'ConditionEnvironment' in section 'Unit' Feb 05 04:35:46 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm systemd[1]: /usr/lib/systemd/system/cloud-init.service:18: Unknown lvalue 'ConditionEnvironment' in section 'Unit' Feb 05 04:35:46 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm systemd[1]: /usr/lib/systemd/system/cloud-init-local.service:12: Unknown lvalue 'ConditionEnvironment' in section 'Unit' Feb 05 04:35:46 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm systemd[1]: /usr/lib/systemd/system/cloud-final.service:7: Unknown lvalue 'ConditionEnvironment' in section 'Unit' Feb 05 04:35:46 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm systemd[1]: /usr/lib/systemd/system/cloud-config.service:9: Unknown lvalue 'ConditionEnvironment' in section 'Unit' Feb 05 04:35:46 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm systemd[1]: Started /usr/bin/systemctl start man-db-cache-update. -- Subject: Unit run-r3d5e2a18afbe413c902e6d4b8680e27a.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit run-r3d5e2a18afbe413c902e6d4b8680e27a.service has finished starting up. -- -- The start-up result is done. Feb 05 04:35:46 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm systemd[1]: Starting man-db-cache-update.service... -- Subject: Unit man-db-cache-update.service has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit man-db-cache-update.service has begun starting up. Feb 05 04:35:46 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm systemd[1]: Reloading. Feb 05 04:35:46 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm systemd[1]: /usr/lib/systemd/system/cloud-init.target:15: Unknown lvalue 'ConditionEnvironment' in section 'Unit' Feb 05 04:35:46 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm systemd[1]: /usr/lib/systemd/system/cloud-init.service:18: Unknown lvalue 'ConditionEnvironment' in section 'Unit' Feb 05 04:35:46 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm systemd[1]: /usr/lib/systemd/system/cloud-init-local.service:12: Unknown lvalue 'ConditionEnvironment' in section 'Unit' Feb 05 04:35:46 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm systemd[1]: /usr/lib/systemd/system/cloud-final.service:7: Unknown lvalue 'ConditionEnvironment' in section 'Unit' Feb 05 04:35:46 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm systemd[1]: /usr/lib/systemd/system/cloud-config.service:9: Unknown lvalue 'ConditionEnvironment' in section 'Unit' Feb 05 04:35:46 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm systemd[1]: man-db-cache-update.service: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit man-db-cache-update.service has successfully entered the 'dead' state. Feb 05 04:35:46 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm systemd[1]: Started man-db-cache-update.service. -- Subject: Unit man-db-cache-update.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit man-db-cache-update.service has finished starting up. -- -- The start-up result is done. Feb 05 04:35:46 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm systemd[1]: run-r3d5e2a18afbe413c902e6d4b8680e27a.service: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit run-r3d5e2a18afbe413c902e6d4b8680e27a.service has successfully entered the 'dead' state. Feb 05 04:35:47 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm python3[10772]: ansible-file Invoked with name=/etc/certmonger//pre-scripts owner=root group=root mode=0700 state=directory path=/etc/certmonger//pre-scripts recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Feb 05 04:35:47 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm python3[10792]: ansible-file Invoked with name=/etc/certmonger//post-scripts owner=root group=root mode=0700 state=directory path=/etc/certmonger//post-scripts recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Feb 05 04:35:47 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm python3[10812]: ansible-systemd Invoked with name=certmonger state=started enabled=True daemon_reload=False daemon_reexec=False no_block=False force=None masked=None user=None scope=None Feb 05 04:35:47 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm systemd[1]: Reloading. Feb 05 04:35:48 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm systemd[1]: /usr/lib/systemd/system/cloud-init.target:15: Unknown lvalue 'ConditionEnvironment' in section 'Unit' Feb 05 04:35:48 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm systemd[1]: /usr/lib/systemd/system/cloud-init.service:18: Unknown lvalue 'ConditionEnvironment' in section 'Unit' Feb 05 04:35:48 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm systemd[1]: /usr/lib/systemd/system/cloud-init-local.service:12: Unknown lvalue 'ConditionEnvironment' in section 'Unit' Feb 05 04:35:48 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm systemd[1]: /usr/lib/systemd/system/cloud-final.service:7: Unknown lvalue 'ConditionEnvironment' in section 'Unit' Feb 05 04:35:48 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm systemd[1]: /usr/lib/systemd/system/cloud-config.service:9: Unknown lvalue 'ConditionEnvironment' in section 'Unit' Feb 05 04:35:48 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm systemd[1]: Starting Certificate monitoring and PKI enrollment... -- Subject: Unit certmonger.service has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit certmonger.service has begun starting up. Feb 05 04:35:48 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm systemd[1]: Started Certificate monitoring and PKI enrollment. -- Subject: Unit certmonger.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit certmonger.service has finished starting up. -- -- The start-up result is done. Feb 05 04:35:48 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm python3[10893]: ansible-fedora.linux_system_roles.certificate_request Invoked with name=quadlet_demo dns=['localhost'] directory=/etc/pki/tls wait=True ca=self-sign __header=# # Ansible managed # # system_role:certificate provider_config_directory=/etc/certmonger provider=certmonger key_usage=['digitalSignature', 'keyEncipherment'] extended_key_usage=['id-kp-serverAuth', 'id-kp-clientAuth'] auto_renew=True ip=None email=None common_name=None country=None state=None locality=None organization=None organizational_unit=None contact_email=None key_size=None owner=None group=None mode=None principal=None run_before=None run_after=None Feb 05 04:35:48 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm certmonger[10842]: 2024-02-05 04:35:48 [10842] Wrote to /var/lib/certmonger/requests/20240205043548 Feb 05 04:35:48 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm certmonger[10842]: 2024-02-05 04:35:48 [10842] Wrote to /var/lib/certmonger/requests/20240205043548 Feb 05 04:35:48 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm rsyslogd[1023]: imjournal: journal files changed, reloading... [v8.2102.0-15.el8 try https://www.rsyslog.com/e/0 ] Feb 05 04:35:48 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm certmonger[10842]: 2024-02-05 04:35:48 [10842] Wrote to /var/lib/certmonger/requests/20240205043548 Feb 05 04:35:48 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm certmonger[10842]: 2024-02-05 04:35:48 [10842] Wrote to /var/lib/certmonger/requests/20240205043548 Feb 05 04:35:48 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm certmonger[10842]: 2024-02-05 04:35:48 [10842] Wrote to /var/lib/certmonger/requests/20240205043548 Feb 05 04:35:48 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm certmonger[10842]: 2024-02-05 04:35:48 [10842] Wrote to /var/lib/certmonger/requests/20240205043548 Feb 05 04:35:48 dbf27722-21c5-42d7-a04d-fe5d7cfbbc1d.testing-farm certmonger[10842]: 2024-02-05 04:35:48 [10842] Wrote to /var/lib/certmonger/requests/20240205043548 Feb 05 04:35:48 dbf27722-21c5-hat.com/ _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 05 04:33:03 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm python3[9616]: ansible-service_facts Invoked Feb 05 04:33:04 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm python3[9722]: ansible-ansible.legacy.command Invoked with _raw_params=rpm -qa _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 05 04:33:06 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm python3[9743]: ansible-service_facts Invoked Feb 05 04:33:07 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm python3[9849]: ansible-ansible.legacy.command Invoked with _raw_params=rpm -qa _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 05 04:33:09 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm python3[9870]: ansible-ansible.legacy.command Invoked with _raw_params=for repofile in $(ls -1 /etc/yum.repos.d/); do echo "# vvvvv $repofile vvvvv vvvvv"; cat "/etc/yum.repos.d/$repofile"; echo "# ^^^^^ ^^^^^ ^^^^^ ^^^^^"; echo ""; done _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 05 04:33:09 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm python3[9913]: ansible-ansible.legacy.command Invoked with _raw_params=yum -q repolist _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 05 04:33:09 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm python3[9935]: ansible-ansible.legacy.command Invoked with _raw_params=dnf clean all --enablerepo=* _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 05 04:33:10 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm python3[9957]: ansible-ansible.legacy.command Invoked with _raw_params=dnf makecache --enablerepo=* _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 05 04:35:38 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm python3[10168]: ansible-ansible.legacy.setup Invoked with gather_subset=['all'] gather_timeout=10 filter=[] fact_path=/etc/ansible/facts.d Feb 05 04:35:38 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm python3[10217]: ansible-stat Invoked with path=/run/ostree-booted follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Feb 05 04:35:39 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm python3[10237]: ansible-ansible.legacy.dnf Invoked with name=['python3-pyasn1', 'python3-cryptography', 'python3-dbus'] state=present allow_downgrade=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 allowerasing=False nobest=False use_backend=auto conf_file=None disable_excludes=None download_dir=None list=None releasever=None Feb 05 04:35:42 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm python3[10263]: ansible-ansible.legacy.dnf Invoked with name=['certmonger'] state=present allow_downgrade=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 allowerasing=False nobest=False use_backend=auto conf_file=None disable_excludes=None download_dir=None list=None releasever=None Feb 05 04:35:45 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm dbus-daemon[628]: [system] Reloaded configuration Feb 05 04:35:45 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm dbus-daemon[628]: [system] Reloaded configuration Feb 05 04:35:45 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm dbus-daemon[628]: [system] Reloaded configuration Feb 05 04:35:45 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm dbus-daemon[628]: [system] Reloaded configuration Feb 05 04:35:45 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm dbus-daemon[628]: [system] Reloaded configuration Feb 05 04:35:45 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm systemd[1]: Reloading. Feb 05 04:35:45 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm systemd[1]: /usr/lib/systemd/system/cloud-init.target:15: Unknown lvalue 'ConditionEnvironment' in section 'Unit' Feb 05 04:35:45 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm systemd[1]: /usr/lib/systemd/system/cloud-init.service:18: Unknown lvalue 'ConditionEnvironment' in section 'Unit' Feb 05 04:35:45 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm systemd[1]: /usr/lib/systemd/system/cloud-init-local.service:12: Unknown lvalue 'ConditionEnvironment' in section 'Unit' Feb 05 04:35:45 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm systemd[1]: /usr/lib/systemd/system/cloud-final.service:7: Unknown lvalue 'ConditionEnvironment' in section 'Unit' Feb 05 04:35:45 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm systemd[1]: /usr/lib/systemd/system/cloud-config.service:9: Unknown lvalue 'ConditionEnvironment' in section 'Unit' Feb 05 04:35:45 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm systemd[1]: Started /usr/bin/systemctl start man-db-cache-update. -- Subject: Unit run-r171c5213cc8240f7aa3ebaa21ec93406.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit run-r171c5213cc8240f7aa3ebaa21ec93406.service has finished starting up. -- -- The start-up result is done. Feb 05 04:35:45 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm systemd[1]: Starting man-db-cache-update.service... -- Subject: Unit man-db-cache-update.service has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit man-db-cache-update.service has begun starting up. Feb 05 04:35:45 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm systemd[1]: Reloading. Feb 05 04:35:46 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm systemd[1]: /usr/lib/systemd/system/cloud-init.target:15: Unknown lvalue 'ConditionEnvironment' in section 'Unit' Feb 05 04:35:46 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm systemd[1]: /usr/lib/systemd/system/cloud-init.service:18: Unknown lvalue 'ConditionEnvironment' in section 'Unit' Feb 05 04:35:46 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm systemd[1]: /usr/lib/systemd/system/cloud-init-local.service:12: Unknown lvalue 'ConditionEnvironment' in section 'Unit' Feb 05 04:35:46 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm systemd[1]: /usr/lib/systemd/system/cloud-final.service:7: Unknown lvalue 'ConditionEnvironment' in section 'Unit' Feb 05 04:35:46 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm systemd[1]: /usr/lib/systemd/system/cloud-config.service:9: Unknown lvalue 'ConditionEnvironment' in section 'Unit' Feb 05 04:35:46 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm systemd[1]: man-db-cache-update.service: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit man-db-cache-update.service has successfully entered the 'dead' state. Feb 05 04:35:46 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm systemd[1]: Started man-db-cache-update.service. -- Subject: Unit man-db-cache-update.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit man-db-cache-update.service has finished starting up. -- -- The start-up result is done. Feb 05 04:35:46 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm systemd[1]: run-r171c5213cc8240f7aa3ebaa21ec93406.service: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit run-r171c5213cc8240f7aa3ebaa21ec93406.service has successfully entered the 'dead' state. Feb 05 04:35:46 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm python3[10787]: ansible-file Invoked with name=/etc/certmonger//pre-scripts owner=root group=root mode=0700 state=directory path=/etc/certmonger//pre-scripts recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Feb 05 04:35:47 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm python3[10807]: ansible-file Invoked with name=/etc/certmonger//post-scripts owner=root group=root mode=0700 state=directory path=/etc/certmonger//post-scripts recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Feb 05 04:35:47 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm python3[10827]: ansible-ansible.legacy.systemd Invoked with name=certmonger state=started enabled=True daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Feb 05 04:35:47 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm systemd[1]: Reloading. Feb 05 04:35:47 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm systemd[1]: /usr/lib/systemd/system/cloud-init.target:15: Unknown lvalue 'ConditionEnvironment' in section 'Unit' Feb 05 04:35:47 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm systemd[1]: /usr/lib/systemd/system/cloud-init.service:18: Unknown lvalue 'ConditionEnvironment' in section 'Unit' Feb 05 04:35:47 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm systemd[1]: /usr/lib/systemd/system/cloud-init-local.service:12: Unknown lvalue 'ConditionEnvironment' in section 'Unit' Feb 05 04:35:47 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm systemd[1]: /usr/lib/systemd/system/cloud-final.service:7: Unknown lvalue 'ConditionEnvironment' in section 'Unit' Feb 05 04:35:47 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm systemd[1]: /usr/lib/systemd/system/cloud-config.service:9: Unknown lvalue 'ConditionEnvironment' in section 'Unit' Feb 05 04:35:47 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm systemd[1]: Starting Certificate monitoring and PKI enrollment... -- Subject: Unit certmonger.service has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit certmonger.service has begun starting up. Feb 05 04:35:47 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm systemd[1]: Started Certificate monitoring and PKI enrollment. -- Subject: Unit certmonger.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit certmonger.service has finished starting up. -- -- The start-up result is done. Feb 05 04:35:48 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm python3[10908]: ansible-fedora.linux_system_roles.certificate_request Invoked with name=quadlet_demo dns=['localhost'] directory=/etc/pki/tls wait=True ca=self-sign __header=# # Ansible managed # # system_role:certificate provider_config_directory=/etc/certmonger provider=certmonger key_usage=['digitalSignature', 'keyEncipherment'] extended_key_usage=['id-kp-serverAuth', 'id-kp-clientAuth'] auto_renew=True ip=None email=None common_name=None country=None state=None locality=None organization=None organizational_unit=None contact_email=None key_size=None owner=None group=None mode=None principal=None run_before=None run_after=None Feb 05 04:35:48 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm certmonger[10857]: 2024-02-05 04:35:48 [10857] Wrote to /var/lib/certmonger/requests/20240205043548 Feb 05 04:35:48 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm certmonger[10857]: 2024-02-05 04:35:48 [10857] Wrote to /var/lib/certmonger/requests/20240205043548 Feb 05 04:35:48 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm rsyslogd[1034]: imjournal: journal files changed, reloading... [v8.2102.0-15.el8 try https://www.rsyslog.com/e/0 ] Feb 05 04:35:48 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm certmonger[10857]: 2024-02-05 04:35:48 [10857] Wrote to /var/lib/certmonger/requests/20240205043548 Feb 05 04:35:48 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm certmonger[10857]: 2024-02-05 04:35:48 [10857] Wrote to /var/lib/certmonger/requests/20240205043548 Feb 05 04:35:48 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm certmonger[10857]: 2024-02-05 04:35:48 [10857] Wrote to /var/lib/certmonger/requests/20240205043548 Feb 05 04:35:48 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm certmonger[10857]: 2024-02-05 04:35:48 [10857] Wrote to /var/lib/certmonger/requests/20240205043548 Feb 05 04:35:48 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm certmonger[10857]: 2024-02-05 04:35:48 [10857] Wrote to /var/lib/certmonger/requests/20240205043548 Feb 05 04:35:48 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm certmonger[10857]: 2024-02-05 04:35:48 [10857] Wrote to /var/lib/certmonger/requests/20240205043548 Feb 05 04:35:48 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm certmonger[10857]: 2024-02-05 04:35:48 [10857] Wrote to /var/lib/certmonger/requests/20240205043548 Feb 05 04:35:48 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm certmonger[10857]: 2024-02-05 04:35:48 [10857] Wrote to /var/lib/certmonger/requests/20240205043548 Feb 05 04:35:48 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm certmonger[10857]: 2024-02-05 04:35:48 [10857] Wrote to /var/lib/certmonger/requests/20240205043548 Feb 05 04:35:48 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm certmonger[10857]: 2024-02-05 04:35:48 [10857] Wrote to /var/lib/certmonger/requests/20240205043548 Feb 05 04:35:48 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm certmonger[10857]: 2024-02-05 04:35:48 [10857] Wrote to /var/lib/certmonger/requests/20240205043548 Feb 05 04:35:48 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm certmonger[10857]: 2024-02-05 04:35:48 [10857] Wrote to /var/lib/certmonger/requests/20240205043548 Feb 05 04:35:48 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm certmonger[10857]: 2024-02-05 04:35:48 [10857] Wrote to /var/lib/certmonger/requests/20240205043548 Feb 05 04:35:48 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm certmonger[10857]: 2024-02-05 04:35:48 [10857] Wrote to /var/lib/certmonger/requests/20240205043548 Feb 05 04:35:48 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm certmonger[10857]: 2024-02-05 04:35:48 [10857] Wrote to /var/lib/certmonger/requests/20240205043548 Feb 05 04:35:48 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm certmonger[10857]: 2024-02-05 04:35:48 [10857] Wrote to /var/lib/certmonger/requests/20240205043548 Feb 05 04:35:48 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm certmonger[10857]: 2024-02-05 04:35:48 [10857] Wrote to /var/lib/certmonger/requests/20240205043548 Feb 05 04:35:48 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm certmonger[10857]: 2024-02-05 04:35:48 [10857] Wrote to /var/lib/certmonger/requests/20240205043548 Feb 05 04:35:48 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm certmonger[10857]: 2024-02-05 04:35:48 [10857] Wrote to /var/lib/certmonger/requests/20240205043548 Feb 05 04:35:48 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm certmonger[10857]: 2024-02-05 04:35:48 [10857] Wrote to /var/lib/certmonger/requests/20240205043548 Feb 05 04:35:48 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm certmonger[10857]: 2024-02-05 04:35:48 [10857] Wrote to /var/lib/certmonger/requests/20240205043548 Feb 05 04:35:48 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm certmonger[10857]: 2024-02-05 04:35:48 [10857] Wrote to /var/lib/certmonger/requests/20240205043548 Feb 05 04:35:48 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm certmonger[10857]: 2024-02-05 04:35:48 [10857] Wrote to /var/lib/certmonger/requests/20240205043548 Feb 05 04:35:48 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm certmonger[10928]: Certificate in file "/etc/pki/tls/certs/quadlet_demo.crt" issued by CA and saved. Feb 05 04:35:48 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm certmonger[10857]: 2024-02-05 04:35:48 [10857] Wrote to /var/lib/certmonger/requests/20240205043548 Feb 05 04:35:48 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm python3[10947]: ansible-slurp Invoked with path=/etc/pki/tls/certs/quadlet_demo.crt src=/etc/pki/tls/certs/quadlet_demo.crt Feb 05 04:35:49 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm python3[10967]: ansible-slurp Invoked with path=/etc/pki/tls/private/quadlet_demo.key src=/etc/pki/tls/private/quadlet_demo.key Feb 05 04:35:49 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm python3[10987]: ansible-slurp Invoked with path=/etc/pki/tls/certs/quadlet_demo.crt src=/etc/pki/tls/certs/quadlet_demo.crt Feb 05 04:35:49 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm python3[11007]: ansible-ansible.legacy.command Invoked with _raw_params=getcert stop-tracking -f /etc/pki/tls/certs/quadlet_demo.crt _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 05 04:35:49 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm certmonger[10857]: 2024-02-05 04:35:49 [10857] Wrote to /var/lib/certmonger/requests/20240205043548 Feb 05 04:35:49 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm python3[11028]: ansible-file Invoked with path=/etc/pki/tls/certs/quadlet_demo.crt state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Feb 05 04:35:49 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm python3[11048]: ansible-file Invoked with path=/etc/pki/tls/private/quadlet_demo.key state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Feb 05 04:35:50 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm python3[11068]: ansible-file Invoked with path=/etc/pki/tls/certs/quadlet_demo.crt state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Feb 05 04:35:50 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm python3[11088]: ansible-stat Invoked with path=/run/ostree-booted follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Feb 05 04:35:52 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm python3[11130]: ansible-ansible.legacy.dnf Invoked with name=['crun', 'podman', 'podman-plugins', 'shadow-utils-subid'] state=present allow_downgrade=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 allowerasing=False nobest=False use_backend=auto conf_file=None disable_excludes=None download_dir=None list=None releasever=None Feb 05 04:36:00 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm dbus-daemon[628]: [system] Reloaded configuration Feb 05 04:36:01 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm setsebool[11165]: The virt_use_nfs policy boolean was changed to 1 by root Feb 05 04:36:01 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm setsebool[11165]: The virt_sandbox_use_all_caps policy boolean was changed to 1 by root Feb 05 04:36:19 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm kernel: SELinux: Converting 389 SID table entries... Feb 05 04:36:19 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm kernel: SELinux: policy capability network_peer_controls=1 Feb 05 04:36:19 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm kernel: SELinux: policy capability open_perms=1 Feb 05 04:36:19 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm kernel: SELinux: policy capability extended_socket_class=1 Feb 05 04:36:19 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm kernel: SELinux: policy capability always_check_network=0 Feb 05 04:36:19 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm kernel: SELinux: policy capability cgroup_seclabel=1 Feb 05 04:36:19 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm kernel: SELinux: policy capability nnp_nosuid_transition=1 Feb 05 04:36:19 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm dbus-daemon[628]: [system] Reloaded configuration Feb 05 04:36:19 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm kernel: fuse: init (API version 7.34) Feb 05 04:36:19 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm systemd[1]: Mounting FUSE Control File System... -- Subject: Unit sys-fs-fuse-connections.mount has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit sys-fs-fuse-connections.mount has begun starting up. Feb 05 04:36:19 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm systemd[1]: Mounted FUSE Control File System. -- Subject: Unit sys-fs-fuse-connections.mount has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit sys-fs-fuse-connections.mount has finished starting up. -- -- The start-up result is done. Feb 05 04:36:20 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm dbus-daemon[628]: [system] Reloaded configuration Feb 05 04:36:20 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm dbus-daemon[628]: [system] Reloaded configuration Feb 05 04:36:50 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm systemd[1]: Started /usr/bin/systemctl start man-db-cache-update. -- Subject: Unit run-r0aa8ed571cbb47118a58af695bd60834.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit run-r0aa8ed571cbb47118a58af695bd60834.service has finished starting up. -- -- The start-up result is done. Feb 05 04:36:50 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm systemd[1]: Starting man-db-cache-update.service... -- Subject: Unit man-db-cache-update.service has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit man-db-cache-update.service has begun starting up. Feb 05 04:36:50 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm systemd[1]: Reloading. Feb 05 04:36:50 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm systemd[1]: /usr/lib/systemd/system/cloud-init.target:15: Unknown lvalue 'ConditionEnvironment' in section 'Unit' Feb 05 04:36:50 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm systemd[1]: /usr/lib/systemd/system/cloud-init.service:18: Unknown lvalue 'ConditionEnvironment' in section 'Unit' Feb 05 04:36:50 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm systemd[1]: /usr/lib/systemd/system/cloud-init-local.service:12: Unknown lvalue 'ConditionEnvironment' in section 'Unit' Feb 05 04:36:50 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm systemd[1]: /usr/lib/systemd/system/cloud-final.service:7: Unknown lvalue 'ConditionEnvironment' in section 'Unit' Feb 05 04:36:50 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm systemd[1]: /usr/lib/systemd/system/cloud-config.service:9: Unknown lvalue 'ConditionEnvironment' in section 'Unit' Feb 05 04:36:51 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm python3[13274]: ansible-ansible.legacy.command Invoked with _raw_params=podman --version _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 05 04:36:52 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm systemd[1]: man-db-cache-update.service: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit man-db-cache-update.service has successfully entered the 'dead' state. Feb 05 04:36:52 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm systemd[1]: Started man-db-cache-update.service. -- Subject: Unit man-db-cache-update.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit man-db-cache-update.service has finished starting up. -- -- The start-up result is done. Feb 05 04:36:52 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm systemd[1]: run-r0aa8ed571cbb47118a58af695bd60834.service: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit run-r0aa8ed571cbb47118a58af695bd60834.service has successfully entered the 'dead' state. Feb 05 04:36:52 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm python3[13580]: ansible-getent Invoked with database=passwd key=root fail_key=False service=None split=None Feb 05 04:36:53 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm python3[13601]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None Feb 05 04:36:53 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm python3[13622]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Feb 05 04:36:54 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm python3[13644]: ansible-stat Invoked with path=/run/ostree-booted follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Feb 05 04:36:54 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm python3[13664]: ansible-ansible.legacy.dnf Invoked with name=['firewalld'] state=present allow_downgrade=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 allowerasing=False nobest=False use_backend=auto conf_file=None disable_excludes=None download_dir=None list=None releasever=None Feb 05 04:36:57 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm python3[13685]: ansible-systemd Invoked with name=firewalld masked=False daemon_reload=False daemon_reexec=False scope=system no_block=False state=None enabled=None force=None Feb 05 04:36:57 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm python3[13709]: ansible-ansible.legacy.systemd Invoked with name=firewalld state=started enabled=True daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Feb 05 04:36:57 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm systemd[1]: Reloading. Feb 05 04:36:57 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm systemd[1]: /usr/lib/systemd/system/cloud-init.target:15: Unknown lvalue 'ConditionEnvironment' in section 'Unit' Feb 05 04:36:57 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm systemd[1]: /usr/lib/systemd/system/cloud-init.service:18: Unknown lvalue 'ConditionEnvironment' in section 'Unit' Feb 05 04:36:57 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm systemd[1]: /usr/lib/systemd/system/cloud-init-local.service:12: Unknown lvalue 'ConditionEnvironment' in section 'Unit' Feb 05 04:36:57 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm systemd[1]: /usr/lib/systemd/system/cloud-final.service:7: Unknown lvalue 'ConditionEnvironment' in section 'Unit' Feb 05 04:36:57 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm systemd[1]: /usr/lib/systemd/system/cloud-config.service:9: Unknown lvalue 'ConditionEnvironment' in section 'Unit' Feb 05 04:36:57 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm systemd[1]: Starting firewalld - dynamic firewall daemon... -- Subject: Unit firewalld.service has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit firewalld.service has begun starting up. Feb 05 04:36:58 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm systemd[1]: Started firewalld - dynamic firewall daemon. -- Subject: Unit firewalld.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit firewalld.service has finished starting up. -- -- The start-up result is done. Feb 05 04:36:58 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm firewalld[13744]: WARNING: AllowZoneDrifting is enabled. This is considered an insecure configuration option. It will be removed in a future release. Please consider disabling it now. Feb 05 04:36:58 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm python3[13774]: ansible-fedora.linux_system_roles.firewall_lib Invoked with port=['8000/tcp'] permanent=True runtime=True state=enabled __report_changed=True service=[] source_port=[] forward_port=[] rich_rule=[] source=[] interface=[] interface_pci_id=[] icmp_block=[] timeout=0 ipset_entries=[] protocol=[] helper_module=[] destination=[] firewalld_conf=None masquerade=None icmp_block_inversion=None target=None zone=None set_default_zone=None ipset=None ipset_type=None description=None short=None Feb 05 04:36:59 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm python3[13852]: ansible-fedora.linux_system_roles.firewall_lib Invoked with port=['9000/tcp'] permanent=True runtime=True state=enabled __report_changed=True service=[] source_port=[] forward_port=[] rich_rule=[] source=[] interface=[] interface_pci_id=[] icmp_block=[] timeout=0 ipset_entries=[] protocol=[] helper_module=[] destination=[] firewalld_conf=None masquerade=None icmp_block_inversion=None target=None zone=None set_default_zone=None ipset=None ipset_type=None description=None short=None Feb 05 04:37:01 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm python3[13950]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None Feb 05 04:37:01 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm python3[13971]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Feb 05 04:37:02 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm python3[13993]: ansible-file Invoked with path=/etc/containers/systemd state=directory owner=root group=0 mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Feb 05 04:37:02 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm python3[14055]: ansible-ansible.legacy.stat Invoked with path=/etc/containers/systemd/quadlet-demo.network follow=False get_checksum=True checksum_algorithm=sha1 get_mime=True get_attributes=True Feb 05 04:37:03 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm python3[14112]: ansible-ansible.legacy.copy Invoked with src=/root/.ansible/tmp/ansible-tmp-1707107822.6875324-8017-246495456961851/source dest=/etc/containers/systemd/quadlet-demo.network owner=root group=0 mode=0644 _original_basename=quadlet-demo.network follow=False checksum=e57c08d49aff4bae8daab138d913aeddaa8682a0 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Feb 05 04:37:03 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm python3[14153]: ansible-systemd Invoked with daemon_reload=True scope=system daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None Feb 05 04:37:03 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm systemd[1]: Reloading. Feb 05 04:37:03 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm systemd[1]: /usr/lib/systemd/system/cloud-init.target:15: Unknown lvalue 'ConditionEnvironment' in section 'Unit' Feb 05 04:37:03 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm systemd[1]: /usr/lib/systemd/system/cloud-init.service:18: Unknown lvalue 'ConditionEnvironment' in section 'Unit' Feb 05 04:37:03 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm systemd[1]: /usr/lib/systemd/system/cloud-init-local.service:12: Unknown lvalue 'ConditionEnvironment' in section 'Unit' Feb 05 04:37:03 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm systemd[1]: /usr/lib/systemd/system/cloud-final.service:7: Unknown lvalue 'ConditionEnvironment' in section 'Unit' Feb 05 04:37:03 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm systemd[1]: /usr/lib/systemd/system/cloud-config.service:9: Unknown lvalue 'ConditionEnvironment' in section 'Unit' Feb 05 04:37:04 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm python3[14203]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None Feb 05 04:37:04 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm python3[14224]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Feb 05 04:37:05 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm python3[14246]: ansible-file Invoked with path=/etc/containers/systemd state=directory owner=root group=0 mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Feb 05 04:37:05 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm python3[14308]: ansible-ansible.legacy.stat Invoked with path=/etc/containers/systemd/quadlet-demo-mysql.volume follow=False get_checksum=True checksum_algorithm=sha1 get_mime=True get_attributes=True Feb 05 04:37:05 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm python3[14365]: ansible-ansible.legacy.copy Invoked with src=/root/.ansible/tmp/ansible-tmp-1707107825.1552148-8065-231360590828526/source dest=/etc/containers/systemd/quadlet-demo-mysql.volume owner=root group=0 mode=0644 _original_basename=quadlet-demo-mysql.volume follow=False checksum=585f8cbdf0ec73000f9227dcffbef71e9552ea4a backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Feb 05 04:37:05 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm python3[14406]: ansible-systemd Invoked with daemon_reload=True scope=system daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None Feb 05 04:37:05 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm systemd[1]: Reloading. Feb 05 04:37:06 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm systemd[1]: /usr/lib/systemd/system/cloud-init.target:15: Unknown lvalue 'ConditionEnvironment' in section 'Unit' Feb 05 04:37:06 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm systemd[1]: /usr/lib/systemd/system/cloud-init.service:18: Unknown lvalue 'ConditionEnvironment' in section 'Unit' Feb 05 04:37:06 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm systemd[1]: /usr/lib/systemd/system/cloud-init-local.service:12: Unknown lvalue 'ConditionEnvironment' in section 'Unit' Feb 05 04:37:06 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm systemd[1]: /usr/lib/systemd/system/cloud-final.service:7: Unknown lvalue 'ConditionEnvironment' in section 'Unit' Feb 05 04:37:06 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm systemd[1]: /usr/lib/systemd/system/cloud-config.service:9: Unknown lvalue 'ConditionEnvironment' in section 'Unit' Feb 05 04:37:06 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm python3[14456]: ansible-systemd Invoked with name=quadlet-demo-mysql-volume.service scope=system state=started daemon_reload=False daemon_reexec=False no_block=False enabled=None force=None masked=None Feb 05 04:37:06 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm systemd[1]: Starting quadlet-demo-mysql-volume.service... -- Subject: Unit quadlet-demo-mysql-volume.service has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit quadlet-demo-mysql-volume.service has begun starting up. Feb 05 04:37:06 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm systemd[1]: var-lib-containers-storage-overlay-metacopy\x2dcheck4123922290-merged.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay-metacopy\x2dcheck4123922290-merged.mount has successfully entered the 'dead' state. Feb 05 04:37:06 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm quadlet-demo-mysql-volume[14462]: systemd-quadlet-demo-mysql Feb 05 04:37:06 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm systemd[1]: Started quadlet-demo-mysql-volume.service. -- Subject: Unit quadlet-demo-mysql-volume.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit quadlet-demo-mysql-volume.service has finished starting up. -- -- The start-up result is done. Feb 05 04:37:07 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm python3[14491]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None Feb 05 04:37:07 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm python3[14512]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Feb 05 04:37:07 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Feb 05 04:37:07 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm python3[14534]: ansible-file Invoked with path=/tmp/quadlet_demo state=directory owner=root group=root mode=0777 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Feb 05 04:37:08 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm python3[14554]: ansible-containers.podman.podman_image Invoked with name=quay.io/linux-system-roles/mysql:5.6 force=True tag=latest pull=True push=False state=present executable=podman build={'force_rm': False, 'format': 'oci', 'cache': True, 'rm': True, 'annotation': None, 'file': None, 'volume': None, 'extra_args': None, 'target': None} push_args={'compress': None, 'format': None, 'remove_signatures': None, 'sign_by': None, 'dest': None, 'transport': None} arch=None path=None validate_certs=None auth_file=None username=None password=NOT_LOGGING_PARAMETER ca_cert_dir=None Feb 05 04:37:08 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm python3[14554]: ansible-containers.podman.podman_image PODMAN-IMAGE-DEBUG: /usr/bin/podman image ls quay.io/linux-system-roles/mysql:5.6 --format json Feb 05 04:37:08 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm python3[14554]: ansible-containers.podman.podman_image PODMAN-IMAGE-DEBUG: /usr/bin/podman image exists quay.io/linux-system-roles/mysql:5.6 Feb 05 04:37:08 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm python3[14554]: ansible-containers.podman.podman_image PODMAN-IMAGE-DEBUG: /usr/bin/podman pull quay.io/linux-system-roles/mysql:5.6 -q Feb 05 04:37:15 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Feb 05 04:37:15 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm python3[14554]: ansible-containers.podman.podman_image PODMAN-IMAGE-DEBUG: /usr/bin/podman inspect dd3b2a5dcb48ff61113592ed5ddd762581be4387c7bc552375a2159422aa6bf5 --format json Feb 05 04:37:15 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Feb 05 04:37:16 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm python3[14672]: ansible-file Invoked with path=/etc/containers/systemd state=directory owner=root group=0 mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Feb 05 04:37:16 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm python3[14734]: ansible-ansible.legacy.stat Invoked with path=/etc/containers/systemd/quadlet-demo-mysql.container follow=False get_checksum=True checksum_algorithm=sha1 get_mime=True get_attributes=True Feb 05 04:37:16 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm python3[14791]: ansible-ansible.legacy.copy Invoked with dest=/etc/containers/systemd/quadlet-demo-mysql.container owner=root group=0 mode=0644 src=/root/.ansible/tmp/ansible-tmp-1707107836.1947784-8117-161871808305394/source _original_basename=tmp_gj6hhx4 follow=False checksum=7bf9312d57274d4dabbd222b01717e2cc0cfa0ed backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Feb 05 04:37:16 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm python3[14832]: ansible-systemd Invoked with daemon_reload=True scope=system daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None Feb 05 04:37:16 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm systemd[1]: Reloading. Feb 05 04:37:17 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm systemd[1]: /usr/lib/systemd/system/cloud-init.target:15: Unknown lvalue 'ConditionEnvironment' in section 'Unit' Feb 05 04:37:17 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm systemd[1]: /usr/lib/systemd/system/cloud-init.service:18: Unknown lvalue 'ConditionEnvironment' in section 'Unit' Feb 05 04:37:17 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm systemd[1]: /usr/lib/systemd/system/cloud-init-local.service:12: Unknown lvalue 'ConditionEnvironment' in section 'Unit' Feb 05 04:37:17 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm systemd[1]: /usr/lib/systemd/system/cloud-final.service:7: Unknown lvalue 'ConditionEnvironment' in section 'Unit' Feb 05 04:37:17 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm systemd[1]: /usr/lib/systemd/system/cloud-config.service:9: Unknown lvalue 'ConditionEnvironment' in section 'Unit' Feb 05 04:37:17 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm python3[14882]: ansible-systemd Invoked with name=quadlet-demo-mysql.service scope=system state=started daemon_reload=False daemon_reexec=False no_block=False enabled=None force=None masked=None Feb 05 04:37:17 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm systemd[1]: Starting quadlet-demo-network.service... -- Subject: Unit quadlet-demo-network.service has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit quadlet-demo-network.service has begun starting up. Feb 05 04:37:17 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm quadlet-demo-network[14888]: systemd-quadlet-demo Feb 05 04:37:17 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm systemd[1]: Started quadlet-demo-network.service. -- Subject: Unit quadlet-demo-network.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit quadlet-demo-network.service has finished starting up. -- -- The start-up result is done. Feb 05 04:37:17 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm systemd[1]: Starting quadlet-demo-mysql.service... -- Subject: Unit quadlet-demo-mysql.service has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit quadlet-demo-mysql.service has begun starting up. Feb 05 04:37:17 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm quadlet-demo-mysql[14896]: Error: running container create option: no secret with name or id "mysql-root-password-container": no such secret Feb 05 04:37:17 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Feb 05 04:37:17 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm systemd[1]: quadlet-demo-mysql.service: Main process exited, code=exited, status=125/n/a Feb 05 04:37:17 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm systemd[1]: quadlet-demo-mysql.service: Failed with result 'exit-code'. -- Subject: Unit failed -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit quadlet-demo-mysql.service has entered the 'failed' state with result 'exit-code'. Feb 05 04:37:17 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm systemd[1]: Failed to start quadlet-demo-mysql.service. -- Subject: Unit quadlet-demo-mysql.service has failed -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit quadlet-demo-mysql.service has failed. -- -- The result is failed. Feb 05 04:37:17 8b813d25-3f6c-402c-86ff-762b252376c0.testing-farm python3[14958]: ansible-ansible.legacy.command Invoked with _raw_params=journalctl -ex _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None TASK [Check] ******************************************************************* task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/tests/tests_quadlet_demo.yml:136 Monday 05 February 2024 04:37:17 +0000 (0:00:00.252) 0:01:40.041 ******* ok: [sut] => { "changed": false, "cmd": [ "podman", "ps", "-a" ], "delta": "0:00:00.039721", "end": "2024-02-05 04:37:18.152232", "rc": 0, "start": "2024-02-05 04:37:18.112511" } STDOUT: CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES TASK [Check pods] ************************************************************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/tests/tests_quadlet_demo.yml:140 Monday 05 February 2024 04:37:18 +0000 (0:00:00.234) 0:01:40.275 ******* ok: [sut] => { "changed": false, "cmd": [ "podman", "pod", "ps", "--ctr-ids", "--ctr-names", "--ctr-status" ], "delta": "0:00:00.038138", "end": "2024-02-05 04:37:18.385150", "failed_when_result": false, "rc": 0, "start": "2024-02-05 04:37:18.347012" } STDOUT: POD ID NAME STATUS CREATED INFRA ID IDS NAMES STATUS TASK [Check systemd] *********************************************************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/tests/tests_quadlet_demo.yml:145 Monday 05 February 2024 04:37:18 +0000 (0:00:00.233) 0:01:40.509 ******* ok: [sut] => { "changed": false, "cmd": "set -euo pipefail; systemctl list-units --all --with-dependencies | grep quadlet", "delta": "0:00:00.008829", "end": "2024-02-05 04:37:18.590294", "failed_when_result": false, "rc": 1, "start": "2024-02-05 04:37:18.581465" } STDERR: systemctl: unrecognized option '--with-dependencies' MSG: non-zero return code TASK [LS] ********************************************************************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/tests/tests_quadlet_demo.yml:153 Monday 05 February 2024 04:37:18 +0000 (0:00:00.202) 0:01:40.711 ******* ok: [sut] => { "changed": false, "cmd": [ "ls", "-alrtF", "/etc/systemd/system" ], "delta": "0:00:00.004010", "end": "2024-02-05 04:37:18.786424", "failed_when_result": false, "rc": 0, "start": "2024-02-05 04:37:18.782414" } STDOUT: total 8 lrwxrwxrwx. 1 root root 9 May 11 2019 systemd-timedated.service -> /dev/null drwxr-xr-x. 4 root root 169 Jan 30 09:30 ../ lrwxrwxrwx. 1 root root 39 Jan 30 09:30 syslog.service -> /usr/lib/systemd/system/rsyslog.service drwxr-xr-x. 2 root root 32 Jan 30 09:30 getty.target.wants/ lrwxrwxrwx. 1 root root 37 Jan 30 09:30 ctrl-alt-del.target -> /usr/lib/systemd/system/reboot.target drwxr-xr-x. 2 root root 48 Jan 30 09:30 network-online.target.wants/ lrwxrwxrwx. 1 root root 57 Jan 30 09:30 dbus-org.freedesktop.nm-dispatcher.service -> /usr/lib/systemd/system/NetworkManager-dispatcher.service lrwxrwxrwx. 1 root root 41 Jan 30 09:31 dbus-org.freedesktop.timedate1.service -> /usr/lib/systemd/system/timedatex.service drwxr-xr-x. 2 root root 61 Jan 30 09:31 timers.target.wants/ drwxr-xr-x. 2 root root 31 Jan 30 09:31 basic.target.wants/ drwxr-xr-x. 2 root root 38 Jan 30 09:31 dev-virtio\x2dports-org.qemu.guest_agent.0.device.wants/ lrwxrwxrwx. 1 root root 41 Jan 30 09:33 default.target -> /usr/lib/systemd/system/multi-user.target drwxr-xr-x. 2 root root 51 Jan 30 09:49 sockets.target.wants/ drwxr-xr-x. 2 root root 31 Jan 30 09:49 remote-fs.target.wants/ drwxr-xr-x. 2 root root 59 Jan 30 09:50 sshd-keygen@.service.d/ drwxr-xr-x. 2 root root 119 Jan 30 09:50 cloud-init.target.wants/ drwxr-xr-x. 2 root root 181 Jan 30 09:52 sysinit.target.wants/ drwxr-xr-x. 2 root root 4096 Feb 5 04:36 multi-user.target.wants/ lrwxrwxrwx. 1 root root 41 Feb 5 04:36 dbus-org.fedoraproject.FirewallD1.service -> /usr/lib/systemd/system/firewalld.service drwxr-xr-x. 13 root root 4096 Feb 5 04:36 ./ TASK [Cleanup] ***************************************************************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/tests/tests_quadlet_demo.yml:158 Monday 05 February 2024 04:37:18 +0000 (0:00:00.197) 0:01:40.908 ******* TASK [fedora.linux_system_roles.podman : Set platform/version specific variables] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:3 Monday 05 February 2024 04:37:18 +0000 (0:00:00.079) 0:01:40.988 ******* included: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml for sut TASK [fedora.linux_system_roles.podman : Ensure ansible_facts used by role] **** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:3 Monday 05 February 2024 04:37:18 +0000 (0:00:00.030) 0:01:41.019 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_required_facts | difference(ansible_facts.keys() | list) | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if system is ostree] ************ task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:11 Monday 05 February 2024 04:37:18 +0000 (0:00:00.024) 0:01:41.043 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_is_ostree is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set flag to indicate system is ostree] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:16 Monday 05 February 2024 04:37:18 +0000 (0:00:00.017) 0:01:41.061 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_is_ostree is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set platform/version specific variables] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:20 Monday 05 February 2024 04:37:18 +0000 (0:00:00.019) 0:01:41.081 ******* ok: [sut] => (item=RedHat.yml) => { "ansible_facts": { "__podman_packages": [ "podman", "shadow-utils-subid" ] }, "ansible_included_var_files": [ "/WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/vars/RedHat.yml" ], "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml" } skipping: [sut] => (item=CentOS.yml) => { "ansible_loop_var": "item", "changed": false, "false_condition": "__vars_file is file", "item": "CentOS.yml", "skip_reason": "Conditional result was False" } ok: [sut] => (item=CentOS_8.yml) => { "ansible_facts": { "__podman_packages": [ "crun", "podman", "podman-plugins", "shadow-utils-subid" ] }, "ansible_included_var_files": [ "/WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/vars/CentOS_8.yml" ], "ansible_loop_var": "item", "changed": false, "item": "CentOS_8.yml" } ok: [sut] => (item=CentOS_8.yml) => { "ansible_facts": { "__podman_packages": [ "crun", "podman", "podman-plugins", "shadow-utils-subid" ] }, "ansible_included_var_files": [ "/WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/vars/CentOS_8.yml" ], "ansible_loop_var": "item", "changed": false, "item": "CentOS_8.yml" } TASK [fedora.linux_system_roles.podman : Gather the package facts] ************* task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:6 Monday 05 February 2024 04:37:19 +0000 (0:00:00.050) 0:01:41.131 ******* ok: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Enable copr if requested] ************* task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:10 Monday 05 February 2024 04:37:20 +0000 (0:00:01.326) 0:01:42.458 ******* skipping: [sut] => { "changed": false, "false_condition": "podman_use_copr | d(false)", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Ensure required packages are installed] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:14 Monday 05 February 2024 04:37:20 +0000 (0:00:00.019) 0:01:42.477 ******* skipping: [sut] => { "changed": false, "false_condition": "(__podman_packages | difference(ansible_facts.packages))", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get podman version] ******************* task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:22 Monday 05 February 2024 04:37:20 +0000 (0:00:00.022) 0:01:42.499 ******* ok: [sut] => { "changed": false, "cmd": [ "podman", "--version" ], "delta": "0:00:00.029113", "end": "2024-02-05 04:37:20.601857", "rc": 0, "start": "2024-02-05 04:37:20.572744" } STDOUT: podman version 4.8.3 TASK [fedora.linux_system_roles.podman : Set podman version] ******************* task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:27 Monday 05 February 2024 04:37:20 +0000 (0:00:00.226) 0:01:42.725 ******* ok: [sut] => { "ansible_facts": { "podman_version": "4.8.3" }, "changed": false } TASK [fedora.linux_system_roles.podman : Podman package version must be 4.2 or later] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:31 Monday 05 February 2024 04:37:20 +0000 (0:00:00.020) 0:01:42.746 ******* skipping: [sut] => { "changed": false, "false_condition": "podman_version is version(\"4.2\", \"<\")", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Podman package version must be 4.4 or later for quadlet, secrets] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:38 Monday 05 February 2024 04:37:20 +0000 (0:00:00.017) 0:01:42.764 ******* skipping: [sut] => { "changed": false, "false_condition": "podman_version is version(\"4.4\", \"<\")", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Podman package version must be 4.4 or later for quadlet, secrets] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:48 Monday 05 February 2024 04:37:20 +0000 (0:00:00.019) 0:01:42.783 ******* META: end_host conditional evaluated to False, continuing execution for sut skipping: [sut] => { "skip_reason": "end_host conditional evaluated to False, continuing execution for sut" } MSG: end_host conditional evaluated to false, continuing execution for sut TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:55 Monday 05 February 2024 04:37:20 +0000 (0:00:00.050) 0:01:42.834 ******* included: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for sut TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Monday 05 February 2024 04:37:20 +0000 (0:00:00.036) 0:01:42.871 ******* skipping: [sut] => { "changed": false, "false_condition": "'getent_passwd' not in ansible_facts or __podman_user not in ansible_facts['getent_passwd']", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Monday 05 February 2024 04:37:20 +0000 (0:00:00.022) 0:01:42.893 ******* skipping: [sut] => { "changed": false, "false_condition": "not ansible_facts[\"getent_passwd\"][__podman_user]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Monday 05 February 2024 04:37:20 +0000 (0:00:00.022) 0:01:42.915 ******* ok: [sut] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get group information] **************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:27 Monday 05 February 2024 04:37:20 +0000 (0:00:00.027) 0:01:42.943 ******* ok: [sut] => { "ansible_facts": { "getent_group": { "root": [ "x", "0", "" ] } }, "changed": false } TASK [fedora.linux_system_roles.podman : Set group name] *********************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:34 Monday 05 February 2024 04:37:21 +0000 (0:00:00.222) 0:01:43.166 ******* ok: [sut] => { "ansible_facts": { "__podman_group_name": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:38 Monday 05 February 2024 04:37:21 +0000 (0:00:00.026) 0:01:43.192 ******* ok: [sut] => { "changed": false, "stat": { "atime": 1707107813.7932265, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "bb5b46ffbafcaa8c4021f3c8b3cb8594f48ef34b", "ctime": 1707107779.304328, "dev": 51713, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 6792906, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-sharedlib", "mode": "0755", "mtime": 1700557386.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 12640, "uid": 0, "version": "3297496774", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************ task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:49 Monday 05 February 2024 04:37:21 +0000 (0:00:00.210) 0:01:43.402 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check group with getsubids] *********** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:53 Monday 05 February 2024 04:37:21 +0000 (0:00:00.018) 0:01:43.420 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if user is in subuid file] ****** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:63 Monday 05 February 2024 04:37:21 +0000 (0:00:00.018) 0:01:43.438 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" }itNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "262144", "LimitNOFILESoft": "1024", "LimitNPROC": "14003", "LimitNPROCSoft": "14003", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "14003", "LimitSIGPENDINGSoft": "14003", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "13727", "MemoryAccounting": "yes", "MemoryCurrent": "40820736", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemorySwapMax": "infinity", "MountAPIVFS": "no", "MountFlags": "", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAMask": "", "NUMAPolicy": "n/a", "Names": "firewalld.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "PermissionsStartOnly": "no", "Perpetual": "no", "PrivateDevices": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "RemainAfterExit": "no", "RemoveIPC": "no", "Requires": "dbus.socket sysinit.target system.slice", "Restart": "no", "RestartUSec": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "Slice": "system.slice", "StandardError": "null", "StandardInput": "null", "StandardInputData": "", "StandardOutput": "null", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StateChangeTimestamp": "Mon 2024-02-05 04:36:57 UTC", "StateChangeTimestampMonotonic": "844023745", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "running", "SuccessAction": "none", "SyslogFacility": "3", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "0", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "2", "TasksMax": "22405", "TimeoutStartUSec": "1min 30s", "TimeoutStopUSec": "1min 30s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "dbus", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "enabled", "UnitFileState": "enabled", "UtmpMode": "init", "WantedBy": "multi-user.target", "Wants": "network-pre.target", "WatchdogTimestamp": "Mon 2024-02-05 04:36:57 UTC", "WatchdogTimestampMonotonic": "844023742", "WatchdogUSec": "0" } } TASK [fedora.linux_system_roles.firewall : Enable and start firewalld service] *** task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:28 Monday 05 February 2024 04:37:24 +0000 (0:00:00.335) 0:01:46.520 ******* ok: [sut] => { "changed": false, "enabled": true, "name": "firewalld", "state": "started", "status": { "ActiveEnterTimestamp": "Mon 2024-02-05 04:36:57 UTC", "ActiveEnterTimestampMonotonic": "844023745", "ActiveExitTimestampMonotonic": "0", "ActiveState": "active", "After": "basic.target dbus.socket dbus.service system.slice polkit.service sysinit.target", "AllowIsolate": "no", "AllowedCPUs": "", "AllowedMemoryNodes": "", "AmbientCapabilities": "", "AssertResult": "yes", "AssertTimestamp": "Mon 2024-02-05 04:36:57 UTC", "AssertTimestampMonotonic": "843333266", "Before": "multi-user.target shutdown.target network-pre.target", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "BusName": "org.fedoraproject.FirewallD1", "CPUAccounting": "no", "CPUAffinity": "", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "[not set]", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "yes", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf", "CollectMode": "inactive", "ConditionResult": "yes", "ConditionTimestamp": "Mon 2024-02-05 04:36:57 UTC", "ConditionTimestampMonotonic": "843333264", "ConfigurationDirectoryMode": "0755", "Conflicts": "ip6tables.service iptables.service ipset.service ebtables.service nftables.service shutdown.target", "ControlGroup": "/system.slice/firewalld.service", "ControlPID": "0", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "Delegate": "no", "Description": "firewalld - dynamic firewall daemon", "DevicePolicy": "auto", "Documentation": "man:firewalld(1)", "DynamicUser": "no", "EffectiveCPUs": "", "EffectiveMemoryNodes": "", "EnvironmentFiles": "/etc/sysconfig/firewalld (ignore_errors=yes)", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "13727", "ExecMainStartTimestamp": "Mon 2024-02-05 04:36:57 UTC", "ExecMainStartTimestampMonotonic": "843334789", "ExecMainStatus": "0", "ExecReload": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStart": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FragmentPath": "/usr/lib/systemd/system/firewalld.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOSchedulingClass": "0", "IOSchedulingPriority": "0", "IOWeight": "[not set]", "IPAccounting": "no", "IPEgressBytes": "18446744073709551615", "IPEgressPackets": "18446744073709551615", "IPIngressBytes": "18446744073709551615", "IPIngressPackets": "18446744073709551615", "Id": "firewalld.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestamp": "Mon 2024-02-05 04:36:57 UTC", "InactiveExitTimestampMonotonic": "843334853", "InvocationID": "adeaf38f5dab4ccbb553c48d522b2165", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "mixed", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "0", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "65536", "LimitMEMLOCKSoft": "65536", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "262144", "LimitNOFILESoft": "1024", "LimitNPROC": "14003", "LimitNPROCSoft": "14003", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "14003", "LimitSIGPENDINGSoft": "14003", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "13727", "MemoryAccounting": "yes", "MemoryCurrent": "40820736", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemorySwapMax": "infinity", "MountAPIVFS": "no", "MountFlags": "", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAMask": "", "NUMAPolicy": "n/a", "Names": "firewalld.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "PermissionsStartOnly": "no", "Perpetual": "no", "PrivateDevices": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "RemainAfterExit": "no", "RemoveIPC": "no", "Requires": "dbus.socket sysinit.target system.slice", "Restart": "no", "RestartUSec": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "Slice": "system.slice", "StandardError": "null", "StandardInput": "null", "StandardInputData": "", "StandardOutput": "null", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StateChangeTimestamp": "Mon 2024-02-05 04:36:57 UTC", "StateChangeTimestampMonotonic": "844023745", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "running", "SuccessAction": "none", "SyslogFacility": "3", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "0", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "2", "TasksMax": "22405", "TimeoutStartUSec": "1min 30s", "TimeoutStopUSec": "1min 30s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "dbus", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "enabled", "UnitFileState": "enabled", "UtmpMode": "init", "WantedBy": "multi-user.target", "Wants": "network-pre.target", "WatchdogTimestamp": "Mon 2024-02-05 04:36:57 UTC", "WatchdogTimestampMonotonic": "844023742", "WatchdogUSec": "0" } } TASK [fedora.linux_system_roles.firewall : Check if previous replaced is defined] *** task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:34 Monday 05 February 2024 04:37:25 +0000 (0:00:00.350) 0:01:46.871 ******* ok: [sut] => { "ansible_facts": { "__firewall_previous_replaced": false, "__firewall_python_cmd": "/usr/bin/python3", "__firewall_report_changed": true }, "changed": false } TASK [fedora.linux_system_roles.firewall : Get config files, checksums before and remove] *** task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:43 Monday 05 February 2024 04:37:25 +0000 (0:00:00.023) 0:01:46.894 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Tell firewall module it is able to report changed] *** task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:55 Monday 05 February 2024 04:37:25 +0000 (0:00:00.020) 0:01:46.914 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Configure firewall] ***************** task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:71 Monday 05 February 2024 04:37:25 +0000 (0:00:00.020) 0:01:46.934 ******* ok: [sut] => (item={'port': '8000/tcp', 'state': 'enabled'}) => { "__firewall_changed": false, "ansible_loop_var": "item", "changed": false, "item": { "port": "8000/tcp", "state": "enabled" } } ok: [sut] => (item={'port': '9000/tcp', 'state': 'enabled'}) => { "__firewall_changed": false, "ansible_loop_var": "item", "changed": false, "item": { "port": "9000/tcp", "state": "enabled" } } TASK [fedora.linux_system_roles.firewall : Gather firewall config information] *** task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:120 Monday 05 February 2024 04:37:26 +0000 (0:00:00.856) 0:01:47.790 ******* skipping: [sut] => (item={'port': '8000/tcp', 'state': 'enabled'}) => { "ansible_loop_var": "item", "changed": false, "item": { "port": "8000/tcp", "state": "enabled" }, "skip_reason": "Conditional result was False" } skipping: [sut] => (item={'port': '9000/tcp', 'state': 'enabled'}) => { "ansible_loop_var": "item", "changed": false, "item": { "port": "9000/tcp", "state": "enabled" }, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Update firewalld_config fact] ******* task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:130 Monday 05 February 2024 04:37:26 +0000 (0:00:00.032) 0:01:47.822 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Gather firewall config if no arguments] *** task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:139 Monday 05 February 2024 04:37:26 +0000 (0:00:00.024) 0:01:47.847 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Update firewalld_config fact] ******* task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:144 Monday 05 February 2024 04:37:26 +0000 (0:00:00.020) 0:01:47.867 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Get config files, checksums after] *** task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:153 Monday 05 February 2024 04:37:26 +0000 (0:00:00.021) 0:01:47.889 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Calculate what has changed] ********* task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:163 Monday 05 February 2024 04:37:26 +0000 (0:00:00.018) 0:01:47.908 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Show diffs] ************************* task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:169 Monday 05 February 2024 04:37:26 +0000 (0:00:00.019) 0:01:47.928 ******* skipping: [sut] => {} TASK [Manage selinux for specified ports] ************************************** task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/main.yml:102 Monday 05 February 2024 04:37:26 +0000 (0:00:00.020) 0:01:47.948 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Keep track of users that need to cancel linger] *** task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/main.yml:109 Monday 05 February 2024 04:37:26 +0000 (0:00:00.019) 0:01:47.968 ******* ok: [sut] => { "ansible_facts": { "__podman_cancel_user_linger": [] }, "changed": false } TASK [linux-system-roles.podman : Handle secrets] ****************************** task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/main.yml:113 Monday 05 February 2024 04:37:26 +0000 (0:00:00.045) 0:01:48.013 ******* included: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/handle_secret.yml for sut included: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/handle_secret.yml for sut included: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/handle_secret.yml for sut TASK [linux-system-roles.podman : Set variables part 0] ************************ task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/handle_secret.yml:3 Monday 05 February 2024 04:37:26 +0000 (0:00:00.053) 0:01:48.067 ******* ok: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [linux-system-roles.podman : Set variables part 1] ************************ task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/handle_secret.yml:12 Monday 05 February 2024 04:37:26 +0000 (0:00:00.022) 0:01:48.089 ******* ok: [sut] => { "ansible_facts": { "__podman_user": "root" }, "changed": false } TASK [linux-system-roles.podman : Set variables part 2] ************************ task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/handle_secret.yml:17 Monday 05 February 2024 04:37:26 +0000 (0:00:00.021) 0:01:48.110 ******* ok: [sut] => { "ansible_facts": { "__podman_rootless": false, "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [linux-system-roles.podman : Manage linger] ******************************* task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/handle_secret.yml:23 Monday 05 February 2024 04:37:26 +0000 (0:00:00.025) 0:01:48.136 ******* included: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/manage_linger.yml for sut TASK [linux-system-roles.podman : Enable linger if needed] ********************* task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/manage_linger.yml:12 Monday 05 February 2024 04:37:26 +0000 (0:00:00.031) 0:01:48.167 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Mark user as not yet needing to cancel linger] *** task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/manage_linger.yml:18 Monday 05 February 2024 04:37:26 +0000 (0:00:00.019) 0:01:48.187 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Mark user for possible linger cancel] ******** task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/manage_linger.yml:22 Monday 05 February 2024 04:37:26 +0000 (0:00:00.019) 0:01:48.206 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Manage each secret] ************************** task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/handle_secret.yml:28 Monday 05 February 2024 04:37:26 +0000 (0:00:00.018) 0:01:48.225 ******* ok: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [linux-system-roles.podman : Set variables part 0] ************************ task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/handle_secret.yml:3 Monday 05 February 2024 04:37:26 +0000 (0:00:00.249) 0:01:48.474 ******* ok: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [linux-system-roles.podman : Set variables part 1] ************************ task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/handle_secret.yml:12 Monday 05 February 2024 04:37:26 +0000 (0:00:00.023) 0:01:48.498 ******* ok: [sut] => { "ansible_facts": { "__podman_user": "root" }, "changed": false } TASK [linux-system-roles.podman : Set variables part 2] ************************ task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/handle_secret.yml:17 Monday 05 February 2024 04:37:26 +0000 (0:00:00.020) 0:01:48.519 ******* ok: [sut] => { "ansible_facts": { "__podman_rootless": false, "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [linux-system-roles.podman : Manage linger] ******************************* task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/handle_secret.yml:23 Monday 05 February 2024 04:37:26 +0000 (0:00:00.027) 0:01:48.547 ******* included: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/manage_linger.yml for sut TASK [linux-system-roles.podman : Enable linger if needed] ********************* task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/manage_linger.yml:12 Monday 05 February 2024 04:37:26 +0000 (0:00:00.030) 0:01:48.578 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Mark user as not yet needing to cancel linger] *** task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/manage_linger.yml:18 Monday 05 February 2024 04:37:26 +0000 (0:00:00.044) 0:01:48.622 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Mark user for possible linger cancel] ******** task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/manage_linger.yml:22 Monday 05 February 2024 04:37:26 +0000 (0:00:00.019) 0:01:48.642 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Manage each secret] ************************** task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/handle_secret.yml:28 Monday 05 February 2024 04:37:26 +0000 (0:00:00.020) 0:01:48.662 ******* ok: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [linux-system-roles.podman : Set variables part 0] ************************ task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/handle_secret.yml:3 Monday 05 February 2024 04:37:27 +0000 (0:00:00.248) 0:01:48.910 ******* ok: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [linux-system-roles.podman : Set variables part 1] ************************ task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/handle_secret.yml:12 Monday 05 February 2024 04:37:27 +0000 (0:00:00.022) 0:01:48.933 ******* ok: [sut] => { "ansible_facts": { "__podman_user": "root" }, "changed": false } TASK [linux-system-roles.podman : Set variables part 2] ************************ task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/handle_secret.yml:17 Monday 05 February 2024 04:37:27 +0000 (0:00:00.021) 0:01:48.954 ******* ok: [sut] => { "ansible_facts": { "__podman_rootless": false, "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [linux-system-roles.podman : Manage linger] ******************************* task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/handle_secret.yml:23 Monday 05 February 2024 04:37:27 +0000 (0:00:00.025) 0:01:48.980 ******* included: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/manage_linger.yml for sut TASK [linux-system-roles.podman : Enable linger if needed] ********************* task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/manage_linger.yml:12 Monday 05 February 2024 04:37:27 +0000 (0:00:00.030) 0:01:49.011 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Mark user as not yet needing to cancel linger] *** task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/manage_linger.yml:18 Monday 05 February 2024 04:37:27 +0000 (0:00:00.019) 0:01:49.030 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Mark user for possible linger cancel] ******** task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/manage_linger.yml:22 Monday 05 February 2024 04:37:27 +0000 (0:00:00.019) 0:01:49.050 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Manage each secret] ************************** task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/handle_secret.yml:28 Monday 05 February 2024 04:37:27 +0000 (0:00:00.019) 0:01:49.070 ******* ok: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [linux-system-roles.podman : Handle Kubernetes specifications] ************ task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/main.yml:120 Monday 05 February 2024 04:37:27 +0000 (0:00:00.250) 0:01:49.321 ******* TASK [linux-system-roles.podman : Handle Quadlet specifications] *************** task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/main.yml:126 Monday 05 February 2024 04:37:27 +0000 (0:00:00.018) 0:01:49.339 ******* included: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml for sut included: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml for sut included: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml for sut included: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml for sut included: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml for sut included: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml for sut TASK [linux-system-roles.podman : Set per-container variables part 0] ********** task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:14 Monday 05 February 2024 04:37:27 +0000 (0:00:00.076) 0:01:49.416 ******* ok: [sut] => { "ansible_facts": { "__podman_quadlet_file_src": "quadlet-demo-mysql.volume", "__podman_quadlet_spec": {}, "__podman_quadlet_str": "[Volume]", "__podman_quadlet_template_src": "" }, "changed": false } TASK [linux-system-roles.podman : Set per-container variables part 1] ********** task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:40 Monday 05 February 2024 04:37:27 +0000 (0:00:00.072) 0:01:49.488 ******* ok: [sut] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_pull_image": true, "__podman_state": "absent", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [linux-system-roles.podman : Fail if no quadlet spec is given] ************ task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/handle_quae, "__firewall_python_cmd": "/usr/bin/python3", "__firewall_report_changed": true }, "changed": false } TASK [fedora.linux_system_roles.firewall : Get config files, checksums before and remove] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:43 Monday 05 February 2024 04:37:25 +0000 (0:00:00.026) 0:01:47.183 ******* skipping: [sut] => { "changed": false, "false_condition": "__firewall_previous_replaced | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Tell firewall module it is able to report changed] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:55 Monday 05 February 2024 04:37:25 +0000 (0:00:00.017) 0:01:47.201 ******* skipping: [sut] => { "changed": false, "false_condition": "__firewall_previous_replaced | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Configure firewall] ***************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:71 Monday 05 February 2024 04:37:25 +0000 (0:00:00.018) 0:01:47.219 ******* ok: [sut] => (item={'port': '8000/tcp', 'state': 'enabled'}) => { "__firewall_changed": false, "ansible_loop_var": "item", "changed": false, "item": { "port": "8000/tcp", "state": "enabled" } } ok: [sut] => (item={'port': '9000/tcp', 'state': 'enabled'}) => { "__firewall_changed": false, "ansible_loop_var": "item", "changed": false, "item": { "port": "9000/tcp", "state": "enabled" } } TASK [fedora.linux_system_roles.firewall : Gather firewall config information] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:120 Monday 05 February 2024 04:37:25 +0000 (0:00:00.833) 0:01:48.052 ******* skipping: [sut] => (item={'port': '8000/tcp', 'state': 'enabled'}) => { "ansible_loop_var": "item", "changed": false, "false_condition": "firewall | length == 1", "item": { "port": "8000/tcp", "state": "enabled" }, "skip_reason": "Conditional result was False" } skipping: [sut] => (item={'port': '9000/tcp', 'state': 'enabled'}) => { "ansible_loop_var": "item", "changed": false, "false_condition": "firewall | length == 1", "item": { "port": "9000/tcp", "state": "enabled" }, "skip_reason": "Conditional result was False" } skipping: [sut] => { "changed": false } MSG: All items skipped TASK [fedora.linux_system_roles.firewall : Update firewalld_config fact] ******* task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:130 Monday 05 February 2024 04:37:25 +0000 (0:00:00.029) 0:01:48.082 ******* skipping: [sut] => { "changed": false, "false_condition": "firewall | length == 1", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Gather firewall config if no arguments] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:139 Monday 05 February 2024 04:37:26 +0000 (0:00:00.019) 0:01:48.102 ******* skipping: [sut] => { "changed": false, "false_condition": "firewall == None or firewall | length == 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Update firewalld_config fact] ******* task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:144 Monday 05 February 2024 04:37:26 +0000 (0:00:00.019) 0:01:48.121 ******* skipping: [sut] => { "changed": false, "false_condition": "firewall == None or firewall | length == 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Get config files, checksums after] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:153 Monday 05 February 2024 04:37:26 +0000 (0:00:00.018) 0:01:48.140 ******* skipping: [sut] => { "changed": false, "false_condition": "__firewall_previous_replaced | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Calculate what has changed] ********* task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:163 Monday 05 February 2024 04:37:26 +0000 (0:00:00.017) 0:01:48.157 ******* skipping: [sut] => { "changed": false, "false_condition": "__firewall_previous_replaced | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Show diffs] ************************* task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:169 Monday 05 February 2024 04:37:26 +0000 (0:00:00.018) 0:01:48.176 ******* skipping: [sut] => { "false_condition": "__firewall_previous_replaced | bool" } TASK [Manage selinux for specified ports] ************************************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:92 Monday 05 February 2024 04:37:26 +0000 (0:00:00.059) 0:01:48.236 ******* skipping: [sut] => { "changed": false, "false_condition": "podman_selinux_ports | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Keep track of users that need to cancel linger] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:99 Monday 05 February 2024 04:37:26 +0000 (0:00:00.019) 0:01:48.255 ******* ok: [sut] => { "ansible_facts": { "__podman_cancel_user_linger": [] }, "changed": false } TASK [fedora.linux_system_roles.podman : Handle secrets] *********************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:103 Monday 05 February 2024 04:37:26 +0000 (0:00:00.019) 0:01:48.274 ******* included: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml for sut => (item=(censored due to no_log)) included: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml for sut => (item=(censored due to no_log)) included: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml for sut => (item=(censored due to no_log)) TASK [fedora.linux_system_roles.podman : Set variables part 0] ***************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:3 Monday 05 February 2024 04:37:26 +0000 (0:00:00.052) 0:01:48.327 ******* ok: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Set variables part 1] ***************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:10 Monday 05 February 2024 04:37:26 +0000 (0:00:00.024) 0:01:48.351 ******* ok: [sut] => { "ansible_facts": { "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set variables part 2] ***************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:14 Monday 05 February 2024 04:37:26 +0000 (0:00:00.021) 0:01:48.372 ******* ok: [sut] => { "ansible_facts": { "__podman_rootless": false, "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:20 Monday 05 February 2024 04:37:26 +0000 (0:00:00.025) 0:01:48.398 ******* included: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for sut TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Monday 05 February 2024 04:37:26 +0000 (0:00:00.030) 0:01:48.428 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Monday 05 February 2024 04:37:26 +0000 (0:00:00.018) 0:01:48.447 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Monday 05 February 2024 04:37:26 +0000 (0:00:00.017) 0:01:48.464 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Manage each secret] ******************* task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:25 Monday 05 February 2024 04:37:26 +0000 (0:00:00.019) 0:01:48.484 ******* ok: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Set variables part 0] ***************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:3 Monday 05 February 2024 04:37:26 +0000 (0:00:00.246) 0:01:48.731 ******* ok: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Set variables part 1] ***************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:10 Monday 05 February 2024 04:37:26 +0000 (0:00:00.022) 0:01:48.753 ******* ok: [sut] => { "ansible_facts": { "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set variables part 2] ***************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:14 Monday 05 February 2024 04:37:26 +0000 (0:00:00.020) 0:01:48.774 ******* ok: [sut] => { "ansible_facts": { "__podman_rootless": false, "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:20 Monday 05 February 2024 04:37:26 +0000 (0:00:00.023) 0:01:48.798 ******* included: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for sut TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Monday 05 February 2024 04:37:26 +0000 (0:00:00.028) 0:01:48.826 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Monday 05 February 2024 04:37:26 +0000 (0:00:00.050) 0:01:48.876 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Monday 05 February 2024 04:37:26 +0000 (0:00:00.019) 0:01:48.895 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Manage each secret] ******************* task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:25 Monday 05 February 2024 04:37:26 +0000 (0:00:00.018) 0:01:48.914 ******* ok: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Set variables part 0] ***************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:3 Monday 05 February 2024 04:37:27 +0000 (0:00:00.246) 0:01:49.160 ******* ok: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Set variables part 1] ***************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:10 Monday 05 February 2024 04:37:27 +0000 (0:00:00.022) 0:01:49.183 ******* ok: [sut] => { "ansible_facts": { "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set variables part 2] ***************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:14 Monday 05 February 2024 04:37:27 +0000 (0:00:00.020) 0:01:49.203 ******* ok: [sut] => { "ansible_facts": { "__podman_rootless": false, "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:20 Monday 05 February 2024 04:37:27 +0000 (0:00:00.024) 0:01:49.228 ******* included: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for sut TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Monday 05 February 2024 04:37:27 +0000 (0:00:00.028) 0:01:49.257 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Monday 05 February 2024 04:37:27 +0000 (0:00:00.018) 0:01:49.275 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Monday 05 February 2024 04:37:27 +0000 (0:00:00.017) 0:01:49.292 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Manage each secret] ******************* task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:25 Monday 05 February 2024 04:37:27 +0000 (0:00:00.018) 0:01:49.311 ******* ok: [sut] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Handle Kubernetes specifications] ***** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:110 Monday 05 February 2024 04:37:27 +0000 (0:00:00.244) 0:01:49.556 ******* skipping: [sut] => { "changed": false, "skipped_reason": "No items in the list" } TASK [fedora.linux_system_roles.podman : Handle Quadlet specifications] ******** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:116 Monday 05 February 2024 04:37:27 +0000 (0:00:00.017) 0:01:49.573 ******* included: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for sut => (item={'file_src': 'quadlet-demo-mysql.volume', 'state': 'absent'}) included: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for sut => (item={'template_src': 'quadlet-demo-mysql.container.j2', 'state': 'absent'}) included: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for sut => (item={'file_src': 'envoy-proxy-configmap.yml', 'state': 'absent'}) included: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for sut => (item={'template_src': 'quadlet-demo.yml.j2', 'state': 'absent'}) included: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for sut => (item={'file_src': 'quadlet-demo.kube', 'state': 'absent'}) included: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for sut => (item={'file_src': 'quadlet-demo.network', 'state': 'absent'}) TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14 Monday 05 February 2024 04:37:27 +0000 (0:00:00.081) 0:01:49.654 ******* ok: [sut] => { "ansible_facts": { "__podman_quadlet_file_src": "quadlet-demo-mysql.volume", "__podman_quadlet_spec": {}, "__podman_quadlet_str": "[Volume]", "__podman_quadlet_template_src": "" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25 Monday 05 February 2024 04:37:27 +0000 (0:00:00.027) 0:01:49.681 ******* ok: [sut] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_pull_image": true, "__podman_state": "absent", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] ***** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35 Monday 05 February 2024 04:37:27 +0000 (0:00:00.023) 0:01:49.705 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_quadlet_file_src", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48 Monday 05 February 2024 04:37:27 +0000 (0:00:00.018) 0:01:49.724 ******* ok: [sut] => { "ansible_facts": { "__podman_quadlet_name": "quadlet-demo-mysql", "__podman_quadlet_type": "volume", "__podman_rootless": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57 Monday 05 February 2024 04:37:27 +0000 (0:00:00.062) 0:01:49.786 ******* included: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for sut TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Monday 05 February 2024 04:37:27 +0000 (0:00:00.035) 0:01:49.822 ******* skipping: [sut] => { "changed": false, "false_condition": "'getent_passwd' not in ansible_facts or __podman_user not in ansible_facts['getent_passwd']", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Monday 05 February 2024 04:37:27 +0000 (0:00:00.023) 0:01:49.846 ******* skipping: [sut] => { "changed": false, "false_condition": "not ansible_facts[\"getent_passwd\"][__podman_user]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Monday 05 February 2024 04:37:27 +0000 (0:00:00.023) 0:01:49.869 ******* ok: [sut] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get group information] **************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:27 Monday 05 February 2024 04:37:27 +0000 (0:00:00.027) 0:01:49.897 ******* ok: [sut] => { "ansible_facts": { "getent_group": { "root": [ "x", "0", "" ] } }, "changed": false } TASK [fedora.linux_system_roles.podman : Set group name] *********************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:34 Monday 05 February 2024 04:37:28 +0000 (0:00:00.204) 0:01:50.101 ******* ok: [sut] => { "ansible_facts": { "__podman_group_name": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:38 Monday 05 February 2024 04:37:28 +0000 (0:00:00.026) 0:01:50.128 ******* ok: [sut] => { "changed": false, "stat": { "atime": 1707107813.7932265, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "bb5b46ffbafcaa8c4021f3c8b3cb8594f48ef34b", "ctime": 1707107779.304328, "dev": 51713, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 6792906, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-sharedlib", "mode": "0755", "mtime": 1700557386.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 12640, "uid": 0, "version": "3297496774", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************ task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:49 Monday 05 February 2024 04:37:28 +0000 (0:00:00.201) 0:01:50.330 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check group with getsubids] *********** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:53 Monday 05 February 2024 04:37:28 +0000 (0:00:00.018) 0:01:50.348 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if user is in subuid file] ****** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:63 Monday 05 February 2024 04:37:28 +0000 (0:00:00.017) 0:01:50.365 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:71 Monday 05 February 2024 04:37:28 +0000 (0:00:00.017) 0:01:50.383 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if group is in subgid file] ***** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:78 Monday 05 February 2024 04:37:28 +0000 (0:00:00.018) 0:01:50.401 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] ***** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:86 Monday 05 February 2024 04:37:28 +0000 (0:00:00.017) 0:01:50.419 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62 Monday 05 February 2024 04:37:28 +0000 (0:00:00.017) 0:01:50.436 ******* ok: [sut] => { "ansible_facts": { "__podman_activate_systemd_unit": true, "__podman_images_found": [], "__podman_kube_yamls_raw": "", "__podman_service_name": "quadlet-demo-mysql-volume.service", "__podman_systemd_scope": "system", "__podman_user_home_dir": "/root", "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73 Monday 05 February 2024 04:37:28 +0000 (0:00:00.032) 0:01:50.469 ******* ok: [sut] => { "ansible_facts": { "__podman_quadlet_path": "/etc/containers/systemd" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get kube yaml contents] *************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77 Monday 05 February 2024 04:37:28 +0000 (0:00:00.020) 0:01:50.489 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_state != \"absent\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87 Monday 05 February 2024 04:37:28 +0000 (0:00:00.016) 0:01:50.506 ******* ok: [sut] => { "ansible_facts": { "__podman_images": [], "__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo-mysql.volume", "__podman_volumes": [] }, "changed": false } TASK [fedora.linux_system_roles.podman : Cleanup quadlets] ********************* task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:103 Monday 05 February 2024 04:37:28 +0000 (0:00:00.046) 0:01:50.552 ******* included: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml for sut TASK [fedora.linux_system_roles.podman : Stop and disable service] ************* task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:4 Monday 05 February 2024 04:37:28 +0000 (0:00:00.073) 0:01:50.625 ******* changed: [sut] => { "changed": true, "enabled": false, "failed_when_result": false, "name": "quadlet-demo-mysql-volume.service", "state": "stopped", "status": { "ActiveEnterTimestamp": "Mon 2024-02-05 04:37:06 UTC", "ActiveEnterTimestampMonotonic": "843736262", "ActiveExitTimestampMonotonic": "0", "ActiveState": "active", "After": "sysinit.target system.slice -.mount basic.target systemd-journald.socket", "AllowIsolate": "no", "AllowedCPUs": "", "AllowedMemoryNodes": "", "AmbientCapabilities": "", "AssertResult": "yes", "AssertTimestamp": "Mon 2024-02-05 04:37:06 UTC", "AssertTimestampMonotonic": "843623458", "Before": "quadlet-demo-mysql.service shutdown.target", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "CPUAccounting": "no", "CPUAffinity": "", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "[not set]", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "no", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf", "CollectMode": "inactive", "ConditionResult": "yes", "ConditionTimestamp": "Mon 2024-02-05 04:37:06 UTC", "ConditionTimestampMonotonic": "843623457", "ConfigurationDirectoryMode": "0755", "Conflicts": "shutdown.target", "ControlGroup": "/system.slice/quadlet-demo-mysql-volume.service", "ControlPID": "0", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "Delegate": "no", "Description": "quadlet-demo-mysql-volume.service", "DevicePolicy": "auto", "DynamicUser": "no", "EffectiveCPUs": "", "EffectiveMemoryNodes": "", "ExecMainCode": "1", "ExecMainExitTimestamp": "Mon 2024-02-05 04:37:06 UTC", "ExecMainExitTimestampMonotonic": "843735903", "ExecMainPID": "14462", "ExecMainStartTimestamp": "Mon 2024-02-05 04:37:06 UTC", "ExecMainStartTimestampMonotonic": "843624362", "ExecMainStatus": "0", "ExecStart": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman volume create --ignore systemd-quadlet-demo-mysql ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FragmentPath": "/run/systemd/generator/quadlet-demo-mysql-volume.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOSchedulingClass": "0", "IOSchedulingPriority": "0", "IOWeight": "[not set]", "IPAccounting": "no", "IPEgressBytes": "18446744073709551615", "IPEgressPackets": "18446744073709551615", "IPIngressBytes": "18446744073709551615", "IPIngressPackets": "18446744073709551615", "Id": "quadlet-demo-mysql-volume.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestamp": "Mon 2024-02-05 04:37:06 UTC", "InactiveExitTimestampMonotonic": "843624426", "InvocationID": "074002164b1143a5a1fd2002d1a65e80", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "control-group", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "0", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "65536", "LimitMEMLOCKSoft": "65536", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "262144", "LimitNOFILESoft": "1024", "LimitNPROC": "14004", "LimitNPROCSoft": "14004", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "14004", "LimitSIGPENDINGSoft": "14004", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "0", "MemoryAccounting": "yes", "MemoryCurrent": "0", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemorySwapMax": "infinity", "MountAPIVFS": "no", "MountFlags": "", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAMask": "", "NUMAPolicy": "n/a", "Names": "quadlet-demo-mysql-volume.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "PermissionsStartOnly": "no", "Perpetual": "no", "PrivateDevices": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "RemainAfterExit": "yes", "RemoveIPC": "no", "RequiredBy": "quadlet-demo-mysql.service", "Requires": "system.slice -.mount sysinit.target", "RequiresMountsFor": "/run/containers", "Restart": "no", "RestartUSec": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "Slice": "system.slice", "StandardError": "inherit", "StandardInput": "null", "StandardInputData": "", "StandardOutput": "journal", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StateChangeTimestamp": "Mon 2024-02-05 04:37:06 UTC", "StateChangeTimestampMonotonic": "843736262", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "exited", "SuccessAction": "none", "SyslogFacility": "3", "SyslogIdentifier": "quadlet-demo-mysql-volume", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "0", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "0", "TasksMax": "22406", "TimeoutStartUSec": "infinity", "TimeoutStopUSec": "1min 30s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "oneshot", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "disabled", "UnitFileState": "generated", "UtmpMode": "init", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "0" } } TASK [fedora.linux_system_roles.podman : Remove quadlet file] ****************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:22 Monday 05 February 2024 04:37:29 +0000 (0:00:00.516) 0:01:51.142 ******* changed: [sut] => { "changed": true, "path": "/etc/containers/systemd/quadlet-demo-mysql.volume", "state": "absent" } TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:28 Monday 05 February 2024 04:37:29 +0000 (0:00:00.202) 0:01:51.345 ******* included: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for sut TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Monday 05 February 2024 04:37:29 +0000 (0:00:00.029) 0:01:51.375 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Monday 05 February 2024 04:37:29 +0000 (0:00:00.018) 0:01:51.394 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Monday 05 February 2024 04:37:29 +0000 (0:00:00.018) 0:01:51.412 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Reload systemctl] ********************* task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:36 Monday 05 February 2024 04:37:29 +0000 (0:00:00.017) 0:01:51.430 ******* ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.podman : Prune images no longer in use] ******** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:45 Monday 05 February 2024 04:37:29 +0000 (0:00:00.474) 0:01:51.904 ******* changed: [sut] => { "changed": true, "cmd": [ "podman", "image", "prune", "-f" ], "delta": "0:00:00.039286", "end": "2024-02-05 04:37:30.021327", "rc": 0, "start": "2024-02-05 04:37:29.982041" } TASK [fedora.linux_system_roles.podman : Remove network] *********************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:53 Monday 05 February 2024 04:37:30 +0000 (0:00:00.240) 0:01:52.145 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_quadlet_type == \"network\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create and update quadlets] *********** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:107 Monday 05 February 2024 04:37:30 +0000 (0:00:00.020) 0:01:52.165 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_state != \"absent\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14 Monday 05 February 2024 04:37:30 +0000 (0:00:00.017) 0:01:52.183 ******* ok: [sut] => { "ansible_facts": { "__podman_quadlet_file_src": "", "__podman_quadlet_spec": {}, "__podman_quadlet_str": "[Install]\nWantedBy=default.target\n\n[Container]\nImage=quay.io/linux-system-roles/mysql:5.6\nContainerName=quadlet-demo-mysql\nVolume=quadlet-demo-mysql.volume:/var/lib/mysql\nVolume=/tmp/quadlet_demo:/var/lib/quadlet_demo:Z\nNetwork=quadlet-demo.network\nSecret=mysql-root-password-container,type=env,target=MYSQL_ROOT_PASSWORD\nHealthCmd=/usr/bin/true\nHealthOnFailure=kill\n", "__podman_quadlet_template_src": "quadlet-demo-mysql.container.j2" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25 Monday 05 February 2024 04:37:30 +0000 (0:00:00.058) 0:01:52.242 ******* ok: [sut] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_pull_image": true, "__podman_state": "absent", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] ***** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35 Monday 05 February 2024 04:37:30 +0000 (0:00:00.024) 0:01:52.266 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_quadlet_str", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48 Monday 05 February 2024 04:37:30 +0000 (0:00:00.020) 0:01:52.286 ******* ok: [sut] => { "ansible_facts": { "__podman_quadlet_name": "quadlet-demo-mysql", "__podman_quadlet_type": "container", "__podman_rootless": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57 Monday 05 February 2024 04:37:30 +0000 (0:00:00.029) 0:01:52.315 ******* included: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for sut TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Monday 05 February 2024 04:37:30 +0000 (0:00:00.035) 0:01:52.351 ******* skipping: [sut] => { "changed": false, "false_condition": "'getent_passwd' not in ansible_facts or __podman_user not in ansible_facts['getent_passwd']", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Monday 05 February 2024 04:37:30 +0000 (0:00:00.021) 0:01:52.373 ******* skipping: [sut] => { "changed": false, "false_condition": "not ansible_facts[\"getent_passwd\"][__podman_user]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Monday 05 February 2024 04:37:30 +0000 (0:00:00.054) 0:01:52.427 ******* ok: [sut] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get group information] **************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:27 Monday 05 February 2024 04:37:30 +0000 (0:00:00.027) 0:01:52.455 ******* ok: [sut] => { "ansible_facts": { "getent_group": { "root": [ "x", "0", "" ] } }, "changed": false } TASK [fedora.linux_system_roles.podman : Set group name] *********************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:34 Monday 05 February 2024 04:37:30 +0000 (0:00:00.203) 0:01:52.658 ******* ok: [sut] => { "ansible_facts": { "__podman_group_name": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:38 Monday 05 February 2024 04:37:30 +0000 (0:00:00.025) 0:01:52.684 ******* ok: [sut] => { "changed": false, "stat": { "atime": 1707107813.7932265, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "bb5b46ffbafcaa8c4021f3c8b3cb8594f48ef34b", "ctime": 1707107779.304328, "dev": 51713, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 6792906, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-sharedlib", "mode": "0755", "mtime": 1700557386.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 12640, "uid": 0, "version": "3297496774", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************ task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:49 Monday 05 February 2024 04:37:30 +0000 (0:00:00.203) 0:01:52.888 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check group with getsubids] *********** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:53 Monday 05 February 2024 04:37:30 +0000 (0:00:00.018) 0:01:52.907 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if user is in subuid file] ****** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:63 Monday 05 February 2024 04:37:30 +0000 (0:00:00.017) 0:01:52.924 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:71 Monday 05 February 2024 04:37:30 +0000 (0:00:00.018) 0:01:52.942 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if group is in subgid file] ***** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:78 Monday 05 February 2024 04:37:30 +0000 (0:00:00.017) 0:01:52.960 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] ***** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:86 Monday 05 February 2024 04:37:30 +0000 (0:00:00.018) 0:01:52.978 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62 Monday 05 February 2024 04:37:30 +0000 (0:00:00.017) 0:01:52.996 ******* ok: [sut] => { "ansible_facts": { "__podman_activate_systemd_unit": true, "__podman_images_found": [ "quay.io/linux-system-roles/mysql:5.6" ], "__podman_kube_yamls_raw": "", "__podman_service_name": "quadlet-demo-mysql.service", "__podman_systemd_scope": "system", "__podman_user_home_dir": "/root", "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73 Monday 05 February 2024 04:37:30 +0000 (0:00:00.033) 0:01:53.029 ******* ok: [sut] => { "ansible_facts": { "__podman_quadlet_path": "/etc/containers/systemd" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get kube yaml contents] *************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77 Monday 05 February 2024 04:37:30 +0000 (0:00:00.019) 0:01:53.049 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_state != \"absent\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87 Monday 05 February 2024 04:37:30 +0000 (0:00:00.018) 0:01:53.068 ******* ok: [sut] => { "ansible_facts": { "__podman_images": [ "quay.io/linux-system-roles/mysql:5.6" ], "__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo-mysql.container", "__podman_volumes": [ "/tmp/quadlet_demo" ] }, "changed": false } TASK [fedora.linux_system_roles.podman : Cleanup quadlets] ********************* task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:103 Monday 05 February 2024 04:37:31 +0000 (0:00:00.047) 0:01:53.115 ******* included: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml for sut TASK [fedora.linux_system_roles.podman : Stop and disable service] ************* task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:4 Monday 05 February 2024 04:37:31 +0000 (0:00:00.038) 0:01:53.154 ******* changed: [sut] => { "changed": true, "enabled": false, "failed_when_result": false, "name": "quadlet-demo-mysql.service", "state": "stopped", "status": { "ActiveEnterTimestampMonotonic": "0", "ActiveExitTimestampMonotonic": "0", "ActiveState": "failed", "After": "-.mount basic.target system.slice systemd-journald.socket quadlet-demo-network.service tmp.mount quadlet-demo-mysql-volume.service sysinit.target", "AllowIsolate": "no", "AllowedCPUs": "", "AllowedMemoryNodes": "", "AmbientCapabilities": "", "AssertResult": "yes", "AssertTimestamp": "Mon 2024-02-05 04:37:17 UTC", "AssertTimestampMonotonic": "854709597", "Before": "shutdown.target", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "CPUAccounting": "no", "CPUAffinity": "", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "[not set]", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "no", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf", "CollectMode": "inactive", "ConditionResult": "yes", "ConditionTimestamp": "Mon 2024-02-05 04:37:17 UTC", "ConditionTimestampMonotonic": "854709595", "ConfigurationDirectoryMode": "0755", "Conflicts": "shutdown.target", "ControlPID": "0", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "Delegate": "yes", "DelegateControllers": "cpu cpuacct cpuset io blkio memory devices pids", "Description": "quadlet-demo-mysql.service", "DevicePolicy": "auto", "DynamicUser": "no", "EffectiveCPUs": "", "EffectiveMemoryNodes": "", "Environment": "PODMAN_SYSTEMD_UNIT=quadlet-demo-mysql.service", "ExecMainCode": "1", "ExecMainExitTimestamp": "Mon 2024-02-05 04:37:17 UTC", "ExecMainExitTimestampMonotonic": "854779831", "ExecMainPID": "14896", "ExecMainStartTimestamp": "Mon 2024-02-05 04:37:17 UTC", "ExecMainStartTimestampMonotonic": "854710808", "ExecMainStatus": "125", "ExecStart": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman run --name=quadlet-demo-mysql --cidfile=/run/quadlet-demo-mysql.cid --replace --rm --cgroups=split --network=systemd-quadlet-demo --sdnotify=conmon -d -v systemd-quadlet-demo-mysql:/var/lib/mysql -v /tmp/quadlet_demo:/var/lib/quadlet_demo:Z --secret mysql-root-password-container,type=env,target=MYSQL_ROOT_PASSWORD --health-cmd /usr/bin/true --health-on-failure kill quay.io/linux-system-roles/mysql:5.6 ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStop": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman rm -v -f -i --cidfile=/run/quadlet-demo-mysql.cid ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStopPost": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman rm -v -f -i --cidfile=/run/quadlet-demo-mysql.cid ; ignore_errors=yes ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FragmentPath": "/run/systemd/generator/quadlet-demo-mysql.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOSchedulingClass": "0", "IOSchedulingPriority": "0", "IOWeight": "[not set]", "IPAccounting": "no", "IPEgressBytes": "18446744073709551615", "IPEgressPackets": "18446744073709551615", "IPIngressBytes": "18446744073709551615", "IPIngressPackets": "18446744073709551615", "Id": "quadlet-demo-mysql.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestamp": "Mon 2024-02-05 04:37:17 UTC", "InactiveEnterTimestampMonotonic": "854823045", "InactiveExitTimestamp": "Mon 2024-02-05 04:37:17 UTC", "InactiveExitTimestampMonotonic": "854710850", "InvocationID": "478d3738f5694ce1a6a1ee467c432c73", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "mixed", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "0", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "65536", "LimitMEMLOCKSoft": "65536", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "262144", "LimitNOFILESoft": "1024", "LimitNPROC": "14004", "LimitNPROCSoft": "14004", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "14004", "LimitSIGPENDINGSoft": "14004", "LimitSTACK": "infinity", "LimitSTACKSoft": "8 volumeMounts:\n - name: config-volume\n mountPath: /etc/envoy\n - name: certificates\n mountPath: /etc/envoy-certificates\n env:\n - name: ENVOY_UID\n value: \"0\"\n resources:\n requests:\n memory: \"64Mi\"\n cpu: \"250m\"\n limits:\n memory: \"128Mi\"\n cpu: \"500m\"\n volumes:\n - name: config-volume\n configMap:\n name: envoy-proxy-config\n - name: certificates\n secret:\n secretName: envoy-certificates\n - name: wordpress-persistent-storage\n persistentVolumeClaim:\n claimName: wp-pv-claim\n - name: www # not used - for testing hostpath\n hostPath:\n path: /tmp/httpd3\n - name: create # not used - for testing hostpath\n hostPath:\n path: /tmp/httpd3-create\n", "__podman_quadlet_template_src": "quadlet-demo.yml.j2" }, "changed": false } TASK [linux-system-roles.podman : Set per-container variables part 1] ********** task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:40 Monday 05 February 2024 04:37:34 +0000 (0:00:00.060) 0:01:56.153 ******* ok: [sut] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_pull_image": true, "__podman_state": "absent", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [linux-system-roles.podman : Fail if no quadlet spec is given] ************ task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:57 Monday 05 February 2024 04:37:34 +0000 (0:00:00.024) 0:01:56.178 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Set per-container variables part 2] ********** task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:70 Monday 05 February 2024 04:37:34 +0000 (0:00:00.021) 0:01:56.200 ******* ok: [sut] => { "ansible_facts": { "__podman_quadlet_name": "quadlet-demo", "__podman_quadlet_type": "yml", "__podman_rootless": false }, "changed": false } TASK [linux-system-roles.podman : Check user and group information] ************ task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:97 Monday 05 February 2024 04:37:34 +0000 (0:00:00.028) 0:01:56.229 ******* included: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml for sut TASK [linux-system-roles.podman : Get user information] ************************ task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:2 Monday 05 February 2024 04:37:34 +0000 (0:00:00.034) 0:01:56.263 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Fail if user does not exist] ***************** task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:10 Monday 05 February 2024 04:37:34 +0000 (0:00:00.085) 0:01:56.348 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Set group for podman user] ******************* task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:17 Monday 05 February 2024 04:37:34 +0000 (0:00:00.025) 0:01:56.374 ******* ok: [sut] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [linux-system-roles.podman : Get group information] *********************** task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:28 Monday 05 February 2024 04:37:34 +0000 (0:00:00.027) 0:01:56.401 ******* ok: [sut] => { "ansible_facts": { "getent_group": { "root": [ "x", "0", "" ] } }, "changed": false } TASK [linux-system-roles.podman : Set group name] ****************************** task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:36 Monday 05 February 2024 04:37:34 +0000 (0:00:00.217) 0:01:56.618 ******* ok: [sut] => { "ansible_facts": { "__podman_group_name": "root" }, "changed": false } TASK [linux-system-roles.podman : See if getsubids exists] ********************* task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:41 Monday 05 February 2024 04:37:34 +0000 (0:00:00.026) 0:01:56.644 ******* ok: [sut] => { "changed": false, "stat": { "atime": 1707107812.8985565, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "bb5b46ffbafcaa8c4021f3c8b3cb8594f48ef34b", "ctime": 1707107780.3803086, "dev": 51713, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 6792906, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-sharedlib", "mode": "0755", "mtime": 1700557386.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 12640, "uid": 0, "version": "3213196921", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [linux-system-roles.podman : Check user with getsubids] ******************* task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:52 Monday 05 February 2024 04:37:35 +0000 (0:00:00.242) 0:01:56.887 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Check group with getsubids] ****************** task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:56 Monday 05 February 2024 04:37:35 +0000 (0:00:00.020) 0:01:56.907 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Check if user is in subuid file] ************* task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:66 Monday 05 February 2024 04:37:35 +0000 (0:00:00.019) 0:01:56.927 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Fail if user not in subuid file] ************* task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:74 Monday 05 February 2024 04:37:35 +0000 (0:00:00.020) 0:01:56.948 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Check if group is in subgid file] ************ task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:81 Monday 05 February 2024 04:37:35 +0000 (0:00:00.020) 0:01:56.968 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Fail if group not in subgid file] ************ task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:89 Monday 05 February 2024 04:37:35 +0000 (0:00:00.020) 0:01:56.988 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Set per-container variables part 3] ********** task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:102 Monday 05 February 2024 04:37:35 +0000 (0:00:00.020) 0:01:57.009 ******* ok: [sut] => { "ansible_facts": { "__podman_activate_systemd_unit": true, "__podman_images_found": [], "__podman_kube_yamls_raw": "", "__podman_service_name": "", "__podman_systemd_scope": "system", "__podman_user_home_dir": "/root", "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [linux-system-roles.podman : Set per-container variables part 4] ********** task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:134 Monday 05 February 2024 04:37:35 +0000 (0:00:00.035) 0:01:57.045 ******* ok: [sut] => { "ansible_facts": { "__podman_quadlet_path": "/etc/containers/systemd" }, "changed": false } TASK [linux-system-roles.podman : Get kube yaml contents] ********************** task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:140 Monday 05 February 2024 04:37:35 +0000 (0:00:00.021) 0:01:57.066 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Set per-container variables part 5] ********** task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:151 Monday 05 February 2024 04:37:35 +0000 (0:00:00.019) 0:01:57.086 ******* ok: [sut] => { "ansible_facts": { "__podman_images": [], "__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo.yml", "__podman_volumes": [] }, "changed": false } TASK [linux-system-roles.podman : Cleanup quadlets] **************************** task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:205 Monday 05 February 2024 04:37:35 +0000 (0:00:00.048) 0:01:57.135 ******* included: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/cleanup_quadlet_spec.yml for sut TASK [linux-system-roles.podman : Stop and disable service] ******************** task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/cleanup_quadlet_spec.yml:4 Monday 05 February 2024 04:37:35 +0000 (0:00:00.076) 0:01:57.211 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Remove quadlet file] ************************* task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/cleanup_quadlet_spec.yml:22 Monday 05 February 2024 04:37:35 +0000 (0:00:00.021) 0:01:57.233 ******* ok: [sut] => { "changed": false, "path": "/etc/containers/systemd/quadlet-demo.yml", "state": "absent" } TASK [linux-system-roles.podman : Manage linger] ******************************* task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/cleanup_quadlet_spec.yml:28 Monday 05 February 2024 04:37:35 +0000 (0:00:00.301) 0:01:57.534 ******* included: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/manage_linger.yml for sut TASK [linux-system-roles.podman : Enable linger if needed] ********************* task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/manage_linger.yml:12 Monday 05 February 2024 04:37:35 +0000 (0:00:00.033) 0:01:57.567 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Mark user as not yet needing to cancel linger] *** task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/manage_linger.yml:18 Monday 05 February 2024 04:37:35 +0000 (0:00:00.020) 0:01:57.587 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Mark user for possible linger cancel] ******** task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/manage_linger.yml:22 Monday 05 February 2024 04:37:35 +0000 (0:00:00.020) 0:01:57.607 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Reload systemctl] **************************** task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/cleanup_quadlet_spec.yml:36 Monday 05 February 2024 04:37:35 +0000 (0:00:00.019) 0:01:57.627 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Prune images no longer in use] *************** task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/cleanup_quadlet_spec.yml:45 Monday 05 February 2024 04:37:35 +0000 (0:00:00.021) 0:01:57.648 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Remove network] ****************************** task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/cleanup_quadlet_spec.yml:53 Monday 05 February 2024 04:37:35 +0000 (0:00:00.020) 0:01:57.669 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Create and update quadlets] ****************** task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:209 Monday 05 February 2024 04:37:35 +0000 (0:00:00.021) 0:01:57.690 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Set per-container variables part 0] ********** task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:14 Monday 05 February 2024 04:37:36 +0000 (0:00:00.018) 0:01:57.708 ******* ok: [sut] => { "ansible_facts": { "__podman_quadlet_file_src": "quadlet-demo.kube", "__podman_quadlet_spec": {}, "__podman_quadlet_str": "[Install]\nWantedBy=default.target\n\n[Unit]\nRequires=quadlet-demo-mysql.service\nAfter=quadlet-demo-mysql.service\n\n[Kube]\n# Point to the yaml file in the same directory\nYaml=quadlet-demo.yml\n# Use the quadlet-demo network\nNetwork=quadlet-demo.network\n# Publish the envoy proxy data port\nPublishPort=8000:8080\n# Publish the envoy proxy admin port\nPublishPort=9000:9901\n# Use the envoy proxy config map in the same directory\nConfigMap=envoy-proxy-configmap.yml", "__podman_quadlet_template_src": "" }, "changed": false } TASK [linux-system-roles.podman : Set per-container variables part 1] ********** task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:40 Monday 05 February 2024 04:37:36 +0000 (0:00:00.029) 0:01:57.737 ******* ok: [sut] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_pull_image": true, "__podman_state": "absent", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [linux-system-roles.podman : Fail if no quadlet spec is given] ************ task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:57 Monday 05 February 2024 04:37:36 +0000 (0:00:00.025) 0:01:57.763 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Set per-container variables part 2] ********** task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:70 Monday 05 February 2024 04:37:36 +0000 (0:00:00.020) 0:01:57.783 ******* ok: [sut] => { "ansible_facts": { "__podman_quadlet_name": "quadlet-demo", "__podman_quadlet_type": "kube", "__podman_rootless": false }, "changed": false } TASK [linux-system-roles.podman : Check user and group information] ************ task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:97 Monday 05 February 2024 04:37:36 +0000 (0:00:00.030) 0:01:57.813 ******* included: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml for sut TASK [linux-system-roles.podman : Get user information] ************************ task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:2 Monday 05 February 2024 04:37:36 +0000 (0:00:00.064) 0:01:57.877 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Fail if user does not exist] ***************** task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:10 Monday 05 February 2024 04:37:36 +0000 (0:00:00.025) 0:01:57.903 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Set group for podman user] ******************* task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:17 Monday 05 February 2024 04:37:36 +0000 (0:00:00.024) 0:01:57.928 ******* ok: [sut] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [linux-system-roles.podman : Get group information] *********************** task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:28 Monday 05 February 2024 04:37:36 +0000 (0:00:00.029) 0:01:57.957 ******* ok: [sut] => { "ansible_facts": { "getent_group": { "root": [ "x", "0", "" ] } }, "changed": false } TASK [linux-system-roles.podman : Set group name] ****************************** task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:36 Monday 05 February 2024 04:37:36 +0000 (0:00:00.289) 0:01:58.247 ******* ok: [sut] => { "ansible_facts": { "__podman_group_name": "root" }, "changed": false } TASK [linux-system-roles.podman : See if getsubids exists] ********************* task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:41 Monday 05 February 2024 04:37:36 +0000 (0:00:00.025) 0:01:58.272 ******* ok: [sut] => { "changed": false, "stat": { "atime": 1707107812.8985565, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "bb5b46ffbafcaa8c4021f3c8b3cb8594f48ef34b", "ctime": 1707107780.3803086, "dev": 51713, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 6792906, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-sharedlib", "mode": "0755", "mtime": 1700557386.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 12640, "uid": 0, "version": "3213196921", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [linux-system-roles.podman : Check user with getsubids] ******************* task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:52 Monday 05 February 2024 04:37:36 +0000 (0:00:00.293) 0:01:58.566 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Check group with getsubids] ****************** task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:56 Monday 05 February 2024 04:37:36 +0000 (0:00:00.027) 0:01:58.594 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Check if user is in subuid file] ************* task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:66 Monday 05 February 2024 04:37:36 +0000 (0:00:00.020) 0:01:58.614 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Fail if user not in subuid file] ************* task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:74 Monday 05 February 2024 04:37:36 +0000 (0:00:00.020) 0:01:58.635 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Check if group is in subgid file] ************ task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:81 Monday 05 February 2024 04:37:36 +0000 (0:00:00.020) 0:01:58.655 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Fail if group not in subgid file] ************ task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:89 Monday 05 February 2024 04:37:36 +0000 (0:00:00.020) 0:01:58.676 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Set per-container variables part 3] ********** task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:102 Monday 05 February 2024 04:37:37 +0000 (0:00:00.020) 0:01:58.696 ******* ok: [sut] => { "ansible_facts": { "__podman_activate_systemd_unit": true, "__podman_images_found": [], "__podman_kube_yamls_raw": [ "quadlet-demo.yml" ], "__podman_service_name": "quadlet-demo.service", "__podman_systemd_scope": "system", "__podman_user_home_dir": "/root", "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [linux-system-roles.podman : Set per-container variables part 4] ********** task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:134 Monday 05 February 2024 04:37:37 +0000 (0:00:00.036) 0:01:58.732 ******* ok: [sut] => { "ansible_facts": { "__podman_quadlet_path": "/etc/containers/systemd" }, "changed": false } TASK [linux-system-roles.podman : Get kube yaml contents] ********************** task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:140 Monday 05 February 2024 04:37:37 +0000 (0:00:00.021) 0:01:58.754 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Set per-container variables part 5] ********** task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:151 Monday 05 February 2024 04:37:37 +0000 (0:00:00.018) 0:01:58.773 ******* ok: [sut] => { "ansible_facts": { "__podman_images": [], "__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo.kube", "__podman_volumes": [] }, "changed": false } TASK [linux-system-roles.podman : Cleanup quadlets] **************************** task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:205 Monday 05 February 2024 04:37:37 +0000 (0:00:00.074) 0:01:58.847 ******* included: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/cleanup_quadlet_spec.yml for sut TASK [linux-system-roles.podman : Stop and disable service] ******************** task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/cleanup_quadlet_spec.yml:4 Monday 05 February 2024 04:37:37 +0000 (0:00:00.041) 0:01:58.889 ******* ok: [sut] => { "changed": false, "failed_when_result": false } MSG: Could not find the requested service quadlet-demo.service: host TASK [linux-system-roles.podman : Remove quadlet file] ************************* task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/cleanup_quadlet_spec.yml:22 Monday 05 February 2024 04:37:37 +0000 (0:00:00.319) 0:01:59.208 ******* ok: [sut] => { "changed": false, "path": "/etc/containers/systemd/quadlet-demo.kube", "state": "absent" } TASK [linux-system-roles.podman : Manage linger] ******************************* task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/cleanup_quadlet_spec.yml:28 Monday 05 February 2024 04:37:37 +0000 (0:00:00.199) 0:01:59.408 ******* included: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/manage_linger.yml for sut TASK [linux-system-roles.podman : Enable linger if needed] ********************* task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/manage_linger.yml:12 Monday 05 February 2024 04:37:37 +0000 (0:00:00.034) 0:01:59.442 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Mark user as not yet needing to cancel linger] *** task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/manage_linger.yml:18 Monday 05 February 2024 04:37:37 +0000 (0:00:00.020) 0:01:59.462 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Mark user for possible linger cancel] ******** task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/manage_linger.yml:22 Monday 05 February 2024 04:37:37 +0000 (0:00:00.020) 0:01:59.483 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Reload systemctl] **************************** task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/cleanup_quadlet_spec.yml:36 Monday 05 February 2024 04:37:37 +0000 (0:00:00.024) 0:01:59.507 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Prune images no longer in use] *************** task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/cleanup_quadlet_spec.yml:45 Monday 05 February 2024 04:37:37 +0000 (0:00:00.020) 0:01:59.528 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Remove network] ****************************** task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/cleanup_quadlet_spec.yml:53 Monday 05 February 2024 04:37:37 +0000 (0:00:00.021) 0:01:59.550 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Create and update quadlets] ****************** task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:209 Monday 05 February 2024 04:37:37 +0000 (0:00:00.021) 0:01:59.571 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Set per-container variables part 0] ********** task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:14 Monday 05 February 2024 04:37:37 +0000 (0:00:00.020) 0:01:59.591 ******* ok: [sut] => { "ansible_facts": { "__podman_quadlet_file_src": "quadlet-demo.network", "__podman_quadlet_spec": {}, "__podman_quadlet_str": "[Network]\nSubnet=192.168.30.0/24\nGateway=192.168.30.1\nLabel=app=wordpress", "__podman_quadlet_template_src": "" }, "changed": false } TASK [linux-system-roles.podman : Set per-container variables part 1] ********** task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:40 Monday 05 February 2024 04:37:37 +0000 (0:00:00.028) 0:01:59.620 ******* ok: [sut] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_pull_image": true, "__podman_state": "absent", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [linux-system-roles.podman : Fail if no quadlet spec is given] ************ task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:57 Monday 05 February 2024 04:37:37 +0000 (0:00:00.024) 0:01:59.644 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Set per-container variables part 2] ********** task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:70 Monday 05 February 2024 04:37:38 +0000 (0:00:00.052) 0:01:59.697 ******* ok: [sut] => { "ansible_facts": { "__podman_quadlet_name": "quadlet-demo", "__podman_quadlet_type": "network", "__podman_rootless": false }, "changed": false } TASK [linux-system-roles.podman : Check user and group information] ************ task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/handle_quadlet_spec.yml:97 Monday 05 February 2024 04:37:38 +0000 (0:00:00.029) 0:01:59.726 ******* included: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml for sut TASK [linux-system-roles.podman : Get user information] ************************ task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:2 Monday 05 February 2024 04:37:38 +0000 (0:00:00.037) 0:01:59.764 ******* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.podman : Fail if user does not exist] ***************** task path: /WORKDIR/git-fix-linger-handlinglwssa1kt/tests/roles/linux-system-roles.podman/tasks/handle_user_group.yml:10 Monday 05 February 2024 04:37:38 +0000 (0:00:00.025) 0:01:59.789 ******* skipping: [sut]llections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:27 Monday 05 February 2024 04:37:34 +0000 (0:00:00.026) 0:01:56.332 ******* ok: [sut] => { "ansible_facts": { "getent_group": { "root": [ "x", "0", "" ] } }, "changed": false } TASK [fedora.linux_system_roles.podman : Set group name] *********************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:34 Monday 05 February 2024 04:37:34 +0000 (0:00:00.201) 0:01:56.533 ******* ok: [sut] => { "ansible_facts": { "__podman_group_name": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:38 Monday 05 February 2024 04:37:34 +0000 (0:00:00.026) 0:01:56.560 ******* ok: [sut] => { "changed": false, "stat": { "atime": 1707107813.7932265, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "bb5b46ffbafcaa8c4021f3c8b3cb8594f48ef34b", "ctime": 1707107779.304328, "dev": 51713, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 6792906, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-sharedlib", "mode": "0755", "mtime": 1700557386.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 12640, "uid": 0, "version": "3297496774", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************ task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:49 Monday 05 February 2024 04:37:34 +0000 (0:00:00.202) 0:01:56.762 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check group with getsubids] *********** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:53 Monday 05 February 2024 04:37:34 +0000 (0:00:00.018) 0:01:56.781 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if user is in subuid file] ****** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:63 Monday 05 February 2024 04:37:34 +0000 (0:00:00.017) 0:01:56.799 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:71 Monday 05 February 2024 04:37:34 +0000 (0:00:00.018) 0:01:56.817 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if group is in subgid file] ***** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:78 Monday 05 February 2024 04:37:34 +0000 (0:00:00.051) 0:01:56.869 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] ***** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:86 Monday 05 February 2024 04:37:34 +0000 (0:00:00.018) 0:01:56.887 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62 Monday 05 February 2024 04:37:34 +0000 (0:00:00.018) 0:01:56.905 ******* ok: [sut] => { "ansible_facts": { "__podman_activate_systemd_unit": true, "__podman_images_found": [], "__podman_kube_yamls_raw": "", "__podman_service_name": "", "__podman_systemd_scope": "system", "__podman_user_home_dir": "/root", "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73 Monday 05 February 2024 04:37:34 +0000 (0:00:00.033) 0:01:56.939 ******* ok: [sut] => { "ansible_facts": { "__podman_quadlet_path": "/etc/containers/systemd" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get kube yaml contents] *************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77 Monday 05 February 2024 04:37:34 +0000 (0:00:00.021) 0:01:56.961 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_state != \"absent\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87 Monday 05 February 2024 04:37:34 +0000 (0:00:00.017) 0:01:56.979 ******* ok: [sut] => { "ansible_facts": { "__podman_images": [], "__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo.yml", "__podman_volumes": [] }, "changed": false } TASK [fedora.linux_system_roles.podman : Cleanup quadlets] ********************* task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:103 Monday 05 February 2024 04:37:34 +0000 (0:00:00.046) 0:01:57.025 ******* included: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml for sut TASK [fedora.linux_system_roles.podman : Stop and disable service] ************* task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:4 Monday 05 February 2024 04:37:34 +0000 (0:00:00.038) 0:01:57.064 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_service_name | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Remove quadlet file] ****************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:22 Monday 05 February 2024 04:37:34 +0000 (0:00:00.020) 0:01:57.085 ******* ok: [sut] => { "changed": false, "path": "/etc/containers/systemd/quadlet-demo.yml", "state": "absent" } TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:28 Monday 05 February 2024 04:37:35 +0000 (0:00:00.201) 0:01:57.287 ******* included: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for sut TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Monday 05 February 2024 04:37:35 +0000 (0:00:00.030) 0:01:57.317 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Monday 05 February 2024 04:37:35 +0000 (0:00:00.018) 0:01:57.336 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Monday 05 February 2024 04:37:35 +0000 (0:00:00.017) 0:01:57.354 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Reload systemctl] ********************* task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:36 Monday 05 February 2024 04:37:35 +0000 (0:00:00.018) 0:01:57.372 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_file_removed is changed", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Prune images no longer in use] ******** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:45 Monday 05 February 2024 04:37:35 +0000 (0:00:00.019) 0:01:57.392 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_file_removed is changed", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Remove network] *********************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:53 Monday 05 February 2024 04:37:35 +0000 (0:00:00.018) 0:01:57.410 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_file_removed is changed", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create and update quadlets] *********** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:107 Monday 05 February 2024 04:37:35 +0000 (0:00:00.019) 0:01:57.430 ******* skipping: [sut] => { "changed": false, "false_condition": "__podman_state != \"absent\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14 Monday 05 February 2024 04:37:35 +0000 (0:00:00.017) 0:01:57.448 ******* ok: [sut] => { "ansible_facts": { "__podman_quadlet_file_src": "quadlet-demo.kube", "__podman_quadlet_spec": {}, "__podman_quadlet_str": "[Install]\nWantedBy=default.target\n\n[Unit]\nRequires=quadlet-demo-mysql.service\nAfter=quadlet-demo-mysql.service\n\n[Kube]\n# Point to the yaml file in the same directory\nYaml=quadlet-demo.yml\n# Use the quadlet-demo network\nNetwork=quadlet-demo.network\n# Publish the envoy proxy data port\nPublishPort=8000:8080\n# Publish the envoy proxy admin port\nPublishPort=9000:9901\n# Use the envoy proxy config map in the same directory\nConfigMap=envoy-proxy-configmap.yml", "__podman_quadlet_template_src": "" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25 Monday 05 February 2024 04:37:35 +0000 (0:00:00.063) 0:01:57.512 ******* ok: [sut] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_pull_image": true, "__podman_state": "absent", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] ***** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35 Monday 05 February 2024 04:37:35 +0000 (0:00:00.025) 0:01:57.538 ******* skipping: [sut] => { "changed": false, "false_condition": "not __podman_quadlet_file_src", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] *** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48 Monday 05 February 2024 04:37:35 +0000 (0:00:00.020) 0:01:57.558 ******* ok: [sut] => { "ansible_facts": { "__podman_quadlet_name": "quadlet-demo", "__podman_quadlet_type": "kube", "__podman_rootless": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57 Monday 05 February 2024 04:37:35 +0000 (0:00:00.029) 0:01:57.588 ******* included: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for sut TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Monday 05 February 2024 04:37:35 +0000 (0:00:00.036) 0:01:57.625 ******* skipping: [sut] => { "changed": false, "false_condition": "'getent_passwd' not in ansible_facts or __podman_user not in ansible_facts['getent_passwd']", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Monday 05 February 2024 04:37:35 +0000 (0:00:00.021) 0:01:57.646 ******* skipping: [sut] => { "changed": false, "false_condition": "not ansible_facts[\"getent_passwd\"][__podman_user]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /WORKDIR/git-fix-linger-handlinggb26lsz1/.collection/ansible_col