ansible-playbook 2.9.27 config file = None configured module search path = ['/root/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules'] ansible python module location = /usr/local/lib/python3.9/site-packages/ansible executable location = /usr/local/bin/ansible-playbook python version = 3.9.19 (main, May 16 2024, 11:40:09) [GCC 8.5.0 20210514 (Red Hat 8.5.0-22)] No config file found; using defaults [WARNING]: running playbook inside collection fedora.linux_system_roles Skipping callback 'actionable', as we already have a stdout callback. Skipping callback 'counter_enabled', as we already have a stdout callback. Skipping callback 'debug', as we already have a stdout callback. Skipping callback 'dense', as we already have a stdout callback. Skipping callback 'dense', as we already have a stdout callback. Skipping callback 'full_skip', as we already have a stdout callback. Skipping callback 'json', as we already have a stdout callback. Skipping callback 'minimal', as we already have a stdout callback. Skipping callback 'null', as we already have a stdout callback. Skipping callback 'oneline', as we already have a stdout callback. Skipping callback 'selective', as we already have a stdout callback. Skipping callback 'skippy', as we already have a stdout callback. Skipping callback 'stderr', as we already have a stdout callback. Skipping callback 'unixy', as we already have a stdout callback. Skipping callback 'yaml', as we already have a stdout callback. PLAYBOOK: tests_quadlet_demo.yml *********************************************** 2 plays in /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_demo.yml PLAY [all] ********************************************************************* META: ran handlers TASK [Include vault variables] ************************************************* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_demo.yml:5 Saturday 24 August 2024 12:33:51 -0400 (0:00:00.030) 0:00:00.030 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_test_password": { "__ansible_vault": "$ANSIBLE_VAULT;1.1;AES256\n35383939616163653333633431363463313831383037386236646138333162396161356130303461\n3932623930643263313563336163316337643562333936360a363538636631313039343233383732\n38666530383538656639363465313230343533386130303833336434303438333161656262346562\n3362626538613031640a663330613638366132356534363534353239616666653466353961323533\n6565\n" }, "mysql_container_root_password": { "__ansible_vault": "$ANSIBLE_VAULT;1.1;AES256\n61333932373230333539663035366431326163363166363036323963623131363530326231303634\n6635326161643165363366323062333334363730376631660a393566366139353861656364656661\n38653463363837336639363032646433666361646535366137303464623261313663643336306465\n6264663730656337310a343962353137386238383064646533366433333437303566656433386233\n34343235326665646661623131643335313236313131353661386338343366316261643634653633\n3832313034366536616531323963333234326461353130303532\n" } }, "ansible_included_var_files": [ "/tmp/podman-mRR/tests/vars/vault-variables.yml" ], "changed": false } META: ran handlers META: ran handlers PLAY [Deploy the quadlet demo app] ********************************************* TASK [Gathering Facts] ********************************************************* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_demo.yml:9 Saturday 24 August 2024 12:33:51 -0400 (0:00:00.042) 0:00:00.073 ******* ok: [managed_node1] META: ran handlers TASK [Test is only supported on x86_64] **************************************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_demo.yml:38 Saturday 24 August 2024 12:33:52 -0400 (0:00:01.126) 0:00:01.200 ******* skipping: [managed_node1] => {} META: TASK [Generate certificates] *************************************************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_demo.yml:51 Saturday 24 August 2024 12:33:52 -0400 (0:00:00.091) 0:00:01.292 ******* TASK [fedora.linux_system_roles.certificate : Set version specific variables] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:2 Saturday 24 August 2024 12:33:52 -0400 (0:00:00.069) 0:00:01.362 ******* included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/set_vars.yml for managed_node1 TASK [fedora.linux_system_roles.certificate : Ensure ansible_facts used by role] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/set_vars.yml:2 Saturday 24 August 2024 12:33:52 -0400 (0:00:00.037) 0:00:01.399 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.certificate : Check if system is ostree] ******* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/set_vars.yml:10 Saturday 24 August 2024 12:33:52 -0400 (0:00:00.021) 0:00:01.420 ******* ok: [managed_node1] => { "changed": false, "stat": { "exists": false } } TASK [fedora.linux_system_roles.certificate : Set flag to indicate system is ostree] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/set_vars.yml:15 Saturday 24 August 2024 12:33:53 -0400 (0:00:00.510) 0:00:01.931 ******* ok: [managed_node1] => { "ansible_facts": { "__certificate_is_ostree": false }, "changed": false } TASK [fedora.linux_system_roles.certificate : Set platform/version specific variables] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/set_vars.yml:19 Saturday 24 August 2024 12:33:53 -0400 (0:00:00.035) 0:00:01.966 ******* skipping: [managed_node1] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [managed_node1] => (item=CentOS.yml) => { "ansible_loop_var": "item", "changed": false, "item": "CentOS.yml", "skip_reason": "Conditional result was False" } skipping: [managed_node1] => (item=CentOS_8.yml) => { "ansible_loop_var": "item", "changed": false, "item": "CentOS_8.yml", "skip_reason": "Conditional result was False" } skipping: [managed_node1] => (item=CentOS_8.yml) => { "ansible_loop_var": "item", "changed": false, "item": "CentOS_8.yml", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.certificate : Ensure certificate role dependencies are installed] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:5 Saturday 24 August 2024 12:33:53 -0400 (0:00:00.055) 0:00:02.022 ******* ok: [managed_node1] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do lsrpackages: python3-cryptography python3-dbus python3-pyasn1 TASK [fedora.linux_system_roles.certificate : Ensure provider packages are installed] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:23 Saturday 24 August 2024 12:33:56 -0400 (0:00:03.099) 0:00:05.121 ******* ok: [managed_node1] => (item=certmonger) => { "__certificate_provider": "certmonger", "ansible_loop_var": "__certificate_provider", "changed": false, "rc": 0, "results": [] } MSG: Nothing to do lsrpackages: certmonger TASK [fedora.linux_system_roles.certificate : Ensure pre-scripts hooks directory exists] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:35 Saturday 24 August 2024 12:33:59 -0400 (0:00:02.809) 0:00:07.930 ******* ok: [managed_node1] => (item=certmonger) => { "__certificate_provider": "certmonger", "ansible_loop_var": "__certificate_provider", "changed": false, "gid": 0, "group": "root", "mode": "0700", "owner": "root", "path": "/etc/certmonger//pre-scripts", "secontext": "unconfined_u:object_r:etc_t:s0", "size": 6, "state": "directory", "uid": 0 } TASK [fedora.linux_system_roles.certificate : Ensure post-scripts hooks directory exists] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:61 Saturday 24 August 2024 12:34:00 -0400 (0:00:00.731) 0:00:08.661 ******* ok: [managed_node1] => (item=certmonger) => { "__certificate_provider": "certmonger", "ansible_loop_var": "__certificate_provider", "changed": false, "gid": 0, "group": "root", "mode": "0700", "owner": "root", "path": "/etc/certmonger//post-scripts", "secontext": "unconfined_u:object_r:etc_t:s0", "size": 6, "state": "directory", "uid": 0 } TASK [fedora.linux_system_roles.certificate : Ensure provider service is running] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:90 Saturday 24 August 2024 12:34:00 -0400 (0:00:00.468) 0:00:09.130 ******* ok: [managed_node1] => (item=certmonger) => { "__certificate_provider": "certmonger", "ansible_loop_var": "__certificate_provider", "changed": false, "enabled": true, "name": "certmonger", "state": "started", "status": { "ActiveEnterTimestamp": "Sat 2024-08-24 12:29:03 EDT", "ActiveEnterTimestampMonotonic": "346295924", "ActiveExitTimestampMonotonic": "0", "ActiveState": "active", "After": "sysinit.target dbus.service systemd-journald.socket network.target basic.target dbus.socket system.slice syslog.target", "AllowIsolate": "no", "AllowedCPUs": "", "AllowedMemoryNodes": "", "AmbientCapabilities": "", "AssertResult": "yes", "AssertTimestamp": "Sat 2024-08-24 12:29:03 EDT", "AssertTimestampMonotonic": "346284245", "Before": "shutdown.target multi-user.target", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "BusName": "org.fedorahosted.certmonger", "CPUAccounting": "no", "CPUAffinity": "", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "[not set]", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "no", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf", "CollectMode": "inactive", "ConditionResult": "yes", "ConditionTimestamp": "Sat 2024-08-24 12:29:03 EDT", "ConditionTimestampMonotonic": "346284243", "ConfigurationDirectoryMode": "0755", "Conflicts": "shutdown.target", "ControlGroup": "/system.slice/certmonger.service", "ControlPID": "0", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "Delegate": "no", "Description": "Certificate monitoring and PKI enrollment", "DevicePolicy": "auto", "DynamicUser": "no", "EffectiveCPUs": "", "EffectiveMemoryNodes": "", "EnvironmentFiles": "/etc/sysconfig/certmonger (ignore_errors=yes)", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "12107", "ExecMainStartTimestamp": "Sat 2024-08-24 12:29:03 EDT", "ExecMainStartTimestampMonotonic": "346285051", "ExecMainStatus": "0", "ExecStart": "{ path=/usr/sbin/certmonger ; argv[]=/usr/sbin/certmonger -S -p /run/certmonger.pid -n $OPTS ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FragmentPath": "/usr/lib/systemd/system/certmonger.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOSchedulingClass": "0", "IOSchedulingPriority": "0", "IOWeight": "[not set]", "IPAccounting": "no", "IPEgressBytes": "18446744073709551615", "IPEgressPackets": "18446744073709551615", "IPIngressBytes": "18446744073709551615", "IPIngressPackets": "18446744073709551615", "Id": "certmonger.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestamp": "Sat 2024-08-24 12:29:03 EDT", "InactiveExitTimestampMonotonic": "346285098", "InvocationID": "f4bd1c0399d64c7cb53a9acf6dd73285", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "control-group", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "0", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "65536", "LimitMEMLOCKSoft": "65536", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "262144", "LimitNOFILESoft": "1024", "LimitNPROC": "14003", "LimitNPROCSoft": "14003", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "14003", "LimitSIGPENDINGSoft": "14003", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "12107", "MemoryAccounting": "yes", "MemoryCurrent": "3100672", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemorySwapMax": "infinity", "MountAPIVFS": "no", "MountFlags": "", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAMask": "", "NUMAPolicy": "n/a", "Names": "certmonger.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "PIDFile": "/run/certmonger.pid", "PartOf": "dbus.service", "PermissionsStartOnly": "no", "Perpetual": "no", "PrivateDevices": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "RemainAfterExit": "no", "RemoveIPC": "no", "Requires": "sysinit.target system.slice dbus.socket", "Restart": "no", "RestartUSec": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "Slice": "system.slice", "StandardError": "inherit", "StandardInput": "null", "StandardInputData": "", "StandardOutput": "journal", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StateChangeTimestamp": "Sat 2024-08-24 12:29:03 EDT", "StateChangeTimestampMonotonic": "346295924", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "running", "SuccessAction": "none", "SyslogFacility": "3", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "0", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "1", "TasksMax": "22405", "TimeoutStartUSec": "1min 30s", "TimeoutStopUSec": "1min 30s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "dbus", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "disabled", "UnitFileState": "enabled", "UtmpMode": "init", "WantedBy": "multi-user.target", "WatchdogTimestamp": "Sat 2024-08-24 12:29:03 EDT", "WatchdogTimestampMonotonic": "346295922", "WatchdogUSec": "0" } } TASK [fedora.linux_system_roles.certificate : Ensure certificate requests] ***** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:101 Saturday 24 August 2024 12:34:01 -0400 (0:00:01.061) 0:00:10.192 ******* changed: [managed_node1] => (item={'name': 'quadlet_demo', 'dns': ['localhost'], 'ca': 'self-sign'}) => { "ansible_loop_var": "item", "changed": true, "item": { "ca": "self-sign", "dns": [ "localhost" ], "name": "quadlet_demo" } } MSG: Certificate requested (new). TASK [fedora.linux_system_roles.certificate : Slurp the contents of the files] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:152 Saturday 24 August 2024 12:34:02 -0400 (0:00:01.256) 0:00:11.448 ******* ok: [managed_node1] => (item=['cert', {'name': 'quadlet_demo', 'dns': ['localhost'], 'ca': 'self-sign'}]) => { "ansible_loop_var": "item", "changed": false, "content": "LS0tLS1CRUdJTiBDRVJUSUZJQ0FURS0tLS0tCk1JSURnekNDQW11Z0F3SUJBZ0lSQUxxZnN3V1ZSa25NdTg3clBDa0JaVnN3RFFZSktvWklodmNOQVFFTEJRQXcKVURFZ01CNEdBMVVFQXd3WFRHOWpZV3dnVTJsbmJtbHVaeUJCZFhSb2IzSnBkSGt4TERBcUJnTlZCQU1NSTJKaApPV1ppTXpBMUxUazFORFkwT1dOakxXSmlZMlZsWWpOakxUSTVNREUyTlRVNU1CNFhEVEkwTURneU5ERTJNelF3Ck1sb1hEVEkxTURneU5ERTJNamt3TTFvd0ZERVNNQkFHQTFVRUF4TUpiRzlqWVd4b2IzTjBNSUlCSWpBTkJna3EKaGtpRzl3MEJBUUVGQUFPQ0FROEFNSUlCQ2dLQ0FRRUFrekNka1R1VnAxSzRxSmxtWFhGTG1Bb3ZqbVVhT1FYRApVRk5UOVkyeHM2VHZleEFHaEh0S1Z4b3VxSW8zdUhOT2VTd3Z0SEU3NXBJMzVmeDlPdVpHbE9sWEhsUFVycmpiCm5JOVY3amtSWUhWdi9iSlJxSzZOYmc3dXF0c0VuN0FwVDVadFdXWWw5amM0Ykg4OEpDb2Rka0wrLy9zYllwRzkKa2hQaGo0NWhESVkzUTg4cW0vU0R1Z1dybUNHa0lMaFZHclQrM0RpaWI0ZFpGYVZZKzB4MnBOeWQ3YXdYZW1hQwphd3A0UHlJeEhSbjJuUUxjblZYa0hVVHhnOUJPcTNLWVM0dUsvY0VjclFHeEVNV0pkR0lmUmMxUUx6dzcvT3ArCnRNWDBVdUZLOFFwT2ZBYlIwNjlNdUVhSXlTZzFtMmFTK1FxNjhGV3BOMURBNmloR05CMkU3UUlEQVFBQm80R1QKTUlHUU1Bc0dBMVVkRHdRRUF3SUZvREFVQmdOVkhSRUVEVEFMZ2dsc2IyTmhiR2h2YzNRd0hRWURWUjBsQkJZdwpGQVlJS3dZQkJRVUhBd0VHQ0NzR0FRVUZCd01DTUF3R0ExVWRFd0VCL3dRQ01BQXdIUVlEVlIwT0JCWUVGSUVzClVIenQ5dTRIMzVKNzFsQzJyeUNkdmtabE1COEdBMVVkSXdRWU1CYUFGSG95TXRWMVNheE1UVW4wcm42UVkwNUwKM28rd01BMEdDU3FHU0liM0RRRUJDd1VBQTRJQkFRQnpVY0l6OHozQVNOVXg5MnYxYlc2OFFsdlpCditROHJZUQo1M0JYKzFYOSsyVENUVE91bUxabkloZnFlQVJLbkJIdjE3Lzg4ZFBlQXNPVTdlT2JXWXpwOGU2TVJaSHIvWW45CjJXMXJQakQvLzRERFZaeUduOXZLc1Yzc1RiR1ExZi9HN2UrZVljd0JDSWNKOTNOajlZM292QnRwWUJpWTdFNnoKOUhjT2txSGlWSnJxaS9VWmtLUk1TcjRobU9CeFhvc280ZXhxUUVaU0xocjB1NnNKTWx4REtlMk5CLysxbGNPaQo0bHZ0ODVHbTZONXU1QWJ0amx1Y1ZGNXpKTjFtVlRpd1FaZTdFakUweXA0NkxrVVBkMlRwQ3ROSmRhWmpENVZICmYyUTFIWHpuN0VPSjVvODlHcWhEaExhUi9VVjN0SU4rbEFUUXQwREVIQ1k4TFpBNDZNOGQKLS0tLS1FTkQgQ0VSVElGSUNBVEUtLS0tLQo=", "encoding": "base64", "item": [ "cert", { "ca": "self-sign", "dns": [ "localhost" ], "name": "quadlet_demo" } ], "source": "/etc/pki/tls/certs/quadlet_demo.crt" } ok: [managed_node1] => (item=['key', {'name': 'quadlet_demo', 'dns': ['localhost'], 'ca': 'self-sign'}]) => { "ansible_loop_var": "item", "changed": false, "content": "LS0tLS1CRUdJTiBQUklWQVRFIEtFWS0tLS0tCk1JSUV2QUlCQURBTkJna3Foa2lHOXcwQkFRRUZBQVNDQktZd2dnU2lBZ0VBQW9JQkFRQ1RNSjJSTzVXblVyaW8KbVdaZGNVdVlDaStPWlJvNUJjTlFVMVAxamJHenBPOTdFQWFFZTBwWEdpNm9pamU0YzA1NUxDKzBjVHZta2pmbAovSDA2NWthVTZWY2VVOVN1dU51Y2oxWHVPUkZnZFcvOXNsR29ybzF1RHU2cTJ3U2ZzQ2xQbG0xWlppWDJOemhzCmZ6d2tLaDEyUXY3Lyt4dGlrYjJTRStHUGptRU1oamREenlxYjlJTzZCYXVZSWFRZ3VGVWF0UDdjT0tKdmgxa1YKcFZqN1RIYWszSjN0ckJkNlpvSnJDbmcvSWpFZEdmYWRBdHlkVmVRZFJQR0QwRTZyY3BoTGk0cjl3Unl0QWJFUQp4WWwwWWg5RnpWQXZQRHY4Nm42MHhmUlM0VXJ4Q2s1OEJ0SFRyMHk0Um9qSktEV2JacEw1Q3Jyd1ZhazNVTURxCktFWTBIWVR0QWdNQkFBRUNnZ0VBVEFJWDdDNjcybHNpVnBaWXdUcEhVaTA0WDYwZi9JQllXdnVjNkxvU0lQVzAKWVhWbUdEWGVydG9yYkFkc1BNajJwaVpHUUVVVkxMMG5OcEhMVFlZcXZxMDh1c0dKbSt5RUkvWDA1djc3VlRkdgo3SytQOGE2cUJFS2J4d2t3K3hHcVU1b1h2OG03SXNsalRzdnBqN2RiM1gwOVY0MGhtQ1l3b2VXNFhqeWxLeVMvCm1VUUZPK3VpaGd5VUR3NEszdHJEZUhNUXhYVlRQTkNRY3pxSXk4eXB5aFJoVXBQdGtzdGxkbHQ4K1dHa29qNloKdTBDRUNGSkpvK2FEYnRWMUhPbEVqK1NUTllxd3BTSUw0RUlMQ0FyOHAxZ3BKZHE1V1NQNlpaNXNtYmtZVm5Rego2QnkyN2pLb1VuYXRJOW8zM0FDZm5HMkM5Nm1sRmdLNWFTUlBjUGFaRlFLQmdRREQ4dXlPUnJ6c0xTM1czVTBGCnRZM3hneEtsVDNyOWFjeGNIY1NIRi82akdXYXNSZzI5U2JLbTdxZUhiVWp0bHB6bkk1VWN5cTlIYVR3Z1U5OHIKdGVuekRYQTUzbXhJZ3ZtdGRlRlk3NnZ3YzhYbFlaRTEyQUFoa0RpSHVTOEl1M2NuN2tEZ1FUd2UwZnZUWWxEMgpHWWwwYmhtL1JnbzlTbWYyNUtWR0x3and6d0tCZ1FEQVRGT243ZkV6MHdSSE9maWlHbTB6MEhmMUl3RjJFN2ZvClQ4YWxwaXhQMWg2V2dUaFdwUWpvckNJTE1LUHRiN25EdEJTcllwN1o5VGNYMzVnbWFjbUpVRmZpUUQ2Y1l0V1QKTWxad2E5cm5Dalh6QloxcUNqZnNyZUI0c2RReEdpWmxOUFNoWkFYcWdCajd6bjRGY3dUVDFGQ0VpSGhGd0RTYwpaUG5QQzlYRmd3S0JnR1Y3ck5STEFwRjNIS1NKWE5pMkhOdVdQTlB0QkErMithWkRySU1qQkRoSWNZcUxoWlloClV5dHRzc3BLNlZoVGh0T1c5Q1ZETytCS09sK3IwSXQ4M2J1Sm5EUDhaQ3h4MlFuallXQ0VVSkpxQTNZM1A5T1oKR1FoSS8xNGJwS1pySU4xSFljWkRRdlhiWHdjdGhpelBNUWRoOW1lL3lKSFJ0V0tETEszYXNsQzFBb0dBQ0s2cApqRWg2L25zRmsyMnpveVlNVGkyWWJQSUNxK0VpQ1pRb2hPd216Z2Y5NDQrOENuVnpISVpCRno0LytJVjEzcmN2Cnc3azFNT1FZYzdSOXV5WkZtalNOYzBLeFVRQkpZbGNBZmh6d2lrUDB3Wjg4UjBCazNBUFJ3Y0ZOc2xYYTYvRGoKSnJKTWJoVVFpN1g4REtUdno5eWpNZ2l6WFhyYkpVQ040dUpHNlo4Q2dZQVgwOExCd1FuWHJiOHZEVE9scHRHbwpCMHU2NnVKN3pWelFLNmxSMXR0N3MzMk9RSjJLdWt3WkVCbVZZMkFjckkxNW1rc2dqVTNWWFY1d1prWTl2K2tiCmpFaFNJT2RxZGZIa2x4ZnBwWTVneEY2TjA2WUpOLzdJNzNEWWFvZXdIVU9BTVVRSTdxZVB5SEpURDZjbHp3Z2QKeEtpWlFsM3JRMytlOU9qNVFITmlYdz09Ci0tLS0tRU5EIFBSSVZBVEUgS0VZLS0tLS0K", "encoding": "base64", "item": [ "key", { "ca": "self-sign", "dns": [ "localhost" ], "name": "quadlet_demo" } ], "source": "/etc/pki/tls/private/quadlet_demo.key" } ok: [managed_node1] => (item=['ca', {'name': 'quadlet_demo', 'dns': ['localhost'], 'ca': 'self-sign'}]) => { "ansible_loop_var": "item", "changed": false, "content": "LS0tLS1CRUdJTiBDRVJUSUZJQ0FURS0tLS0tCk1JSURnekNDQW11Z0F3SUJBZ0lSQUxxZnN3V1ZSa25NdTg3clBDa0JaVnN3RFFZSktvWklodmNOQVFFTEJRQXcKVURFZ01CNEdBMVVFQXd3WFRHOWpZV3dnVTJsbmJtbHVaeUJCZFhSb2IzSnBkSGt4TERBcUJnTlZCQU1NSTJKaApPV1ppTXpBMUxUazFORFkwT1dOakxXSmlZMlZsWWpOakxUSTVNREUyTlRVNU1CNFhEVEkwTURneU5ERTJNelF3Ck1sb1hEVEkxTURneU5ERTJNamt3TTFvd0ZERVNNQkFHQTFVRUF4TUpiRzlqWVd4b2IzTjBNSUlCSWpBTkJna3EKaGtpRzl3MEJBUUVGQUFPQ0FROEFNSUlCQ2dLQ0FRRUFrekNka1R1VnAxSzRxSmxtWFhGTG1Bb3ZqbVVhT1FYRApVRk5UOVkyeHM2VHZleEFHaEh0S1Z4b3VxSW8zdUhOT2VTd3Z0SEU3NXBJMzVmeDlPdVpHbE9sWEhsUFVycmpiCm5JOVY3amtSWUhWdi9iSlJxSzZOYmc3dXF0c0VuN0FwVDVadFdXWWw5amM0Ykg4OEpDb2Rka0wrLy9zYllwRzkKa2hQaGo0NWhESVkzUTg4cW0vU0R1Z1dybUNHa0lMaFZHclQrM0RpaWI0ZFpGYVZZKzB4MnBOeWQ3YXdYZW1hQwphd3A0UHlJeEhSbjJuUUxjblZYa0hVVHhnOUJPcTNLWVM0dUsvY0VjclFHeEVNV0pkR0lmUmMxUUx6dzcvT3ArCnRNWDBVdUZLOFFwT2ZBYlIwNjlNdUVhSXlTZzFtMmFTK1FxNjhGV3BOMURBNmloR05CMkU3UUlEQVFBQm80R1QKTUlHUU1Bc0dBMVVkRHdRRUF3SUZvREFVQmdOVkhSRUVEVEFMZ2dsc2IyTmhiR2h2YzNRd0hRWURWUjBsQkJZdwpGQVlJS3dZQkJRVUhBd0VHQ0NzR0FRVUZCd01DTUF3R0ExVWRFd0VCL3dRQ01BQXdIUVlEVlIwT0JCWUVGSUVzClVIenQ5dTRIMzVKNzFsQzJyeUNkdmtabE1COEdBMVVkSXdRWU1CYUFGSG95TXRWMVNheE1UVW4wcm42UVkwNUwKM28rd01BMEdDU3FHU0liM0RRRUJDd1VBQTRJQkFRQnpVY0l6OHozQVNOVXg5MnYxYlc2OFFsdlpCditROHJZUQo1M0JYKzFYOSsyVENUVE91bUxabkloZnFlQVJLbkJIdjE3Lzg4ZFBlQXNPVTdlT2JXWXpwOGU2TVJaSHIvWW45CjJXMXJQakQvLzRERFZaeUduOXZLc1Yzc1RiR1ExZi9HN2UrZVljd0JDSWNKOTNOajlZM292QnRwWUJpWTdFNnoKOUhjT2txSGlWSnJxaS9VWmtLUk1TcjRobU9CeFhvc280ZXhxUUVaU0xocjB1NnNKTWx4REtlMk5CLysxbGNPaQo0bHZ0ODVHbTZONXU1QWJ0amx1Y1ZGNXpKTjFtVlRpd1FaZTdFakUweXA0NkxrVVBkMlRwQ3ROSmRhWmpENVZICmYyUTFIWHpuN0VPSjVvODlHcWhEaExhUi9VVjN0SU4rbEFUUXQwREVIQ1k4TFpBNDZNOGQKLS0tLS1FTkQgQ0VSVElGSUNBVEUtLS0tLQo=", "encoding": "base64", "item": [ "ca", { "ca": "self-sign", "dns": [ "localhost" ], "name": "quadlet_demo" } ], "source": "/etc/pki/tls/certs/quadlet_demo.crt" } TASK [fedora.linux_system_roles.certificate : Create return data] ************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:160 Saturday 24 August 2024 12:34:04 -0400 (0:00:01.392) 0:00:12.841 ******* ok: [managed_node1] => { "ansible_facts": { "certificate_test_certs": { "quadlet_demo": { "ca": "/etc/pki/tls/certs/quadlet_demo.crt", "ca_content": "-----BEGIN CERTIFICATE-----\nMIIDgzCCAmugAwIBAgIRALqfswWVRknMu87rPCkBZVswDQYJKoZIhvcNAQELBQAw\nUDEgMB4GA1UEAwwXTG9jYWwgU2lnbmluZyBBdXRob3JpdHkxLDAqBgNVBAMMI2Jh\nOWZiMzA1LTk1NDY0OWNjLWJiY2VlYjNjLTI5MDE2NTU5MB4XDTI0MDgyNDE2MzQw\nMloXDTI1MDgyNDE2MjkwM1owFDESMBAGA1UEAxMJbG9jYWxob3N0MIIBIjANBgkq\nhkiG9w0BAQEFAAOCAQ8AMIIBCgKCAQEAkzCdkTuVp1K4qJlmXXFLmAovjmUaOQXD\nUFNT9Y2xs6TvexAGhHtKVxouqIo3uHNOeSwvtHE75pI35fx9OuZGlOlXHlPUrrjb\nnI9V7jkRYHVv/bJRqK6Nbg7uqtsEn7ApT5ZtWWYl9jc4bH88JCoddkL+//sbYpG9\nkhPhj45hDIY3Q88qm/SDugWrmCGkILhVGrT+3Diib4dZFaVY+0x2pNyd7awXemaC\nawp4PyIxHRn2nQLcnVXkHUTxg9BOq3KYS4uK/cEcrQGxEMWJdGIfRc1QLzw7/Op+\ntMX0UuFK8QpOfAbR069MuEaIySg1m2aS+Qq68FWpN1DA6ihGNB2E7QIDAQABo4GT\nMIGQMAsGA1UdDwQEAwIFoDAUBgNVHREEDTALgglsb2NhbGhvc3QwHQYDVR0lBBYw\nFAYIKwYBBQUHAwEGCCsGAQUFBwMCMAwGA1UdEwEB/wQCMAAwHQYDVR0OBBYEFIEs\nUHzt9u4H35J71lC2ryCdvkZlMB8GA1UdIwQYMBaAFHoyMtV1SaxMTUn0rn6QY05L\n3o+wMA0GCSqGSIb3DQEBCwUAA4IBAQBzUcIz8z3ASNUx92v1bW68QlvZBv+Q8rYQ\n53BX+1X9+2TCTTOumLZnIhfqeARKnBHv17/88dPeAsOU7eObWYzp8e6MRZHr/Yn9\n2W1rPjD//4DDVZyGn9vKsV3sTbGQ1f/G7e+eYcwBCIcJ93Nj9Y3ovBtpYBiY7E6z\n9HcOkqHiVJrqi/UZkKRMSr4hmOBxXoso4exqQEZSLhr0u6sJMlxDKe2NB/+1lcOi\n4lvt85Gm6N5u5AbtjlucVF5zJN1mVTiwQZe7EjE0yp46LkUPd2TpCtNJdaZjD5VH\nf2Q1HXzn7EOJ5o89GqhDhLaR/UV3tIN+lATQt0DEHCY8LZA46M8d\n-----END CERTIFICATE-----\n", "cert": "/etc/pki/tls/certs/quadlet_demo.crt", "cert_content": "-----BEGIN CERTIFICATE-----\nMIIDgzCCAmugAwIBAgIRALqfswWVRknMu87rPCkBZVswDQYJKoZIhvcNAQELBQAw\nUDEgMB4GA1UEAwwXTG9jYWwgU2lnbmluZyBBdXRob3JpdHkxLDAqBgNVBAMMI2Jh\nOWZiMzA1LTk1NDY0OWNjLWJiY2VlYjNjLTI5MDE2NTU5MB4XDTI0MDgyNDE2MzQw\nMloXDTI1MDgyNDE2MjkwM1owFDESMBAGA1UEAxMJbG9jYWxob3N0MIIBIjANBgkq\nhkiG9w0BAQEFAAOCAQ8AMIIBCgKCAQEAkzCdkTuVp1K4qJlmXXFLmAovjmUaOQXD\nUFNT9Y2xs6TvexAGhHtKVxouqIo3uHNOeSwvtHE75pI35fx9OuZGlOlXHlPUrrjb\nnI9V7jkRYHVv/bJRqK6Nbg7uqtsEn7ApT5ZtWWYl9jc4bH88JCoddkL+//sbYpG9\nkhPhj45hDIY3Q88qm/SDugWrmCGkILhVGrT+3Diib4dZFaVY+0x2pNyd7awXemaC\nawp4PyIxHRn2nQLcnVXkHUTxg9BOq3KYS4uK/cEcrQGxEMWJdGIfRc1QLzw7/Op+\ntMX0UuFK8QpOfAbR069MuEaIySg1m2aS+Qq68FWpN1DA6ihGNB2E7QIDAQABo4GT\nMIGQMAsGA1UdDwQEAwIFoDAUBgNVHREEDTALgglsb2NhbGhvc3QwHQYDVR0lBBYw\nFAYIKwYBBQUHAwEGCCsGAQUFBwMCMAwGA1UdEwEB/wQCMAAwHQYDVR0OBBYEFIEs\nUHzt9u4H35J71lC2ryCdvkZlMB8GA1UdIwQYMBaAFHoyMtV1SaxMTUn0rn6QY05L\n3o+wMA0GCSqGSIb3DQEBCwUAA4IBAQBzUcIz8z3ASNUx92v1bW68QlvZBv+Q8rYQ\n53BX+1X9+2TCTTOumLZnIhfqeARKnBHv17/88dPeAsOU7eObWYzp8e6MRZHr/Yn9\n2W1rPjD//4DDVZyGn9vKsV3sTbGQ1f/G7e+eYcwBCIcJ93Nj9Y3ovBtpYBiY7E6z\n9HcOkqHiVJrqi/UZkKRMSr4hmOBxXoso4exqQEZSLhr0u6sJMlxDKe2NB/+1lcOi\n4lvt85Gm6N5u5AbtjlucVF5zJN1mVTiwQZe7EjE0yp46LkUPd2TpCtNJdaZjD5VH\nf2Q1HXzn7EOJ5o89GqhDhLaR/UV3tIN+lATQt0DEHCY8LZA46M8d\n-----END CERTIFICATE-----\n", "key": "/etc/pki/tls/private/quadlet_demo.key", "key_content": "-----BEGIN PRIVATE KEY-----\nMIIEvAIBADANBgkqhkiG9w0BAQEFAASCBKYwggSiAgEAAoIBAQCTMJ2RO5WnUrio\nmWZdcUuYCi+OZRo5BcNQU1P1jbGzpO97EAaEe0pXGi6oije4c055LC+0cTvmkjfl\n/H065kaU6VceU9SuuNucj1XuORFgdW/9slGoro1uDu6q2wSfsClPlm1ZZiX2Nzhs\nfzwkKh12Qv7/+xtikb2SE+GPjmEMhjdDzyqb9IO6BauYIaQguFUatP7cOKJvh1kV\npVj7THak3J3trBd6ZoJrCng/IjEdGfadAtydVeQdRPGD0E6rcphLi4r9wRytAbEQ\nxYl0Yh9FzVAvPDv86n60xfRS4UrxCk58BtHTr0y4RojJKDWbZpL5CrrwVak3UMDq\nKEY0HYTtAgMBAAECggEATAIX7C672lsiVpZYwTpHUi04X60f/IBYWvuc6LoSIPW0\nYXVmGDXertorbAdsPMj2piZGQEUVLL0nNpHLTYYqvq08usGJm+yEI/X05v77VTdv\n7K+P8a6qBEKbxwkw+xGqU5oXv8m7IsljTsvpj7db3X09V40hmCYwoeW4XjylKyS/\nmUQFO+uihgyUDw4K3trDeHMQxXVTPNCQczqIy8ypyhRhUpPtkstldlt8+WGkoj6Z\nu0CECFJJo+aDbtV1HOlEj+STNYqwpSIL4EILCAr8p1gpJdq5WSP6ZZ5smbkYVnQz\n6By27jKoUnatI9o33ACfnG2C96mlFgK5aSRPcPaZFQKBgQDD8uyORrzsLS3W3U0F\ntY3xgxKlT3r9acxcHcSHF/6jGWasRg29SbKm7qeHbUjtlpznI5Ucyq9HaTwgU98r\ntenzDXA53mxIgvmtdeFY76vwc8XlYZE12AAhkDiHuS8Iu3cn7kDgQTwe0fvTYlD2\nGYl0bhm/Rgo9Smf25KVGLwjwzwKBgQDATFOn7fEz0wRHOfiiGm0z0Hf1IwF2E7fo\nT8alpixP1h6WgThWpQjorCILMKPtb7nDtBSrYp7Z9TcX35gmacmJUFfiQD6cYtWT\nMlZwa9rnCjXzBZ1qCjfsreB4sdQxGiZlNPShZAXqgBj7zn4FcwTT1FCEiHhFwDSc\nZPnPC9XFgwKBgGV7rNRLApF3HKSJXNi2HNuWPNPtBA+2+aZDrIMjBDhIcYqLhZYh\nUyttsspK6VhThtOW9CVDO+BKOl+r0It83buJnDP8ZCxx2QnjYWCEUJJqA3Y3P9OZ\nGQhI/14bpKZrIN1HYcZDQvXbXwcthizPMQdh9me/yJHRtWKDLK3aslC1AoGACK6p\njEh6/nsFk22zoyYMTi2YbPICq+EiCZQohOwmzgf944+8CnVzHIZBFz4/+IV13rcv\nw7k1MOQYc7R9uyZFmjSNc0KxUQBJYlcAfhzwikP0wZ88R0Bk3APRwcFNslXa6/Dj\nJrJMbhUQi7X8DKTvz9yjMgizXXrbJUCN4uJG6Z8CgYAX08LBwQnXrb8vDTOlptGo\nB0u66uJ7zVzQK6lR1tt7s32OQJ2KukwZEBmVY2AcrI15mksgjU3VXV5wZkY9v+kb\njEhSIOdqdfHklxfppY5gxF6N06YJN/7I73DYaoewHUOAMUQI7qePyHJTD6clzwgd\nxKiZQl3rQ3+e9Oj5QHNiXw==\n-----END PRIVATE KEY-----\n" } } }, "changed": false } TASK [fedora.linux_system_roles.certificate : Stop tracking certificates] ****** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:176 Saturday 24 August 2024 12:34:04 -0400 (0:00:00.025) 0:00:12.867 ******* ok: [managed_node1] => (item={'cert': '/etc/pki/tls/certs/quadlet_demo.crt', 'cert_content': '-----BEGIN CERTIFICATE-----\nMIIDgzCCAmugAwIBAgIRALqfswWVRknMu87rPCkBZVswDQYJKoZIhvcNAQELBQAw\nUDEgMB4GA1UEAwwXTG9jYWwgU2lnbmluZyBBdXRob3JpdHkxLDAqBgNVBAMMI2Jh\nOWZiMzA1LTk1NDY0OWNjLWJiY2VlYjNjLTI5MDE2NTU5MB4XDTI0MDgyNDE2MzQw\nMloXDTI1MDgyNDE2MjkwM1owFDESMBAGA1UEAxMJbG9jYWxob3N0MIIBIjANBgkq\nhkiG9w0BAQEFAAOCAQ8AMIIBCgKCAQEAkzCdkTuVp1K4qJlmXXFLmAovjmUaOQXD\nUFNT9Y2xs6TvexAGhHtKVxouqIo3uHNOeSwvtHE75pI35fx9OuZGlOlXHlPUrrjb\nnI9V7jkRYHVv/bJRqK6Nbg7uqtsEn7ApT5ZtWWYl9jc4bH88JCoddkL+//sbYpG9\nkhPhj45hDIY3Q88qm/SDugWrmCGkILhVGrT+3Diib4dZFaVY+0x2pNyd7awXemaC\nawp4PyIxHRn2nQLcnVXkHUTxg9BOq3KYS4uK/cEcrQGxEMWJdGIfRc1QLzw7/Op+\ntMX0UuFK8QpOfAbR069MuEaIySg1m2aS+Qq68FWpN1DA6ihGNB2E7QIDAQABo4GT\nMIGQMAsGA1UdDwQEAwIFoDAUBgNVHREEDTALgglsb2NhbGhvc3QwHQYDVR0lBBYw\nFAYIKwYBBQUHAwEGCCsGAQUFBwMCMAwGA1UdEwEB/wQCMAAwHQYDVR0OBBYEFIEs\nUHzt9u4H35J71lC2ryCdvkZlMB8GA1UdIwQYMBaAFHoyMtV1SaxMTUn0rn6QY05L\n3o+wMA0GCSqGSIb3DQEBCwUAA4IBAQBzUcIz8z3ASNUx92v1bW68QlvZBv+Q8rYQ\n53BX+1X9+2TCTTOumLZnIhfqeARKnBHv17/88dPeAsOU7eObWYzp8e6MRZHr/Yn9\n2W1rPjD//4DDVZyGn9vKsV3sTbGQ1f/G7e+eYcwBCIcJ93Nj9Y3ovBtpYBiY7E6z\n9HcOkqHiVJrqi/UZkKRMSr4hmOBxXoso4exqQEZSLhr0u6sJMlxDKe2NB/+1lcOi\n4lvt85Gm6N5u5AbtjlucVF5zJN1mVTiwQZe7EjE0yp46LkUPd2TpCtNJdaZjD5VH\nf2Q1HXzn7EOJ5o89GqhDhLaR/UV3tIN+lATQt0DEHCY8LZA46M8d\n-----END CERTIFICATE-----\n', 'key': '/etc/pki/tls/private/quadlet_demo.key', 'key_content': '-----BEGIN PRIVATE KEY-----\nMIIEvAIBADANBgkqhkiG9w0BAQEFAASCBKYwggSiAgEAAoIBAQCTMJ2RO5WnUrio\nmWZdcUuYCi+OZRo5BcNQU1P1jbGzpO97EAaEe0pXGi6oije4c055LC+0cTvmkjfl\n/H065kaU6VceU9SuuNucj1XuORFgdW/9slGoro1uDu6q2wSfsClPlm1ZZiX2Nzhs\nfzwkKh12Qv7/+xtikb2SE+GPjmEMhjdDzyqb9IO6BauYIaQguFUatP7cOKJvh1kV\npVj7THak3J3trBd6ZoJrCng/IjEdGfadAtydVeQdRPGD0E6rcphLi4r9wRytAbEQ\nxYl0Yh9FzVAvPDv86n60xfRS4UrxCk58BtHTr0y4RojJKDWbZpL5CrrwVak3UMDq\nKEY0HYTtAgMBAAECggEATAIX7C672lsiVpZYwTpHUi04X60f/IBYWvuc6LoSIPW0\nYXVmGDXertorbAdsPMj2piZGQEUVLL0nNpHLTYYqvq08usGJm+yEI/X05v77VTdv\n7K+P8a6qBEKbxwkw+xGqU5oXv8m7IsljTsvpj7db3X09V40hmCYwoeW4XjylKyS/\nmUQFO+uihgyUDw4K3trDeHMQxXVTPNCQczqIy8ypyhRhUpPtkstldlt8+WGkoj6Z\nu0CECFJJo+aDbtV1HOlEj+STNYqwpSIL4EILCAr8p1gpJdq5WSP6ZZ5smbkYVnQz\n6By27jKoUnatI9o33ACfnG2C96mlFgK5aSRPcPaZFQKBgQDD8uyORrzsLS3W3U0F\ntY3xgxKlT3r9acxcHcSHF/6jGWasRg29SbKm7qeHbUjtlpznI5Ucyq9HaTwgU98r\ntenzDXA53mxIgvmtdeFY76vwc8XlYZE12AAhkDiHuS8Iu3cn7kDgQTwe0fvTYlD2\nGYl0bhm/Rgo9Smf25KVGLwjwzwKBgQDATFOn7fEz0wRHOfiiGm0z0Hf1IwF2E7fo\nT8alpixP1h6WgThWpQjorCILMKPtb7nDtBSrYp7Z9TcX35gmacmJUFfiQD6cYtWT\nMlZwa9rnCjXzBZ1qCjfsreB4sdQxGiZlNPShZAXqgBj7zn4FcwTT1FCEiHhFwDSc\nZPnPC9XFgwKBgGV7rNRLApF3HKSJXNi2HNuWPNPtBA+2+aZDrIMjBDhIcYqLhZYh\nUyttsspK6VhThtOW9CVDO+BKOl+r0It83buJnDP8ZCxx2QnjYWCEUJJqA3Y3P9OZ\nGQhI/14bpKZrIN1HYcZDQvXbXwcthizPMQdh9me/yJHRtWKDLK3aslC1AoGACK6p\njEh6/nsFk22zoyYMTi2YbPICq+EiCZQohOwmzgf944+8CnVzHIZBFz4/+IV13rcv\nw7k1MOQYc7R9uyZFmjSNc0KxUQBJYlcAfhzwikP0wZ88R0Bk3APRwcFNslXa6/Dj\nJrJMbhUQi7X8DKTvz9yjMgizXXrbJUCN4uJG6Z8CgYAX08LBwQnXrb8vDTOlptGo\nB0u66uJ7zVzQK6lR1tt7s32OQJ2KukwZEBmVY2AcrI15mksgjU3VXV5wZkY9v+kb\njEhSIOdqdfHklxfppY5gxF6N06YJN/7I73DYaoewHUOAMUQI7qePyHJTD6clzwgd\nxKiZQl3rQ3+e9Oj5QHNiXw==\n-----END PRIVATE KEY-----\n', 'ca': '/etc/pki/tls/certs/quadlet_demo.crt', 'ca_content': '-----BEGIN CERTIFICATE-----\nMIIDgzCCAmugAwIBAgIRALqfswWVRknMu87rPCkBZVswDQYJKoZIhvcNAQELBQAw\nUDEgMB4GA1UEAwwXTG9jYWwgU2lnbmluZyBBdXRob3JpdHkxLDAqBgNVBAMMI2Jh\nOWZiMzA1LTk1NDY0OWNjLWJiY2VlYjNjLTI5MDE2NTU5MB4XDTI0MDgyNDE2MzQw\nMloXDTI1MDgyNDE2MjkwM1owFDESMBAGA1UEAxMJbG9jYWxob3N0MIIBIjANBgkq\nhkiG9w0BAQEFAAOCAQ8AMIIBCgKCAQEAkzCdkTuVp1K4qJlmXXFLmAovjmUaOQXD\nUFNT9Y2xs6TvexAGhHtKVxouqIo3uHNOeSwvtHE75pI35fx9OuZGlOlXHlPUrrjb\nnI9V7jkRYHVv/bJRqK6Nbg7uqtsEn7ApT5ZtWWYl9jc4bH88JCoddkL+//sbYpG9\nkhPhj45hDIY3Q88qm/SDugWrmCGkILhVGrT+3Diib4dZFaVY+0x2pNyd7awXemaC\nawp4PyIxHRn2nQLcnVXkHUTxg9BOq3KYS4uK/cEcrQGxEMWJdGIfRc1QLzw7/Op+\ntMX0UuFK8QpOfAbR069MuEaIySg1m2aS+Qq68FWpN1DA6ihGNB2E7QIDAQABo4GT\nMIGQMAsGA1UdDwQEAwIFoDAUBgNVHREEDTALgglsb2NhbGhvc3QwHQYDVR0lBBYw\nFAYIKwYBBQUHAwEGCCsGAQUFBwMCMAwGA1UdEwEB/wQCMAAwHQYDVR0OBBYEFIEs\nUHzt9u4H35J71lC2ryCdvkZlMB8GA1UdIwQYMBaAFHoyMtV1SaxMTUn0rn6QY05L\n3o+wMA0GCSqGSIb3DQEBCwUAA4IBAQBzUcIz8z3ASNUx92v1bW68QlvZBv+Q8rYQ\n53BX+1X9+2TCTTOumLZnIhfqeARKnBHv17/88dPeAsOU7eObWYzp8e6MRZHr/Yn9\n2W1rPjD//4DDVZyGn9vKsV3sTbGQ1f/G7e+eYcwBCIcJ93Nj9Y3ovBtpYBiY7E6z\n9HcOkqHiVJrqi/UZkKRMSr4hmOBxXoso4exqQEZSLhr0u6sJMlxDKe2NB/+1lcOi\n4lvt85Gm6N5u5AbtjlucVF5zJN1mVTiwQZe7EjE0yp46LkUPd2TpCtNJdaZjD5VH\nf2Q1HXzn7EOJ5o89GqhDhLaR/UV3tIN+lATQt0DEHCY8LZA46M8d\n-----END CERTIFICATE-----\n'}) => { "ansible_loop_var": "item", "changed": false, "cmd": [ "getcert", "stop-tracking", "-f", "/etc/pki/tls/certs/quadlet_demo.crt" ], "delta": "0:00:00.030775", "end": "2024-08-24 12:34:04.757078", "item": { "ca": "/etc/pki/tls/certs/quadlet_demo.crt", "ca_content": "-----BEGIN CERTIFICATE-----\nMIIDgzCCAmugAwIBAgIRALqfswWVRknMu87rPCkBZVswDQYJKoZIhvcNAQELBQAw\nUDEgMB4GA1UEAwwXTG9jYWwgU2lnbmluZyBBdXRob3JpdHkxLDAqBgNVBAMMI2Jh\nOWZiMzA1LTk1NDY0OWNjLWJiY2VlYjNjLTI5MDE2NTU5MB4XDTI0MDgyNDE2MzQw\nMloXDTI1MDgyNDE2MjkwM1owFDESMBAGA1UEAxMJbG9jYWxob3N0MIIBIjANBgkq\nhkiG9w0BAQEFAAOCAQ8AMIIBCgKCAQEAkzCdkTuVp1K4qJlmXXFLmAovjmUaOQXD\nUFNT9Y2xs6TvexAGhHtKVxouqIo3uHNOeSwvtHE75pI35fx9OuZGlOlXHlPUrrjb\nnI9V7jkRYHVv/bJRqK6Nbg7uqtsEn7ApT5ZtWWYl9jc4bH88JCoddkL+//sbYpG9\nkhPhj45hDIY3Q88qm/SDugWrmCGkILhVGrT+3Diib4dZFaVY+0x2pNyd7awXemaC\nawp4PyIxHRn2nQLcnVXkHUTxg9BOq3KYS4uK/cEcrQGxEMWJdGIfRc1QLzw7/Op+\ntMX0UuFK8QpOfAbR069MuEaIySg1m2aS+Qq68FWpN1DA6ihGNB2E7QIDAQABo4GT\nMIGQMAsGA1UdDwQEAwIFoDAUBgNVHREEDTALgglsb2NhbGhvc3QwHQYDVR0lBBYw\nFAYIKwYBBQUHAwEGCCsGAQUFBwMCMAwGA1UdEwEB/wQCMAAwHQYDVR0OBBYEFIEs\nUHzt9u4H35J71lC2ryCdvkZlMB8GA1UdIwQYMBaAFHoyMtV1SaxMTUn0rn6QY05L\n3o+wMA0GCSqGSIb3DQEBCwUAA4IBAQBzUcIz8z3ASNUx92v1bW68QlvZBv+Q8rYQ\n53BX+1X9+2TCTTOumLZnIhfqeARKnBHv17/88dPeAsOU7eObWYzp8e6MRZHr/Yn9\n2W1rPjD//4DDVZyGn9vKsV3sTbGQ1f/G7e+eYcwBCIcJ93Nj9Y3ovBtpYBiY7E6z\n9HcOkqHiVJrqi/UZkKRMSr4hmOBxXoso4exqQEZSLhr0u6sJMlxDKe2NB/+1lcOi\n4lvt85Gm6N5u5AbtjlucVF5zJN1mVTiwQZe7EjE0yp46LkUPd2TpCtNJdaZjD5VH\nf2Q1HXzn7EOJ5o89GqhDhLaR/UV3tIN+lATQt0DEHCY8LZA46M8d\n-----END CERTIFICATE-----\n", "cert": "/etc/pki/tls/certs/quadlet_demo.crt", "cert_content": "-----BEGIN CERTIFICATE-----\nMIIDgzCCAmugAwIBAgIRALqfswWVRknMu87rPCkBZVswDQYJKoZIhvcNAQELBQAw\nUDEgMB4GA1UEAwwXTG9jYWwgU2lnbmluZyBBdXRob3JpdHkxLDAqBgNVBAMMI2Jh\nOWZiMzA1LTk1NDY0OWNjLWJiY2VlYjNjLTI5MDE2NTU5MB4XDTI0MDgyNDE2MzQw\nMloXDTI1MDgyNDE2MjkwM1owFDESMBAGA1UEAxMJbG9jYWxob3N0MIIBIjANBgkq\nhkiG9w0BAQEFAAOCAQ8AMIIBCgKCAQEAkzCdkTuVp1K4qJlmXXFLmAovjmUaOQXD\nUFNT9Y2xs6TvexAGhHtKVxouqIo3uHNOeSwvtHE75pI35fx9OuZGlOlXHlPUrrjb\nnI9V7jkRYHVv/bJRqK6Nbg7uqtsEn7ApT5ZtWWYl9jc4bH88JCoddkL+//sbYpG9\nkhPhj45hDIY3Q88qm/SDugWrmCGkILhVGrT+3Diib4dZFaVY+0x2pNyd7awXemaC\nawp4PyIxHRn2nQLcnVXkHUTxg9BOq3KYS4uK/cEcrQGxEMWJdGIfRc1QLzw7/Op+\ntMX0UuFK8QpOfAbR069MuEaIySg1m2aS+Qq68FWpN1DA6ihGNB2E7QIDAQABo4GT\nMIGQMAsGA1UdDwQEAwIFoDAUBgNVHREEDTALgglsb2NhbGhvc3QwHQYDVR0lBBYw\nFAYIKwYBBQUHAwEGCCsGAQUFBwMCMAwGA1UdEwEB/wQCMAAwHQYDVR0OBBYEFIEs\nUHzt9u4H35J71lC2ryCdvkZlMB8GA1UdIwQYMBaAFHoyMtV1SaxMTUn0rn6QY05L\n3o+wMA0GCSqGSIb3DQEBCwUAA4IBAQBzUcIz8z3ASNUx92v1bW68QlvZBv+Q8rYQ\n53BX+1X9+2TCTTOumLZnIhfqeARKnBHv17/88dPeAsOU7eObWYzp8e6MRZHr/Yn9\n2W1rPjD//4DDVZyGn9vKsV3sTbGQ1f/G7e+eYcwBCIcJ93Nj9Y3ovBtpYBiY7E6z\n9HcOkqHiVJrqi/UZkKRMSr4hmOBxXoso4exqQEZSLhr0u6sJMlxDKe2NB/+1lcOi\n4lvt85Gm6N5u5AbtjlucVF5zJN1mVTiwQZe7EjE0yp46LkUPd2TpCtNJdaZjD5VH\nf2Q1HXzn7EOJ5o89GqhDhLaR/UV3tIN+lATQt0DEHCY8LZA46M8d\n-----END CERTIFICATE-----\n", "key": "/etc/pki/tls/private/quadlet_demo.key", "key_content": "-----BEGIN PRIVATE KEY-----\nMIIEvAIBADANBgkqhkiG9w0BAQEFAASCBKYwggSiAgEAAoIBAQCTMJ2RO5WnUrio\nmWZdcUuYCi+OZRo5BcNQU1P1jbGzpO97EAaEe0pXGi6oije4c055LC+0cTvmkjfl\n/H065kaU6VceU9SuuNucj1XuORFgdW/9slGoro1uDu6q2wSfsClPlm1ZZiX2Nzhs\nfzwkKh12Qv7/+xtikb2SE+GPjmEMhjdDzyqb9IO6BauYIaQguFUatP7cOKJvh1kV\npVj7THak3J3trBd6ZoJrCng/IjEdGfadAtydVeQdRPGD0E6rcphLi4r9wRytAbEQ\nxYl0Yh9FzVAvPDv86n60xfRS4UrxCk58BtHTr0y4RojJKDWbZpL5CrrwVak3UMDq\nKEY0HYTtAgMBAAECggEATAIX7C672lsiVpZYwTpHUi04X60f/IBYWvuc6LoSIPW0\nYXVmGDXertorbAdsPMj2piZGQEUVLL0nNpHLTYYqvq08usGJm+yEI/X05v77VTdv\n7K+P8a6qBEKbxwkw+xGqU5oXv8m7IsljTsvpj7db3X09V40hmCYwoeW4XjylKyS/\nmUQFO+uihgyUDw4K3trDeHMQxXVTPNCQczqIy8ypyhRhUpPtkstldlt8+WGkoj6Z\nu0CECFJJo+aDbtV1HOlEj+STNYqwpSIL4EILCAr8p1gpJdq5WSP6ZZ5smbkYVnQz\n6By27jKoUnatI9o33ACfnG2C96mlFgK5aSRPcPaZFQKBgQDD8uyORrzsLS3W3U0F\ntY3xgxKlT3r9acxcHcSHF/6jGWasRg29SbKm7qeHbUjtlpznI5Ucyq9HaTwgU98r\ntenzDXA53mxIgvmtdeFY76vwc8XlYZE12AAhkDiHuS8Iu3cn7kDgQTwe0fvTYlD2\nGYl0bhm/Rgo9Smf25KVGLwjwzwKBgQDATFOn7fEz0wRHOfiiGm0z0Hf1IwF2E7fo\nT8alpixP1h6WgThWpQjorCILMKPtb7nDtBSrYp7Z9TcX35gmacmJUFfiQD6cYtWT\nMlZwa9rnCjXzBZ1qCjfsreB4sdQxGiZlNPShZAXqgBj7zn4FcwTT1FCEiHhFwDSc\nZPnPC9XFgwKBgGV7rNRLApF3HKSJXNi2HNuWPNPtBA+2+aZDrIMjBDhIcYqLhZYh\nUyttsspK6VhThtOW9CVDO+BKOl+r0It83buJnDP8ZCxx2QnjYWCEUJJqA3Y3P9OZ\nGQhI/14bpKZrIN1HYcZDQvXbXwcthizPMQdh9me/yJHRtWKDLK3aslC1AoGACK6p\njEh6/nsFk22zoyYMTi2YbPICq+EiCZQohOwmzgf944+8CnVzHIZBFz4/+IV13rcv\nw7k1MOQYc7R9uyZFmjSNc0KxUQBJYlcAfhzwikP0wZ88R0Bk3APRwcFNslXa6/Dj\nJrJMbhUQi7X8DKTvz9yjMgizXXrbJUCN4uJG6Z8CgYAX08LBwQnXrb8vDTOlptGo\nB0u66uJ7zVzQK6lR1tt7s32OQJ2KukwZEBmVY2AcrI15mksgjU3VXV5wZkY9v+kb\njEhSIOdqdfHklxfppY5gxF6N06YJN/7I73DYaoewHUOAMUQI7qePyHJTD6clzwgd\nxKiZQl3rQ3+e9Oj5QHNiXw==\n-----END PRIVATE KEY-----\n" }, "rc": 0, "start": "2024-08-24 12:34:04.726303" } STDOUT: Request "20240824163402" removed. TASK [fedora.linux_system_roles.certificate : Remove files] ******************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:181 Saturday 24 August 2024 12:34:04 -0400 (0:00:00.562) 0:00:13.429 ******* changed: [managed_node1] => (item=/etc/pki/tls/certs/quadlet_demo.crt) => { "ansible_loop_var": "item", "changed": true, "item": "/etc/pki/tls/certs/quadlet_demo.crt", "path": "/etc/pki/tls/certs/quadlet_demo.crt", "state": "absent" } changed: [managed_node1] => (item=/etc/pki/tls/private/quadlet_demo.key) => { "ansible_loop_var": "item", "changed": true, "item": "/etc/pki/tls/private/quadlet_demo.key", "path": "/etc/pki/tls/private/quadlet_demo.key", "state": "absent" } ok: [managed_node1] => (item=/etc/pki/tls/certs/quadlet_demo.crt) => { "ansible_loop_var": "item", "changed": false, "item": "/etc/pki/tls/certs/quadlet_demo.crt", "path": "/etc/pki/tls/certs/quadlet_demo.crt", "state": "absent" } TASK [Run the role] ************************************************************ task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_demo.yml:62 Saturday 24 August 2024 12:34:06 -0400 (0:00:01.714) 0:00:15.143 ******* TASK [fedora.linux_system_roles.podman : Set platform/version specific variables] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:3 Saturday 24 August 2024 12:34:06 -0400 (0:00:00.117) 0:00:15.260 ******* included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Ensure ansible_facts used by role] **** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:3 Saturday 24 August 2024 12:34:06 -0400 (0:00:00.029) 0:00:15.290 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if system is ostree] ************ task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:11 Saturday 24 August 2024 12:34:06 -0400 (0:00:00.023) 0:00:15.314 ******* ok: [managed_node1] => { "changed": false, "stat": { "exists": false } } TASK [fedora.linux_system_roles.podman : Set flag to indicate system is ostree] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:16 Saturday 24 August 2024 12:34:07 -0400 (0:00:00.419) 0:00:15.733 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_is_ostree": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Check if transactional-update exists in /sbin] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:23 Saturday 24 August 2024 12:34:07 -0400 (0:00:00.031) 0:00:15.764 ******* ok: [managed_node1] => { "changed": false, "stat": { "exists": false } } TASK [fedora.linux_system_roles.podman : Set flag if transactional-update exists] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:28 Saturday 24 August 2024 12:34:07 -0400 (0:00:00.376) 0:00:16.141 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_is_transactional": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Set platform/version specific variables] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:32 Saturday 24 August 2024 12:34:07 -0400 (0:00:00.052) 0:00:16.193 ******* ok: [managed_node1] => (item=RedHat.yml) => { "ansible_facts": { "__podman_packages": [ "podman", "shadow-utils-subid" ] }, "ansible_included_var_files": [ "/tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/vars/RedHat.yml" ], "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml" } skipping: [managed_node1] => (item=CentOS.yml) => { "ansible_loop_var": "item", "changed": false, "item": "CentOS.yml", "skip_reason": "Conditional result was False" } ok: [managed_node1] => (item=CentOS_8.yml) => { "ansible_facts": { "__podman_packages": [ "crun", "podman", "podman-plugins", "shadow-utils-subid" ] }, "ansible_included_var_files": [ "/tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/vars/CentOS_8.yml" ], "ansible_loop_var": "item", "changed": false, "item": "CentOS_8.yml" } ok: [managed_node1] => (item=CentOS_8.yml) => { "ansible_facts": { "__podman_packages": [ "crun", "podman", "podman-plugins", "shadow-utils-subid" ] }, "ansible_included_var_files": [ "/tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/vars/CentOS_8.yml" ], "ansible_loop_var": "item", "changed": false, "item": "CentOS_8.yml" } TASK [fedora.linux_system_roles.podman : Gather the package facts] ************* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:6 Saturday 24 August 2024 12:34:07 -0400 (0:00:00.065) 0:00:16.258 ******* ok: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Enable copr if requested] ************* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:10 Saturday 24 August 2024 12:34:09 -0400 (0:00:01.878) 0:00:18.137 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Ensure required packages are installed] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:14 Saturday 24 August 2024 12:34:09 -0400 (0:00:00.050) 0:00:18.188 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Notify user that reboot is needed to apply changes] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:28 Saturday 24 August 2024 12:34:09 -0400 (0:00:00.081) 0:00:18.270 ******* skipping: [managed_node1] => {} TASK [fedora.linux_system_roles.podman : Reboot transactional update systems] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:33 Saturday 24 August 2024 12:34:09 -0400 (0:00:00.094) 0:00:18.364 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if reboot is needed and not set] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:38 Saturday 24 August 2024 12:34:09 -0400 (0:00:00.089) 0:00:18.453 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get podman version] ******************* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:46 Saturday 24 August 2024 12:34:09 -0400 (0:00:00.073) 0:00:18.527 ******* ok: [managed_node1] => { "changed": false, "cmd": [ "podman", "--version" ], "delta": "0:00:00.027889", "end": "2024-08-24 12:34:10.333993", "rc": 0, "start": "2024-08-24 12:34:10.306104" } STDOUT: podman version 4.9.4-dev TASK [fedora.linux_system_roles.podman : Set podman version] ******************* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:52 Saturday 24 August 2024 12:34:10 -0400 (0:00:00.494) 0:00:19.021 ******* ok: [managed_node1] => { "ansible_facts": { "podman_version": "4.9.4-dev" }, "changed": false } TASK [fedora.linux_system_roles.podman : Podman package version must be 4.2 or later] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:56 Saturday 24 August 2024 12:34:10 -0400 (0:00:00.072) 0:00:19.094 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Podman package version must be 4.4 or later for quadlet, secrets] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:63 Saturday 24 August 2024 12:34:10 -0400 (0:00:00.051) 0:00:19.145 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } META: end_host conditional evaluated to false, continuing execution for managed_node1 TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:80 Saturday 24 August 2024 12:34:10 -0400 (0:00:00.264) 0:00:19.410 ******* included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Saturday 24 August 2024 12:34:10 -0400 (0:00:00.087) 0:00:19.497 ******* ok: [managed_node1] => { "ansible_facts": { "getent_passwd": { "root": [ "x", "0", "0", "root", "/root", "/bin/bash" ] } }, "changed": false } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Saturday 24 August 2024 12:34:11 -0400 (0:00:00.848) 0:00:20.346 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Saturday 24 August 2024 12:34:11 -0400 (0:00:00.110) 0:00:20.456 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get group information] **************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:28 Saturday 24 August 2024 12:34:11 -0400 (0:00:00.112) 0:00:20.569 ******* ok: [managed_node1] => { "ansible_facts": { "getent_group": { "root": [ "x", "0", "" ] } }, "changed": false } TASK [fedora.linux_system_roles.podman : Set group name] *********************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:35 Saturday 24 August 2024 12:34:12 -0400 (0:00:00.670) 0:00:21.239 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_group_name": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:39 Saturday 24 August 2024 12:34:12 -0400 (0:00:00.066) 0:00:21.305 ******* ok: [managed_node1] => { "changed": false, "stat": { "atime": 1724516929.7168462, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "bb5b46ffbafcaa8c4021f3c8b3cb8594f48ef34b", "ctime": 1724516896.725572, "dev": 51713, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 6986657, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-sharedlib", "mode": "0755", "mtime": 1700557386.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 12640, "uid": 0, "version": "1255679238", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************ task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:50 Saturday 24 August 2024 12:34:13 -0400 (0:00:00.469) 0:00:21.774 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check group with getsubids] *********** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:55 Saturday 24 August 2024 12:34:13 -0400 (0:00:00.037) 0:00:21.812 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:60 Saturday 24 August 2024 12:34:13 -0400 (0:00:00.044) 0:00:21.856 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subuid file] ********************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:74 Saturday 24 August 2024 12:34:13 -0400 (0:00:00.044) 0:00:21.901 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subgid file] ********************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:79 Saturday 24 August 2024 12:34:13 -0400 (0:00:00.035) 0:00:21.936 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:84 Saturday 24 August 2024 12:34:13 -0400 (0:00:00.032) 0:00:21.968 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:94 Saturday 24 August 2024 12:34:13 -0400 (0:00:00.031) 0:00:21.999 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] ***** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:101 Saturday 24 August 2024 12:34:13 -0400 (0:00:00.034) 0:00:22.033 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set config file paths] **************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:86 Saturday 24 August 2024 12:34:13 -0400 (0:00:00.034) 0:00:22.067 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_container_conf_file": "/etc/containers/containers.conf.d/50-systemroles.conf", "__podman_policy_json_file": "/etc/containers/policy.json", "__podman_registries_conf_file": "/etc/containers/registries.conf.d/50-systemroles.conf", "__podman_storage_conf_file": "/etc/containers/storage.conf" }, "changed": false } TASK [fedora.linux_system_roles.podman : Handle container.conf.d] ************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:95 Saturday 24 August 2024 12:34:13 -0400 (0:00:00.067) 0:00:22.135 ******* included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_container_conf_d.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Ensure containers.d exists] *********** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_container_conf_d.yml:5 Saturday 24 August 2024 12:34:13 -0400 (0:00:00.062) 0:00:22.197 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Update container config file] ********* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_container_conf_d.yml:13 Saturday 24 August 2024 12:34:13 -0400 (0:00:00.037) 0:00:22.235 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Handle registries.conf.d] ************* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:98 Saturday 24 August 2024 12:34:13 -0400 (0:00:00.052) 0:00:22.287 ******* included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_registries_conf_d.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Ensure registries.d exists] *********** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_registries_conf_d.yml:5 Saturday 24 August 2024 12:34:13 -0400 (0:00:00.128) 0:00:22.416 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Update registries config file] ******** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_registries_conf_d.yml:13 Saturday 24 August 2024 12:34:13 -0400 (0:00:00.041) 0:00:22.458 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Handle storage.conf] ****************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:101 Saturday 24 August 2024 12:34:13 -0400 (0:00:00.040) 0:00:22.498 ******* included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_storage_conf.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Ensure storage.conf parent dir exists] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_storage_conf.yml:5 Saturday 24 August 2024 12:34:13 -0400 (0:00:00.062) 0:00:22.561 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Update storage config file] *********** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_storage_conf.yml:13 Saturday 24 August 2024 12:34:13 -0400 (0:00:00.032) 0:00:22.594 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Handle policy.json] ******************* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:104 Saturday 24 August 2024 12:34:14 -0400 (0:00:00.033) 0:00:22.628 ******* included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Ensure policy.json parent dir exists] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:6 Saturday 24 August 2024 12:34:14 -0400 (0:00:00.078) 0:00:22.706 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Stat the policy.json file] ************ task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:14 Saturday 24 August 2024 12:34:14 -0400 (0:00:00.051) 0:00:22.758 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get the existing policy.json] ********* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:19 Saturday 24 August 2024 12:34:14 -0400 (0:00:00.049) 0:00:22.808 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Write new policy.json file] *********** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:25 Saturday 24 August 2024 12:34:14 -0400 (0:00:00.049) 0:00:22.857 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Manage firewall for specified ports] ************************************* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:110 Saturday 24 August 2024 12:34:14 -0400 (0:00:00.038) 0:00:22.895 ******* TASK [fedora.linux_system_roles.firewall : Setup firewalld] ******************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:2 Saturday 24 August 2024 12:34:14 -0400 (0:00:00.117) 0:00:23.012 ******* included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml for managed_node1 TASK [fedora.linux_system_roles.firewall : Ensure ansible_facts used by role] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:2 Saturday 24 August 2024 12:34:14 -0400 (0:00:00.059) 0:00:23.072 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Check if system is ostree] ********** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:10 Saturday 24 August 2024 12:34:14 -0400 (0:00:00.111) 0:00:23.184 ******* ok: [managed_node1] => { "changed": false, "stat": { "exists": false } } TASK [fedora.linux_system_roles.firewall : Set flag to indicate system is ostree] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:15 Saturday 24 August 2024 12:34:14 -0400 (0:00:00.430) 0:00:23.614 ******* ok: [managed_node1] => { "ansible_facts": { "__firewall_is_ostree": false }, "changed": false } TASK [fedora.linux_system_roles.firewall : Check if transactional-update exists in /sbin] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:22 Saturday 24 August 2024 12:34:15 -0400 (0:00:00.062) 0:00:23.676 ******* ok: [managed_node1] => { "changed": false, "stat": { "exists": false } } TASK [fedora.linux_system_roles.firewall : Set flag if transactional-update exists] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:27 Saturday 24 August 2024 12:34:15 -0400 (0:00:00.428) 0:00:24.105 ******* ok: [managed_node1] => { "ansible_facts": { "__firewall_is_transactional": false }, "changed": false } TASK [fedora.linux_system_roles.firewall : Install firewalld] ****************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:31 Saturday 24 August 2024 12:34:15 -0400 (0:00:00.060) 0:00:24.166 ******* ok: [managed_node1] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do lsrpackages: firewalld TASK [fedora.linux_system_roles.firewall : Notify user that reboot is needed to apply changes] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:43 Saturday 24 August 2024 12:34:18 -0400 (0:00:02.791) 0:00:26.957 ******* skipping: [managed_node1] => {} TASK [fedora.linux_system_roles.firewall : Reboot transactional update systems] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:48 Saturday 24 August 2024 12:34:18 -0400 (0:00:00.066) 0:00:27.024 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Fail if reboot is needed and not set] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:53 Saturday 24 August 2024 12:34:18 -0400 (0:00:00.033) 0:00:27.057 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Collect service facts] ************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:5 Saturday 24 August 2024 12:34:18 -0400 (0:00:00.034) 0:00:27.092 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Attempt to stop and disable conflicting services] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:9 Saturday 24 August 2024 12:34:18 -0400 (0:00:00.038) 0:00:27.131 ******* skipping: [managed_node1] => (item=nftables) => { "ansible_loop_var": "item", "changed": false, "item": "nftables", "skip_reason": "Conditional result was False" } skipping: [managed_node1] => (item=iptables) => { "ansible_loop_var": "item", "changed": false, "item": "iptables", "skip_reason": "Conditional result was False" } skipping: [managed_node1] => (item=ufw) => { "ansible_loop_var": "item", "changed": false, "item": "ufw", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Unmask firewalld service] *********** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:22 Saturday 24 August 2024 12:34:18 -0400 (0:00:00.069) 0:00:27.200 ******* ok: [managed_node1] => { "changed": false, "name": "firewalld", "status": { "ActiveEnterTimestampMonotonic": "0", "ActiveExitTimestampMonotonic": "0", "ActiveState": "inactive", "After": "polkit.service system.slice dbus.service basic.target dbus.socket sysinit.target", "AllowIsolate": "no", "AllowedCPUs": "", "AllowedMemoryNodes": "", "AmbientCapabilities": "", "AssertResult": "no", "AssertTimestampMonotonic": "0", "Before": "shutdown.target network-pre.target", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "BusName": "org.fedoraproject.FirewallD1", "CPUAccounting": "no", "CPUAffinity": "", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "[not set]", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "yes", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf", "CollectMode": "inactive", "ConditionResult": "no", "ConditionTimestampMonotonic": "0", "ConfigurationDirectoryMode": "0755", "Conflicts": "shutdown.target ebtables.service iptables.service ip6tables.service ipset.service nftables.service", "ControlPID": "0", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "Delegate": "no", "Description": "firewalld - dynamic firewall daemon", "DevicePolicy": "auto", "Documentation": "man:firewalld(1)", "DynamicUser": "no", "EffectiveCPUs": "", "EffectiveMemoryNodes": "", "EnvironmentFiles": "/etc/sysconfig/firewalld (ignore_errors=yes)", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "0", "ExecMainStartTimestampMonotonic": "0", "ExecMainStatus": "0", "ExecReload": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStart": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FragmentPath": "/usr/lib/systemd/system/firewalld.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOSchedulingClass": "0", "IOSchedulingPriority": "0", "IOWeight": "[not set]", "IPAccounting": "no", "IPEgressBytes": "18446744073709551615", "IPEgressPackets": "18446744073709551615", "IPIngressBytes": "18446744073709551615", "IPIngressPackets": "18446744073709551615", "Id": "firewalld.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestampMonotonic": "0", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "mixed", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "0", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "65536", "LimitMEMLOCKSoft": "65536", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "262144", "LimitNOFILESoft": "1024", "LimitNPROC": "14003", "LimitNPROCSoft": "14003", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "14003", "LimitSIGPENDINGSoft": "14003", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "0", "MemoryAccounting": "yes", "MemoryCurrent": "[not set]", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemorySwapMax": "infinity", "MountAPIVFS": "no", "MountFlags": "", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAMask": "", "NUMAPolicy": "n/a", "Names": "firewalld.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "PermissionsStartOnly": "no", "Perpetual": "no", "PrivateDevices": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "RemainAfterExit": "no", "RemoveIPC": "no", "Requires": "system.slice sysinit.target dbus.socket", "Restart": "no", "RestartUSec": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "Slice": "system.slice", "StandardError": "null", "StandardInput": "null", "StandardInputData": "", "StandardOutput": "null", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StateChangeTimestampMonotonic": "0", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "dead", "SuccessAction": "none", "SyslogFacility": "3", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "0", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "[not set]", "TasksMax": "22405", "TimeoutStartUSec": "1min 30s", "TimeoutStopUSec": "1min 30s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "dbus", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "enabled", "UnitFileState": "disabled", "UtmpMode": "init", "Wants": "network-pre.target", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "0" } } TASK [fedora.linux_system_roles.firewall : Enable and start firewalld service] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:28 Saturday 24 August 2024 12:34:19 -0400 (0:00:00.592) 0:00:27.793 ******* changed: [managed_node1] => { "changed": true, "enabled": true, "name": "firewalld", "state": "started", "status": { "ActiveEnterTimestampMonotonic": "0", "ActiveExitTimestampMonotonic": "0", "ActiveState": "inactive", "After": "polkit.service sysinit.target system.slice basic.target dbus.service dbus.socket", "AllowIsolate": "no", "AllowedCPUs": "", "AllowedMemoryNodes": "", "AmbientCapabilities": "", "AssertResult": "no", "AssertTimestampMonotonic": "0", "Before": "shutdown.target network-pre.target", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "BusName": "org.fedoraproject.FirewallD1", "CPUAccounting": "no", "CPUAffinity": "", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "[not set]", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "yes", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf", "CollectMode": "inactive", "ConditionResult": "no", "ConditionTimestampMonotonic": "0", "ConfigurationDirectoryMode": "0755", "Conflicts": "shutdown.target iptables.service ebtables.service ip6tables.service nftables.service ipset.service", "ControlPID": "0", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "Delegate": "no", "Description": "firewalld - dynamic firewall daemon", "DevicePolicy": "auto", "Documentation": "man:firewalld(1)", "DynamicUser": "no", "EffectiveCPUs": "", "EffectiveMemoryNodes": "", "EnvironmentFiles": "/etc/sysconfig/firewalld (ignore_errors=yes)", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "0", "ExecMainStartTimestampMonotonic": "0", "ExecMainStatus": "0", "ExecReload": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStart": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FragmentPath": "/usr/lib/systemd/system/firewalld.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOSchedulingClass": "0", "IOSchedulingPriority": "0", "IOWeight": "[not set]", "IPAccounting": "no", "IPEgressBytes": "18446744073709551615", "IPEgressPackets": "18446744073709551615", "IPIngressBytes": "18446744073709551615", "IPIngressPackets": "18446744073709551615", "Id": "firewalld.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestampMonotonic": "0", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "mixed", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "0", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "65536", "LimitMEMLOCKSoft": "65536", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "262144", "LimitNOFILESoft": "1024", "LimitNPROC": "14003", "LimitNPROCSoft": "14003", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "14003", "LimitSIGPENDINGSoft": "14003", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "0", "MemoryAccounting": "yes", "MemoryCurrent": "[not set]", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemorySwapMax": "infinity", "MountAPIVFS": "no", "MountFlags": "", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAMask": "", "NUMAPolicy": "n/a", "Names": "firewalld.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "PermissionsStartOnly": "no", "Perpetual": "no", "PrivateDevices": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "RemainAfterExit": "no", "RemoveIPC": "no", "Requires": "dbus.socket system.slice sysinit.target", "Restart": "no", "RestartUSec": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "Slice": "system.slice", "StandardError": "null", "StandardInput": "null", "StandardInputData": "", "StandardOutput": "null", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StateChangeTimestampMonotonic": "0", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "dead", "SuccessAction": "none", "SyslogFacility": "3", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "0", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "[not set]", "TasksMax": "22405", "TimeoutStartUSec": "1min 30s", "TimeoutStopUSec": "1min 30s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "dbus", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "enabled", "UnitFileState": "disabled", "UtmpMode": "init", "Wants": "network-pre.target", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "0" } } TASK [fedora.linux_system_roles.firewall : Check if previous replaced is defined] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:34 Saturday 24 August 2024 12:34:20 -0400 (0:00:01.206) 0:00:28.999 ******* ok: [managed_node1] => { "ansible_facts": { "__firewall_previous_replaced": false, "__firewall_python_cmd": "/usr/libexec/platform-python", "__firewall_report_changed": true }, "changed": false } TASK [fedora.linux_system_roles.firewall : Get config files, checksums before and remove] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:43 Saturday 24 August 2024 12:34:20 -0400 (0:00:00.049) 0:00:29.049 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Tell firewall module it is able to report changed] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:55 Saturday 24 August 2024 12:34:20 -0400 (0:00:00.054) 0:00:29.104 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Configure firewall] ***************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:71 Saturday 24 August 2024 12:34:20 -0400 (0:00:00.057) 0:00:29.162 ******* changed: [managed_node1] => (item={'port': '8000/tcp', 'state': 'enabled'}) => { "__firewall_changed": true, "ansible_loop_var": "item", "changed": true, "item": { "port": "8000/tcp", "state": "enabled" } } changed: [managed_node1] => (item={'port': '9000/tcp', 'state': 'enabled'}) => { "__firewall_changed": true, "ansible_loop_var": "item", "changed": true, "item": { "port": "9000/tcp", "state": "enabled" } } TASK [fedora.linux_system_roles.firewall : Gather firewall config information] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:120 Saturday 24 August 2024 12:34:22 -0400 (0:00:01.498) 0:00:30.660 ******* skipping: [managed_node1] => (item={'port': '8000/tcp', 'state': 'enabled'}) => { "ansible_loop_var": "item", "changed": false, "item": { "port": "8000/tcp", "state": "enabled" }, "skip_reason": "Conditional result was False" } skipping: [managed_node1] => (item={'port': '9000/tcp', 'state': 'enabled'}) => { "ansible_loop_var": "item", "changed": false, "item": { "port": "9000/tcp", "state": "enabled" }, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Update firewalld_config fact] ******* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:130 Saturday 24 August 2024 12:34:22 -0400 (0:00:00.090) 0:00:30.750 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Gather firewall config if no arguments] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:139 Saturday 24 August 2024 12:34:22 -0400 (0:00:00.060) 0:00:30.811 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Update firewalld_config fact] ******* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:144 Saturday 24 August 2024 12:34:22 -0400 (0:00:00.054) 0:00:30.865 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Get config files, checksums after] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:153 Saturday 24 August 2024 12:34:22 -0400 (0:00:00.057) 0:00:30.922 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Calculate what has changed] ********* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:163 Saturday 24 August 2024 12:34:22 -0400 (0:00:00.053) 0:00:30.976 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Show diffs] ************************* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:169 Saturday 24 August 2024 12:34:22 -0400 (0:00:00.050) 0:00:31.026 ******* skipping: [managed_node1] => {} TASK [Manage selinux for specified ports] ************************************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:117 Saturday 24 August 2024 12:34:22 -0400 (0:00:00.053) 0:00:31.080 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Keep track of users that need to cancel linger] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:124 Saturday 24 August 2024 12:34:22 -0400 (0:00:00.052) 0:00:31.132 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_cancel_user_linger": [] }, "changed": false } TASK [fedora.linux_system_roles.podman : Handle certs.d files - present] ******* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:128 Saturday 24 August 2024 12:34:22 -0400 (0:00:00.049) 0:00:31.182 ******* skipping: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Handle credential files - present] **** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:137 Saturday 24 August 2024 12:34:22 -0400 (0:00:00.034) 0:00:31.217 ******* skipping: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Handle secrets] *********************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:146 Saturday 24 August 2024 12:34:22 -0400 (0:00:00.033) 0:00:31.251 ******* included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml for managed_node1 included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml for managed_node1 included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Set variables part 1] ***************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:3 Saturday 24 August 2024 12:34:22 -0400 (0:00:00.142) 0:00:31.393 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set variables part 2] ***************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:7 Saturday 24 August 2024 12:34:22 -0400 (0:00:00.047) 0:00:31.441 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_rootless": false, "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:13 Saturday 24 August 2024 12:34:22 -0400 (0:00:00.066) 0:00:31.508 ******* included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Saturday 24 August 2024 12:34:23 -0400 (0:00:00.128) 0:00:31.637 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Saturday 24 August 2024 12:34:23 -0400 (0:00:00.050) 0:00:31.687 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Saturday 24 August 2024 12:34:23 -0400 (0:00:00.040) 0:00:31.727 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Stat XDG_RUNTIME_DIR] ***************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:18 Saturday 24 August 2024 12:34:23 -0400 (0:00:00.037) 0:00:31.765 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Manage each secret] ******************* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:34 Saturday 24 August 2024 12:34:23 -0400 (0:00:00.038) 0:00:31.804 ******* [WARNING]: Using a variable for a task's 'args' is unsafe in some situations (see https://docs.ansible.com/ansible/devel/reference_appendices/faq.html#argsplat- unsafe) changed: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": true } TASK [fedora.linux_system_roles.podman : Set variables part 1] ***************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:3 Saturday 24 August 2024 12:34:23 -0400 (0:00:00.632) 0:00:32.436 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set variables part 2] ***************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:7 Saturday 24 August 2024 12:34:23 -0400 (0:00:00.051) 0:00:32.487 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_rootless": false, "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:13 Saturday 24 August 2024 12:34:23 -0400 (0:00:00.046) 0:00:32.533 ******* included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Saturday 24 August 2024 12:34:23 -0400 (0:00:00.068) 0:00:32.602 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Saturday 24 August 2024 12:34:24 -0400 (0:00:00.038) 0:00:32.640 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Saturday 24 August 2024 12:34:24 -0400 (0:00:00.033) 0:00:32.674 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Stat XDG_RUNTIME_DIR] ***************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:18 Saturday 24 August 2024 12:34:24 -0400 (0:00:00.032) 0:00:32.707 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Manage each secret] ******************* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:34 Saturday 24 August 2024 12:34:24 -0400 (0:00:00.033) 0:00:32.740 ******* [WARNING]: Using a variable for a task's 'args' is unsafe in some situations (see https://docs.ansible.com/ansible/devel/reference_appendices/faq.html#argsplat- unsafe) changed: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": true } TASK [fedora.linux_system_roles.podman : Set variables part 1] ***************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:3 Saturday 24 August 2024 12:34:24 -0400 (0:00:00.555) 0:00:33.296 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set variables part 2] ***************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:7 Saturday 24 August 2024 12:34:24 -0400 (0:00:00.033) 0:00:33.329 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_rootless": false, "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:13 Saturday 24 August 2024 12:34:24 -0400 (0:00:00.041) 0:00:33.371 ******* included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Saturday 24 August 2024 12:34:24 -0400 (0:00:00.066) 0:00:33.438 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Saturday 24 August 2024 12:34:24 -0400 (0:00:00.053) 0:00:33.492 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Saturday 24 August 2024 12:34:24 -0400 (0:00:00.051) 0:00:33.543 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Stat XDG_RUNTIME_DIR] ***************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:18 Saturday 24 August 2024 12:34:24 -0400 (0:00:00.053) 0:00:33.596 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Manage each secret] ******************* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:34 Saturday 24 August 2024 12:34:25 -0400 (0:00:00.051) 0:00:33.648 ******* [WARNING]: Using a variable for a task's 'args' is unsafe in some situations (see https://docs.ansible.com/ansible/devel/reference_appendices/faq.html#argsplat- unsafe) changed: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": true } TASK [fedora.linux_system_roles.podman : Handle Kubernetes specifications] ***** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:153 Saturday 24 August 2024 12:34:25 -0400 (0:00:00.556) 0:00:34.205 ******* skipping: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Handle Quadlet specifications] ******** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:160 Saturday 24 August 2024 12:34:25 -0400 (0:00:00.037) 0:00:34.242 ******* included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for managed_node1 included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for managed_node1 included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for managed_node1 included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for managed_node1 included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for managed_node1 included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14 Saturday 24 August 2024 12:34:25 -0400 (0:00:00.142) 0:00:34.384 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_file_src": "quadlet-demo.network", "__podman_quadlet_spec": {}, "__podman_quadlet_str": "[Network]\nSubnet=192.168.30.0/24\nGateway=192.168.30.1\nLabel=app=wordpress", "__podman_quadlet_template_src": "" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25 Saturday 24 August 2024 12:34:25 -0400 (0:00:00.043) 0:00:34.428 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_pull_image": true, "__podman_state": "created", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] ***** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35 Saturday 24 August 2024 12:34:25 -0400 (0:00:00.040) 0:00:34.469 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48 Saturday 24 August 2024 12:34:25 -0400 (0:00:00.072) 0:00:34.541 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_name": "quadlet-demo", "__podman_quadlet_type": "network", "__podman_rootless": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57 Saturday 24 August 2024 12:34:25 -0400 (0:00:00.046) 0:00:34.588 ******* included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Saturday 24 August 2024 12:34:26 -0400 (0:00:00.062) 0:00:34.650 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Saturday 24 August 2024 12:34:26 -0400 (0:00:00.039) 0:00:34.690 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Saturday 24 August 2024 12:34:26 -0400 (0:00:00.041) 0:00:34.731 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get group information] **************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:28 Saturday 24 August 2024 12:34:26 -0400 (0:00:00.044) 0:00:34.776 ******* ok: [managed_node1] => { "ansible_facts": { "getent_group": { "root": [ "x", "0", "" ] } }, "changed": false } TASK [fedora.linux_system_roles.podman : Set group name] *********************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:35 Saturday 24 August 2024 12:34:26 -0400 (0:00:00.377) 0:00:35.153 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_group_name": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:39 Saturday 24 August 2024 12:34:26 -0400 (0:00:00.039) 0:00:35.193 ******* ok: [managed_node1] => { "changed": false, "stat": { "atime": 1724516929.7168462, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "bb5b46ffbafcaa8c4021f3c8b3cb8594f48ef34b", "ctime": 1724516896.725572, "dev": 51713, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 6986657, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-sharedlib", "mode": "0755", "mtime": 1700557386.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 12640, "uid": 0, "version": "1255679238", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************ task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:50 Saturday 24 August 2024 12:34:26 -0400 (0:00:00.392) 0:00:35.585 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check group with getsubids] *********** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:55 Saturday 24 August 2024 12:34:27 -0400 (0:00:00.055) 0:00:35.641 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:60 Saturday 24 August 2024 12:34:27 -0400 (0:00:00.055) 0:00:35.697 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subuid file] ********************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:74 Saturday 24 August 2024 12:34:27 -0400 (0:00:00.053) 0:00:35.751 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subgid file] ********************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:79 Saturday 24 August 2024 12:34:27 -0400 (0:00:00.051) 0:00:35.802 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:84 Saturday 24 August 2024 12:34:27 -0400 (0:00:00.053) 0:00:35.856 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:94 Saturday 24 August 2024 12:34:27 -0400 (0:00:00.050) 0:00:35.906 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] ***** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:101 Saturday 24 August 2024 12:34:27 -0400 (0:00:00.096) 0:00:36.002 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62 Saturday 24 August 2024 12:34:27 -0400 (0:00:00.039) 0:00:36.042 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_activate_systemd_unit": true, "__podman_images_found": [], "__podman_kube_yamls_raw": "", "__podman_service_name": "quadlet-demo-network.service", "__podman_systemd_scope": "system", "__podman_user_home_dir": "/root", "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73 Saturday 24 August 2024 12:34:27 -0400 (0:00:00.066) 0:00:36.109 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_path": "/etc/containers/systemd" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get kube yaml contents] *************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77 Saturday 24 August 2024 12:34:27 -0400 (0:00:00.037) 0:00:36.146 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87 Saturday 24 August 2024 12:34:27 -0400 (0:00:00.034) 0:00:36.181 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_images": [], "__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo.network", "__podman_volumes": [] }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 6] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:103 Saturday 24 August 2024 12:34:27 -0400 (0:00:00.075) 0:00:36.257 ******* ok: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Cleanup quadlets] ********************* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:110 Saturday 24 August 2024 12:34:27 -0400 (0:00:00.039) 0:00:36.297 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create and update quadlets] *********** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:114 Saturday 24 August 2024 12:34:27 -0400 (0:00:00.036) 0:00:36.333 ******* included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:2 Saturday 24 August 2024 12:34:27 -0400 (0:00:00.118) 0:00:36.452 ******* included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Saturday 24 August 2024 12:34:27 -0400 (0:00:00.091) 0:00:36.544 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Saturday 24 August 2024 12:34:27 -0400 (0:00:00.054) 0:00:36.598 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Saturday 24 August 2024 12:34:28 -0400 (0:00:00.052) 0:00:36.651 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create host directories] ************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:7 Saturday 24 August 2024 12:34:28 -0400 (0:00:00.054) 0:00:36.705 ******* TASK [fedora.linux_system_roles.podman : Ensure container images are present] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:18 Saturday 24 August 2024 12:34:28 -0400 (0:00:00.049) 0:00:36.754 ******* skipping: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Ensure the quadlet directory is present] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:39 Saturday 24 August 2024 12:34:28 -0400 (0:00:00.053) 0:00:36.808 ******* ok: [managed_node1] => { "changed": false, "gid": 0, "group": "root", "mode": "0755", "owner": "root", "path": "/etc/containers/systemd", "secontext": "system_u:object_r:etc_t:s0", "size": 6, "state": "directory", "uid": 0 } TASK [fedora.linux_system_roles.podman : Ensure quadlet file is copied] ******** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:48 Saturday 24 August 2024 12:34:28 -0400 (0:00:00.465) 0:00:37.274 ******* changed: [managed_node1] => { "changed": true, "checksum": "e57c08d49aff4bae8daab138d913aeddaa8682a0", "dest": "/etc/containers/systemd/quadlet-demo.network", "gid": 0, "group": "root", "md5sum": "061f3cf318cbd8ab5794bb1173831fb8", "mode": "0644", "owner": "root", "secontext": "system_u:object_r:etc_t:s0", "size": 74, "src": "/root/.ansible/tmp/ansible-tmp-1724517268.7169204-29276-193597116746162/source", "state": "file", "uid": 0 } TASK [fedora.linux_system_roles.podman : Ensure quadlet file content is present] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:58 Saturday 24 August 2024 12:34:29 -0400 (0:00:00.791) 0:00:38.066 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Ensure quadlet file is present] ******* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:70 Saturday 24 August 2024 12:34:29 -0400 (0:00:00.036) 0:00:38.102 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Reload systemctl] ********************* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:82 Saturday 24 August 2024 12:34:29 -0400 (0:00:00.035) 0:00:38.138 ******* ok: [managed_node1] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.podman : Start service] ************************ task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:110 Saturday 24 August 2024 12:34:30 -0400 (0:00:00.601) 0:00:38.739 ******* changed: [managed_node1] => { "changed": true, "name": "quadlet-demo-network.service", "state": "started", "status": { "ActiveEnterTimestampMonotonic": "0", "ActiveExitTimestampMonotonic": "0", "ActiveState": "inactive", "After": "systemd-journald.socket system.slice -.mount sysinit.target basic.target", "AllowIsolate": "no", "AllowedCPUs": "", "AllowedMemoryNodes": "", "AmbientCapabilities": "", "AssertResult": "no", "AssertTimestampMonotonic": "0", "Before": "shutdown.target", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "CPUAccounting": "no", "CPUAffinity": "", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "[not set]", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "no", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf", "CollectMode": "inactive", "ConditionResult": "no", "ConditionTimestampMonotonic": "0", "ConfigurationDirectoryMode": "0755", "Conflicts": "shutdown.target", "ControlPID": "0", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "Delegate": "no", "Description": "quadlet-demo-network.service", "DevicePolicy": "auto", "DynamicUser": "no", "EffectiveCPUs": "", "EffectiveMemoryNodes": "", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "0", "ExecMainStartTimestampMonotonic": "0", "ExecMainStatus": "0", "ExecStart": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman network create --ignore --subnet=192.168.30.0/24 --gateway=192.168.30.1 --label app=wordpress systemd-quadlet-demo ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FragmentPath": "/run/systemd/generator/quadlet-demo-network.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOSchedulingClass": "0", "IOSchedulingPriority": "0", "IOWeight": "[not set]", "IPAccounting": "no", "IPEgressBytes": "18446744073709551615", "IPEgressPackets": "18446744073709551615", "IPIngressBytes": "18446744073709551615", "IPIngressPackets": "18446744073709551615", "Id": "quadlet-demo-network.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestampMonotonic": "0", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "control-group", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "0", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "65536", "LimitMEMLOCKSoft": "65536", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "262144", "LimitNOFILESoft": "1024", "LimitNPROC": "14003", "LimitNPROCSoft": "14003", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "14003", "LimitSIGPENDINGSoft": "14003", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "0", "MemoryAccounting": "yes", "MemoryCurrent": "[not set]", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemorySwapMax": "infinity", "MountAPIVFS": "no", "MountFlags": "", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAMask": "", "NUMAPolicy": "n/a", "Names": "quadlet-demo-network.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "PermissionsStartOnly": "no", "Perpetual": "no", "PrivateDevices": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "RemainAfterExit": "yes", "RemoveIPC": "no", "Requires": "system.slice -.mount sysinit.target", "RequiresMountsFor": "/run/containers", "Restart": "no", "RestartUSec": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "Slice": "system.slice", "StandardError": "inherit", "StandardInput": "null", "StandardInputData": "", "StandardOutput": "journal", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StateChangeTimestampMonotonic": "0", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "dead", "SuccessAction": "none", "SyslogFacility": "3", "SyslogIdentifier": "quadlet-demo-network", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "0", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "[not set]", "TasksMax": "22405", "TimeoutStartUSec": "infinity", "TimeoutStopUSec": "1min 30s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "oneshot", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "disabled", "UnitFileState": "generated", "UtmpMode": "init", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "0" } } TASK [fedora.linux_system_roles.podman : Restart service] ********************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:125 Saturday 24 August 2024 12:34:30 -0400 (0:00:00.566) 0:00:39.305 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14 Saturday 24 August 2024 12:34:30 -0400 (0:00:00.035) 0:00:39.341 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_file_src": "quadlet-demo-mysql.volume", "__podman_quadlet_spec": {}, "__podman_quadlet_str": "[Volume]", "__podman_quadlet_template_src": "" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25 Saturday 24 August 2024 12:34:30 -0400 (0:00:00.051) 0:00:39.392 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_pull_image": true, "__podman_state": "created", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] ***** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35 Saturday 24 August 2024 12:34:30 -0400 (0:00:00.058) 0:00:39.451 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48 Saturday 24 August 2024 12:34:30 -0400 (0:00:00.055) 0:00:39.506 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_name": "quadlet-demo-mysql", "__podman_quadlet_type": "volume", "__podman_rootless": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57 Saturday 24 August 2024 12:34:31 -0400 (0:00:00.127) 0:00:39.633 ******* included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Saturday 24 August 2024 12:34:31 -0400 (0:00:00.098) 0:00:39.731 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Saturday 24 August 2024 12:34:31 -0400 (0:00:00.060) 0:00:39.791 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Saturday 24 August 2024 12:34:31 -0400 (0:00:00.064) 0:00:39.856 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get group information] **************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:28 Saturday 24 August 2024 12:34:31 -0400 (0:00:00.068) 0:00:39.924 ******* ok: [managed_node1] => { "ansible_facts": { "getent_group": { "root": [ "x", "0", "" ] } }, "changed": false } TASK [fedora.linux_system_roles.podman : Set group name] *********************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:35 Saturday 24 August 2024 12:34:31 -0400 (0:00:00.439) 0:00:40.364 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_group_name": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:39 Saturday 24 August 2024 12:34:31 -0400 (0:00:00.066) 0:00:40.431 ******* ok: [managed_node1] => { "changed": false, "stat": { "atime": 1724516929.7168462, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "bb5b46ffbafcaa8c4021f3c8b3cb8594f48ef34b", "ctime": 1724516896.725572, "dev": 51713, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 6986657, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-sharedlib", "mode": "0755", "mtime": 1700557386.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 12640, "uid": 0, "version": "1255679238", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************ task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:50 Saturday 24 August 2024 12:34:32 -0400 (0:00:00.420) 0:00:40.851 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check group with getsubids] *********** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:55 Saturday 24 August 2024 12:34:32 -0400 (0:00:00.056) 0:00:40.908 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:60 Saturday 24 August 2024 12:34:32 -0400 (0:00:00.051) 0:00:40.960 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subuid file] ********************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:74 Saturday 24 August 2024 12:34:32 -0400 (0:00:00.050) 0:00:41.010 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subgid file] ********************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:79 Saturday 24 August 2024 12:34:32 -0400 (0:00:00.060) 0:00:41.071 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:84 Saturday 24 August 2024 12:34:32 -0400 (0:00:00.058) 0:00:41.129 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:94 Saturday 24 August 2024 12:34:32 -0400 (0:00:00.051) 0:00:41.181 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] ***** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:101 Saturday 24 August 2024 12:34:32 -0400 (0:00:00.052) 0:00:41.233 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62 Saturday 24 August 2024 12:34:32 -0400 (0:00:00.050) 0:00:41.283 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_activate_systemd_unit": true, "__podman_images_found": [], "__podman_kube_yamls_raw": "", "__podman_service_name": "quadlet-demo-mysql-volume.service", "__podman_systemd_scope": "system", "__podman_user_home_dir": "/root", "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73 Saturday 24 August 2024 12:34:32 -0400 (0:00:00.206) 0:00:41.490 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_path": "/etc/containers/systemd" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get kube yaml contents] *************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77 Saturday 24 August 2024 12:34:32 -0400 (0:00:00.054) 0:00:41.545 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87 Saturday 24 August 2024 12:34:32 -0400 (0:00:00.054) 0:00:41.599 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_images": [], "__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo-mysql.volume", "__podman_volumes": [] }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 6] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:103 Saturday 24 August 2024 12:34:33 -0400 (0:00:00.099) 0:00:41.699 ******* ok: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Cleanup quadlets] ********************* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:110 Saturday 24 August 2024 12:34:33 -0400 (0:00:00.046) 0:00:41.745 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create and update quadlets] *********** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:114 Saturday 24 August 2024 12:34:33 -0400 (0:00:00.041) 0:00:41.786 ******* included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:2 Saturday 24 August 2024 12:34:33 -0400 (0:00:00.072) 0:00:41.858 ******* included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Saturday 24 August 2024 12:34:33 -0400 (0:00:00.055) 0:00:41.913 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Saturday 24 August 2024 12:34:33 -0400 (0:00:00.033) 0:00:41.946 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Saturday 24 August 2024 12:34:33 -0400 (0:00:00.038) 0:00:41.985 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create host directories] ************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:7 Saturday 24 August 2024 12:34:33 -0400 (0:00:00.042) 0:00:42.028 ******* TASK [fedora.linux_system_roles.podman : Ensure container images are present] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:18 Saturday 24 August 2024 12:34:33 -0400 (0:00:00.048) 0:00:42.076 ******* skipping: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Ensure the quadlet directory is present] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:39 Saturday 24 August 2024 12:34:33 -0400 (0:00:00.058) 0:00:42.135 ******* ok: [managed_node1] => { "changed": false, "gid": 0, "group": "root", "mode": "0755", "owner": "root", "path": "/etc/containers/systemd", "secontext": "system_u:object_r:etc_t:s0", "size": 34, "state": "directory", "uid": 0 } TASK [fedora.linux_system_roles.podman : Ensure quadlet file is copied] ******** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:48 Saturday 24 August 2024 12:34:33 -0400 (0:00:00.413) 0:00:42.548 ******* changed: [managed_node1] => { "changed": true, "checksum": "585f8cbdf0ec73000f9227dcffbef71e9552ea4a", "dest": "/etc/containers/systemd/quadlet-demo-mysql.volume", "gid": 0, "group": "root", "md5sum": "5ddd03a022aeb4502d9bc8ce436b4233", "mode": "0644", "owner": "root", "secontext": "system_u:object_r:etc_t:s0", "size": 9, "src": "/root/.ansible/tmp/ansible-tmp-1724517273.9811473-29551-138782588482123/source", "state": "file", "uid": 0 } TASK [fedora.linux_system_roles.podman : Ensure quadlet file content is present] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:58 Saturday 24 August 2024 12:34:34 -0400 (0:00:00.750) 0:00:43.298 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Ensure quadlet file is present] ******* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:70 Saturday 24 August 2024 12:34:34 -0400 (0:00:00.052) 0:00:43.351 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Reload systemctl] ********************* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:82 Saturday 24 August 2024 12:34:34 -0400 (0:00:00.055) 0:00:43.407 ******* ok: [managed_node1] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.podman : Start service] ************************ task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:110 Saturday 24 August 2024 12:34:35 -0400 (0:00:00.671) 0:00:44.079 ******* changed: [managed_node1] => { "changed": true, "name": "quadlet-demo-mysql-volume.service", "state": "started", "status": { "ActiveEnterTimestampMonotonic": "0", "ActiveExitTimestampMonotonic": "0", "ActiveState": "inactive", "After": "system.slice sysinit.target systemd-journald.socket -.mount basic.target", "AllowIsolate": "no", "AllowedCPUs": "", "AllowedMemoryNodes": "", "AmbientCapabilities": "", "AssertResult": "no", "AssertTimestampMonotonic": "0", "Before": "shutdown.target", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "CPUAccounting": "no", "CPUAffinity": "", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "[not set]", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "no", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf", "CollectMode": "inactive", "ConditionResult": "no", "ConditionTimestampMonotonic": "0", "ConfigurationDirectoryMode": "0755", "Conflicts": "shutdown.target", "ControlPID": "0", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "Delegate": "no", "Description": "quadlet-demo-mysql-volume.service", "DevicePolicy": "auto", "DynamicUser": "no", "EffectiveCPUs": "", "EffectiveMemoryNodes": "", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "0", "ExecMainStartTimestampMonotonic": "0", "ExecMainStatus": "0", "ExecStart": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman volume create --ignore systemd-quadlet-demo-mysql ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FragmentPath": "/run/systemd/generator/quadlet-demo-mysql-volume.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOSchedulingClass": "0", "IOSchedulingPriority": "0", "IOWeight": "[not set]", "IPAccounting": "no", "IPEgressBytes": "18446744073709551615", "IPEgressPackets": "18446744073709551615", "IPIngressBytes": "18446744073709551615", "IPIngressPackets": "18446744073709551615", "Id": "quadlet-demo-mysql-volume.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestampMonotonic": "0", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "control-group", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "0", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "65536", "LimitMEMLOCKSoft": "65536", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "262144", "LimitNOFILESoft": "1024", "LimitNPROC": "14003", "LimitNPROCSoft": "14003", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "14003", "LimitSIGPENDINGSoft": "14003", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "0", "MemoryAccounting": "yes", "MemoryCurrent": "[not set]", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemorySwapMax": "infinity", "MountAPIVFS": "no", "MountFlags": "", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAMask": "", "NUMAPolicy": "n/a", "Names": "quadlet-demo-mysql-volume.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "PermissionsStartOnly": "no", "Perpetual": "no", "PrivateDevices": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "RemainAfterExit": "yes", "RemoveIPC": "no", "Requires": "system.slice sysinit.target -.mount", "RequiresMountsFor": "/run/containers", "Restart": "no", "RestartUSec": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "Slice": "system.slice", "StandardError": "inherit", "StandardInput": "null", "StandardInputData": "", "StandardOutput": "journal", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StateChangeTimestampMonotonic": "0", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "dead", "SuccessAction": "none", "SyslogFacility": "3", "SyslogIdentifier": "quadlet-demo-mysql-volume", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "0", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "[not set]", "TasksMax": "22405", "TimeoutStartUSec": "infinity", "TimeoutStopUSec": "1min 30s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "oneshot", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "disabled", "UnitFileState": "generated", "UtmpMode": "init", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "0" } } TASK [fedora.linux_system_roles.podman : Restart service] ********************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:125 Saturday 24 August 2024 12:34:36 -0400 (0:00:00.581) 0:00:44.660 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14 Saturday 24 August 2024 12:34:36 -0400 (0:00:00.034) 0:00:44.695 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_file_src": "", "__podman_quadlet_spec": {}, "__podman_quadlet_str": "[Install]\nWantedBy=default.target\n\n[Container]\nImage=quay.io/linux-system-roles/mysql:5.6\nContainerName=quadlet-demo-mysql\nVolume=quadlet-demo-mysql.volume:/var/lib/mysql\nVolume=/tmp/quadlet_demo:/var/lib/quadlet_demo:Z\nNetwork=quadlet-demo.network\nSecret=mysql-root-password-container,type=env,target=MYSQL_ROOT_PASSWORD\nHealthCmd=/bin/true\nHealthOnFailure=kill\n", "__podman_quadlet_template_src": "quadlet-demo-mysql.container.j2" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25 Saturday 24 August 2024 12:34:36 -0400 (0:00:00.102) 0:00:44.798 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_pull_image": true, "__podman_state": "created", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] ***** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35 Saturday 24 August 2024 12:34:36 -0400 (0:00:00.053) 0:00:44.851 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48 Saturday 24 August 2024 12:34:36 -0400 (0:00:00.053) 0:00:44.904 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_name": "quadlet-demo-mysql", "__podman_quadlet_type": "container", "__podman_rootless": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57 Saturday 24 August 2024 12:34:36 -0400 (0:00:00.050) 0:00:44.955 ******* included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Saturday 24 August 2024 12:34:36 -0400 (0:00:00.074) 0:00:45.030 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Saturday 24 August 2024 12:34:36 -0400 (0:00:00.089) 0:00:45.119 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Saturday 24 August 2024 12:34:36 -0400 (0:00:00.042) 0:00:45.161 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get group information] **************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:28 Saturday 24 August 2024 12:34:36 -0400 (0:00:00.045) 0:00:45.207 ******* ok: [managed_node1] => { "ansible_facts": { "getent_group": { "root": [ "x", "0", "" ] } }, "changed": false } TASK [fedora.linux_system_roles.podman : Set group name] *********************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:35 Saturday 24 August 2024 12:34:37 -0400 (0:00:00.432) 0:00:45.640 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_group_name": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:39 Saturday 24 August 2024 12:34:37 -0400 (0:00:00.065) 0:00:45.706 ******* ok: [managed_node1] => { "changed": false, "stat": { "atime": 1724516929.7168462, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "bb5b46ffbafcaa8c4021f3c8b3cb8594f48ef34b", "ctime": 1724516896.725572, "dev": 51713, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 6986657, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-sharedlib", "mode": "0755", "mtime": 1700557386.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 12640, "uid": 0, "version": "1255679238", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************ task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:50 Saturday 24 August 2024 12:34:37 -0400 (0:00:00.421) 0:00:46.127 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check group with getsubids] *********** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:55 Saturday 24 August 2024 12:34:37 -0400 (0:00:00.046) 0:00:46.173 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:60 Saturday 24 August 2024 12:34:37 -0400 (0:00:00.035) 0:00:46.209 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subuid file] ********************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:74 Saturday 24 August 2024 12:34:37 -0400 (0:00:00.034) 0:00:46.244 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subgid file] ********************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:79 Saturday 24 August 2024 12:34:37 -0400 (0:00:00.033) 0:00:46.277 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:84 Saturday 24 August 2024 12:34:37 -0400 (0:00:00.034) 0:00:46.311 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:94 Saturday 24 August 2024 12:34:37 -0400 (0:00:00.037) 0:00:46.349 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] ***** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:101 Saturday 24 August 2024 12:34:37 -0400 (0:00:00.051) 0:00:46.400 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62 Saturday 24 August 2024 12:34:37 -0400 (0:00:00.054) 0:00:46.455 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_activate_systemd_unit": true, "__podman_images_found": [ "quay.io/linux-system-roles/mysql:5.6" ], "__podman_kube_yamls_raw": "", "__podman_service_name": "quadlet-demo-mysql.service", "__podman_systemd_scope": "system", "__podman_user_home_dir": "/root", "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73 Saturday 24 August 2024 12:34:37 -0400 (0:00:00.076) 0:00:46.532 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_path": "/etc/containers/systemd" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get kube yaml contents] *************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77 Saturday 24 August 2024 12:34:37 -0400 (0:00:00.043) 0:00:46.575 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87 Saturday 24 August 2024 12:34:38 -0400 (0:00:00.042) 0:00:46.618 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_images": [ "quay.io/linux-system-roles/mysql:5.6" ], "__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo-mysql.container", "__podman_volumes": [ "/tmp/quadlet_demo" ] }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 6] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:103 Saturday 24 August 2024 12:34:38 -0400 (0:00:00.161) 0:00:46.779 ******* ok: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Cleanup quadlets] ********************* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:110 Saturday 24 August 2024 12:34:38 -0400 (0:00:00.039) 0:00:46.819 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create and update quadlets] *********** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:114 Saturday 24 August 2024 12:34:38 -0400 (0:00:00.033) 0:00:46.852 ******* included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:2 Saturday 24 August 2024 12:34:38 -0400 (0:00:00.100) 0:00:46.953 ******* included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Saturday 24 August 2024 12:34:38 -0400 (0:00:00.081) 0:00:47.035 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Saturday 24 August 2024 12:34:38 -0400 (0:00:00.036) 0:00:47.071 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Saturday 24 August 2024 12:34:38 -0400 (0:00:00.037) 0:00:47.109 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create host directories] ************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:7 Saturday 24 August 2024 12:34:38 -0400 (0:00:00.038) 0:00:47.147 ******* [WARNING]: Using a variable for a task's 'args' is unsafe in some situations (see https://docs.ansible.com/ansible/devel/reference_appendices/faq.html#argsplat- unsafe) changed: [managed_node1] => (item=/tmp/quadlet_demo) => { "ansible_loop_var": "item", "changed": true, "gid": 0, "group": "root", "item": "/tmp/quadlet_demo", "mode": "0777", "owner": "root", "path": "/tmp/quadlet_demo", "secontext": "unconfined_u:object_r:user_tmp_t:s0", "size": 6, "state": "directory", "uid": 0 } TASK [fedora.linux_system_roles.podman : Ensure container images are present] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:18 Saturday 24 August 2024 12:34:38 -0400 (0:00:00.414) 0:00:47.562 ******* changed: [managed_node1] => (item=None) => { "attempts": 1, "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": true } changed: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": true } TASK [fedora.linux_system_roles.podman : Ensure the quadlet directory is present] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:39 Saturday 24 August 2024 12:34:46 -0400 (0:00:07.231) 0:00:54.793 ******* ok: [managed_node1] => { "changed": false, "gid": 0, "group": "root", "mode": "0755", "owner": "root", "path": "/etc/containers/systemd", "secontext": "system_u:object_r:etc_t:s0", "size": 67, "state": "directory", "uid": 0 } TASK [fedora.linux_system_roles.podman : Ensure quadlet file is copied] ******** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:48 Saturday 24 August 2024 12:34:46 -0400 (0:00:00.428) 0:00:55.222 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Ensure quadlet file content is present] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:58 Saturday 24 August 2024 12:34:46 -0400 (0:00:00.047) 0:00:55.269 ******* changed: [managed_node1] => { "changed": true, "checksum": "ca62b2ad3cc9afb5b5371ebbf797b9bc4fd7edd4", "dest": "/etc/containers/systemd/quadlet-demo-mysql.container", "gid": 0, "group": "root", "md5sum": "341b473056d2a5dfa35970b0d2e23a5d", "mode": "0644", "owner": "root", "secontext": "system_u:object_r:etc_t:s0", "size": 363, "src": "/root/.ansible/tmp/ansible-tmp-1724517286.7033756-30107-189472742671718/source", "state": "file", "uid": 0 } TASK [fedora.linux_system_roles.podman : Ensure quadlet file is present] ******* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:70 Saturday 24 August 2024 12:34:47 -0400 (0:00:00.718) 0:00:55.988 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Reload systemctl] ********************* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:82 Saturday 24 August 2024 12:34:47 -0400 (0:00:00.036) 0:00:56.024 ******* ok: [managed_node1] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.podman : Start service] ************************ task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:110 Saturday 24 August 2024 12:34:48 -0400 (0:00:00.648) 0:00:56.672 ******* changed: [managed_node1] => { "changed": true, "name": "quadlet-demo-mysql.service", "state": "started", "status": { "ActiveEnterTimestampMonotonic": "0", "ActiveExitTimestampMonotonic": "0", "ActiveState": "inactive", "After": "tmp.mount system.slice basic.target quadlet-demo-mysql-volume.service quadlet-demo-network.service -.mount systemd-journald.socket sysinit.target", "AllowIsolate": "no", "AllowedCPUs": "", "AllowedMemoryNodes": "", "AmbientCapabilities": "", "AssertResult": "no", "AssertTimestampMonotonic": "0", "Before": "shutdown.target", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "CPUAccounting": "no", "CPUAffinity": "", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "[not set]", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "no", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf", "CollectMode": "inactive", "ConditionResult": "no", "ConditionTimestampMonotonic": "0", "ConfigurationDirectoryMode": "0755", "Conflicts": "shutdown.target", "ControlPID": "0", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "Delegate": "yes", "DelegateControllers": "cpu cpuacct cpuset io blkio memory devices pids", "Description": "quadlet-demo-mysql.service", "DevicePolicy": "auto", "DynamicUser": "no", "EffectiveCPUs": "", "EffectiveMemoryNodes": "", "Environment": "PODMAN_SYSTEMD_UNIT=quadlet-demo-mysql.service", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "0", "ExecMainStartTimestampMonotonic": "0", "ExecMainStatus": "0", "ExecStart": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman run --name=quadlet-demo-mysql --cidfile=/run/quadlet-demo-mysql.cid --replace --rm --cgroups=split --network=systemd-quadlet-demo --sdnotify=conmon -d -v systemd-quadlet-demo-mysql:/var/lib/mysql -v /tmp/quadlet_demo:/var/lib/quadlet_demo:Z --secret mysql-root-password-container,type=env,target=MYSQL_ROOT_PASSWORD --health-cmd /bin/true --health-on-failure kill quay.io/linux-system-roles/mysql:5.6 ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStop": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman rm -v -f -i --cidfile=/run/quadlet-demo-mysql.cid ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStopPost": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman rm -v -f -i --cidfile=/run/quadlet-demo-mysql.cid ; ignore_errors=yes ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FragmentPath": "/run/systemd/generator/quadlet-demo-mysql.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOSchedulingClass": "0", "IOSchedulingPriority": "0", "IOWeight": "[not set]", "IPAccounting": "no", "IPEgressBytes": "18446744073709551615", "IPEgressPackets": "18446744073709551615", "IPIngressBytes": "18446744073709551615", "IPIngressPackets": "18446744073709551615", "Id": "quadlet-demo-mysql.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestampMonotonic": "0", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "mixed", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "0", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "65536", "LimitMEMLOCKSoft": "65536", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "262144", "LimitNOFILESoft": "1024", "LimitNPROC": "14003", "LimitNPROCSoft": "14003", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "14003", "LimitSIGPENDINGSoft": "14003", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "0", "MemoryAccounting": "yes", "MemoryCurrent": "[not set]", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemorySwapMax": "infinity", "MountAPIVFS": "no", "MountFlags": "", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAMask": "", "NUMAPolicy": "n/a", "Names": "quadlet-demo-mysql.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "all", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "PermissionsStartOnly": "no", "Perpetual": "no", "PrivateDevices": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "RemainAfterExit": "no", "RemoveIPC": "no", "Requires": "quadlet-demo-mysql-volume.service quadlet-demo-network.service system.slice -.mount sysinit.target", "RequiresMountsFor": "/tmp/quadlet_demo /run/containers", "Restart": "no", "RestartUSec": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "Slice": "system.slice", "SourcePath": "/etc/containers/systemd/quadlet-demo-mysql.container", "StandardError": "inherit", "StandardInput": "null", "StandardInputData": "", "StandardOutput": "journal", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StateChangeTimestampMonotonic": "0", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "dead", "SuccessAction": "none", "SyslogFacility": "3", "SyslogIdentifier": "quadlet-demo-mysql", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "0", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "[not set]", "TasksMax": "22405", "TimeoutStartUSec": "1min 30s", "TimeoutStopUSec": "1min 30s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "notify", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "disabled", "UnitFileState": "generated", "UtmpMode": "init", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "0" } } TASK [fedora.linux_system_roles.podman : Restart service] ********************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:125 Saturday 24 August 2024 12:34:49 -0400 (0:00:01.042) 0:00:57.715 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14 Saturday 24 August 2024 12:34:49 -0400 (0:00:00.041) 0:00:57.756 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_file_src": "envoy-proxy-configmap.yml", "__podman_quadlet_spec": {}, "__podman_quadlet_str": "---\napiVersion: v1\nkind: ConfigMap\nmetadata:\n name: envoy-proxy-config\ndata:\n envoy.yaml: |\n admin:\n address:\n socket_address:\n address: 0.0.0.0\n port_value: 9901\n\n static_resources:\n listeners:\n - name: listener_0\n address:\n socket_address:\n address: 0.0.0.0\n port_value: 8080\n filter_chains:\n - filters:\n - name: envoy.filters.network.http_connection_manager\n typed_config:\n \"@type\": type.googleapis.com/envoy.extensions.filters.network.http_connection_manager.v3.HttpConnectionManager\n stat_prefix: ingress_http\n codec_type: AUTO\n route_config:\n name: local_route\n virtual_hosts:\n - name: local_service\n domains: [\"*\"]\n routes:\n - match:\n prefix: \"/\"\n route:\n cluster: backend\n http_filters:\n - name: envoy.filters.http.router\n typed_config:\n \"@type\": type.googleapis.com/envoy.extensions.filters.http.router.v3.Router\n transport_socket:\n name: envoy.transport_sockets.tls\n typed_config:\n \"@type\": type.googleapis.com/envoy.extensions.transport_sockets.tls.v3.DownstreamTlsContext\n common_tls_context:\n tls_certificates:\n - certificate_chain:\n filename: /etc/envoy-certificates/certificate.pem\n private_key:\n filename: /etc/envoy-certificates/certificate.key\n clusters:\n - name: backend\n connect_timeout: 5s\n type: STATIC\n dns_refresh_rate: 1800s\n lb_policy: ROUND_ROBIN\n load_assignment:\n cluster_name: backend\n endpoints:\n - lb_endpoints:\n - endpoint:\n address:\n socket_address:\n address: 127.0.0.1\n port_value: 80", "__podman_quadlet_template_src": "" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25 Saturday 24 August 2024 12:34:49 -0400 (0:00:00.048) 0:00:57.805 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_pull_image": true, "__podman_state": "created", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] ***** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35 Saturday 24 August 2024 12:34:49 -0400 (0:00:00.047) 0:00:57.852 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48 Saturday 24 August 2024 12:34:49 -0400 (0:00:00.035) 0:00:57.888 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_name": "envoy-proxy-configmap", "__podman_quadlet_type": "yml", "__podman_rootless": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57 Saturday 24 August 2024 12:34:49 -0400 (0:00:00.046) 0:00:57.934 ******* included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Saturday 24 August 2024 12:34:49 -0400 (0:00:00.060) 0:00:57.994 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Saturday 24 August 2024 12:34:49 -0400 (0:00:00.071) 0:00:58.066 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Saturday 24 August 2024 12:34:49 -0400 (0:00:00.051) 0:00:58.117 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get group information] **************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:28 Saturday 24 August 2024 12:34:49 -0400 (0:00:00.067) 0:00:58.184 ******* ok: [managed_node1] => { "ansible_facts": { "getent_group": { "root": [ "x", "0", "" ] } }, "changed": false } TASK [fedora.linux_system_roles.podman : Set group name] *********************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:35 Saturday 24 August 2024 12:34:49 -0400 (0:00:00.417) 0:00:58.601 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_group_name": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:39 Saturday 24 August 2024 12:34:50 -0400 (0:00:00.051) 0:00:58.653 ******* ok: [managed_node1] => { "changed": false, "stat": { "atime": 1724516929.7168462, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "bb5b46ffbafcaa8c4021f3c8b3cb8594f48ef34b", "ctime": 1724516896.725572, "dev": 51713, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 6986657, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-sharedlib", "mode": "0755", "mtime": 1700557386.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 12640, "uid": 0, "version": "1255679238", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************ task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:50 Saturday 24 August 2024 12:34:50 -0400 (0:00:00.408) 0:00:59.062 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check group with getsubids] *********** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:55 Saturday 24 August 2024 12:34:50 -0400 (0:00:00.038) 0:00:59.100 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:60 Saturday 24 August 2024 12:34:50 -0400 (0:00:00.081) 0:00:59.182 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subuid file] ********************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:74 Saturday 24 August 2024 12:34:50 -0400 (0:00:00.031) 0:00:59.214 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subgid file] ********************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:79 Saturday 24 August 2024 12:34:50 -0400 (0:00:00.031) 0:00:59.245 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:84 Saturday 24 August 2024 12:34:50 -0400 (0:00:00.032) 0:00:59.278 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:94 Saturday 24 August 2024 12:34:50 -0400 (0:00:00.034) 0:00:59.312 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] ***** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:101 Saturday 24 August 2024 12:34:50 -0400 (0:00:00.037) 0:00:59.349 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62 Saturday 24 August 2024 12:34:50 -0400 (0:00:00.045) 0:00:59.395 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_activate_systemd_unit": true, "__podman_images_found": [], "__podman_kube_yamls_raw": "", "__podman_service_name": "", "__podman_systemd_scope": "system", "__podman_user_home_dir": "/root", "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73 Saturday 24 August 2024 12:34:50 -0400 (0:00:00.092) 0:00:59.488 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_path": "/etc/containers/systemd" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get kube yaml contents] *************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77 Saturday 24 August 2024 12:34:50 -0400 (0:00:00.053) 0:00:59.541 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87 Saturday 24 August 2024 12:34:50 -0400 (0:00:00.055) 0:00:59.596 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_images": [], "__podman_quadlet_file": "/etc/containers/systemd/envoy-proxy-configmap.yml", "__podman_volumes": [] }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 6] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:103 Saturday 24 August 2024 12:34:51 -0400 (0:00:00.115) 0:00:59.711 ******* ok: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Cleanup quadlets] ********************* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:110 Saturday 24 August 2024 12:34:51 -0400 (0:00:00.059) 0:00:59.770 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create and update quadlets] *********** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:114 Saturday 24 August 2024 12:34:51 -0400 (0:00:00.051) 0:00:59.822 ******* included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:2 Saturday 24 August 2024 12:34:51 -0400 (0:00:00.110) 0:00:59.932 ******* included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Saturday 24 August 2024 12:34:51 -0400 (0:00:00.089) 0:01:00.022 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Saturday 24 August 2024 12:34:51 -0400 (0:00:00.053) 0:01:00.075 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Saturday 24 August 2024 12:34:51 -0400 (0:00:00.050) 0:01:00.126 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create host directories] ************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:7 Saturday 24 August 2024 12:34:51 -0400 (0:00:00.101) 0:01:00.227 ******* TASK [fedora.linux_system_roles.podman : Ensure container images are present] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:18 Saturday 24 August 2024 12:34:51 -0400 (0:00:00.050) 0:01:00.278 ******* skipping: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Ensure the quadlet directory is present] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:39 Saturday 24 August 2024 12:34:51 -0400 (0:00:00.059) 0:01:00.337 ******* ok: [managed_node1] => { "changed": false, "gid": 0, "group": "root", "mode": "0755", "owner": "root", "path": "/etc/containers/systemd", "secontext": "system_u:object_r:etc_t:s0", "size": 103, "state": "directory", "uid": 0 } TASK [fedora.linux_system_roles.podman : Ensure quadlet file is copied] ******** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:48 Saturday 24 August 2024 12:34:52 -0400 (0:00:00.416) 0:01:00.753 ******* changed: [managed_node1] => { "changed": true, "checksum": "d681c7d56f912150d041873e880818b22a90c188", "dest": "/etc/containers/systemd/envoy-proxy-configmap.yml", "gid": 0, "group": "root", "md5sum": "aec75d972c231aac004e1338934544cf", "mode": "0644", "owner": "root", "secontext": "system_u:object_r:etc_t:s0", "size": 2102, "src": "/root/.ansible/tmp/ansible-tmp-1724517292.2012258-30396-55356854144861/source", "state": "file", "uid": 0 } TASK [fedora.linux_system_roles.podman : Ensure quadlet file content is present] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:58 Saturday 24 August 2024 12:34:52 -0400 (0:00:00.740) 0:01:01.493 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Ensure quadlet file is present] ******* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:70 Saturday 24 August 2024 12:34:52 -0400 (0:00:00.034) 0:01:01.528 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Reload systemctl] ********************* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:82 Saturday 24 August 2024 12:34:52 -0400 (0:00:00.041) 0:01:01.570 ******* ok: [managed_node1] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.podman : Start service] ************************ task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:110 Saturday 24 August 2024 12:34:53 -0400 (0:00:00.660) 0:01:02.230 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Restart service] ********************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:125 Saturday 24 August 2024 12:34:53 -0400 (0:00:00.054) 0:01:02.284 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14 Saturday 24 August 2024 12:34:53 -0400 (0:00:00.068) 0:01:02.353 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_file_src": "", "__podman_quadlet_spec": {}, "__podman_quadlet_str": "---\napiVersion: v1\nkind: PersistentVolumeClaim\nmetadata:\n name: wp-pv-claim\n labels:\n app: wordpress\nspec:\n accessModes:\n - ReadWriteOnce\n resources:\n requests:\n storage: 20Gi\n---\napiVersion: v1\nkind: Pod\nmetadata:\n name: quadlet-demo\nspec:\n containers:\n - name: wordpress\n image: quay.io/linux-system-roles/wordpress:4.8-apache\n env:\n - name: WORDPRESS_DB_HOST\n value: quadlet-demo-mysql\n - name: WORDPRESS_DB_PASSWORD\n valueFrom:\n secretKeyRef:\n name: mysql-root-password-kube\n key: password\n volumeMounts:\n - name: wordpress-persistent-storage\n mountPath: /var/www/html\n resources:\n requests:\n memory: \"64Mi\"\n cpu: \"250m\"\n limits:\n memory: \"128Mi\"\n cpu: \"500m\"\n - name: envoy\n image: quay.io/linux-system-roles/envoyproxy:v1.25.0\n volumeMounts:\n - name: config-volume\n mountPath: /etc/envoy\n - name: certificates\n mountPath: /etc/envoy-certificates\n env:\n - name: ENVOY_UID\n value: \"0\"\n resources:\n requests:\n memory: \"64Mi\"\n cpu: \"250m\"\n limits:\n memory: \"128Mi\"\n cpu: \"500m\"\n volumes:\n - name: config-volume\n configMap:\n name: envoy-proxy-config\n - name: certificates\n secret:\n secretName: envoy-certificates\n - name: wordpress-persistent-storage\n persistentVolumeClaim:\n claimName: wp-pv-claim\n - name: www # not used - for testing hostpath\n hostPath:\n path: /tmp/httpd3\n - name: create # not used - for testing hostpath\n hostPath:\n path: /tmp/httpd3-create\n", "__podman_quadlet_template_src": "quadlet-demo.yml.j2" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25 Saturday 24 August 2024 12:34:53 -0400 (0:00:00.201) 0:01:02.555 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_pull_image": true, "__podman_state": "created", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] ***** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35 Saturday 24 August 2024 12:34:54 -0400 (0:00:00.061) 0:01:02.616 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48 Saturday 24 August 2024 12:34:54 -0400 (0:00:00.105) 0:01:02.721 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_name": "quadlet-demo", "__podman_quadlet_type": "yml", "__podman_rootless": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57 Saturday 24 August 2024 12:34:54 -0400 (0:00:00.071) 0:01:02.793 ******* included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Saturday 24 August 2024 12:34:54 -0400 (0:00:00.097) 0:01:02.890 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Saturday 24 August 2024 12:34:54 -0400 (0:00:00.062) 0:01:02.953 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Saturday 24 August 2024 12:34:54 -0400 (0:00:00.061) 0:01:03.014 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get group information] **************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:28 Saturday 24 August 2024 12:34:54 -0400 (0:00:00.067) 0:01:03.082 ******* ok: [managed_node1] => { "ansible_facts": { "getent_group": { "root": [ "x", "0", "" ] } }, "changed": false } TASK [fedora.linux_system_roles.podman : Set group name] *********************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:35 Saturday 24 August 2024 12:34:54 -0400 (0:00:00.418) 0:01:03.501 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_group_name": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:39 Saturday 24 August 2024 12:34:54 -0400 (0:00:00.062) 0:01:03.564 ******* ok: [managed_node1] => { "changed": false, "stat": { "atime": 1724516929.7168462, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "bb5b46ffbafcaa8c4021f3c8b3cb8594f48ef34b", "ctime": 1724516896.725572, "dev": 51713, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 6986657, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-sharedlib", "mode": "0755", "mtime": 1700557386.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 12640, "uid": 0, "version": "1255679238", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************ task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:50 Saturday 24 August 2024 12:34:55 -0400 (0:00:00.478) 0:01:04.043 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check group with getsubids] *********** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:55 Saturday 24 August 2024 12:34:55 -0400 (0:00:00.037) 0:01:04.080 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:60 Saturday 24 August 2024 12:34:55 -0400 (0:00:00.037) 0:01:04.117 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subuid file] ********************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:74 Saturday 24 August 2024 12:34:55 -0400 (0:00:00.033) 0:01:04.151 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subgid file] ********************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:79 Saturday 24 August 2024 12:34:55 -0400 (0:00:00.030) 0:01:04.181 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:84 Saturday 24 August 2024 12:34:55 -0400 (0:00:00.031) 0:01:04.213 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:94 Saturday 24 August 2024 12:34:55 -0400 (0:00:00.030) 0:01:04.243 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] ***** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:101 Saturday 24 August 2024 12:34:55 -0400 (0:00:00.078) 0:01:04.322 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62 Saturday 24 August 2024 12:34:55 -0400 (0:00:00.060) 0:01:04.382 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_activate_systemd_unit": true, "__podman_images_found": [], "__podman_kube_yamls_raw": "", "__podman_service_name": "", "__podman_systemd_scope": "system", "__podman_user_home_dir": "/root", "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73 Saturday 24 August 2024 12:34:55 -0400 (0:00:00.067) 0:01:04.449 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_path": "/etc/containers/systemd" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get kube yaml contents] *************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77 Saturday 24 August 2024 12:34:55 -0400 (0:00:00.040) 0:01:04.490 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87 Saturday 24 August 2024 12:34:55 -0400 (0:00:00.041) 0:01:04.532 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_images": [], "__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo.yml", "__podman_volumes": [] }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 6] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:103 Saturday 24 August 2024 12:34:55 -0400 (0:00:00.073) 0:01:04.605 ******* ok: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Cleanup quadlets] ********************* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:110 Saturday 24 August 2024 12:34:56 -0400 (0:00:00.037) 0:01:04.643 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create and update quadlets] *********** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:114 Saturday 24 August 2024 12:34:56 -0400 (0:00:00.031) 0:01:04.674 ******* included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:2 Saturday 24 August 2024 12:34:56 -0400 (0:00:00.088) 0:01:04.763 ******* included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Saturday 24 August 2024 12:34:56 -0400 (0:00:00.087) 0:01:04.851 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Saturday 24 August 2024 12:34:56 -0400 (0:00:00.054) 0:01:04.905 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Saturday 24 August 2024 12:34:56 -0400 (0:00:00.050) 0:01:04.956 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create host directories] ************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:7 Saturday 24 August 2024 12:34:56 -0400 (0:00:00.050) 0:01:05.006 ******* TASK [fedora.linux_system_roles.podman : Ensure container images are present] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:18 Saturday 24 August 2024 12:34:56 -0400 (0:00:00.048) 0:01:05.055 ******* skipping: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Ensure the quadlet directory is present] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:39 Saturday 24 August 2024 12:34:56 -0400 (0:00:00.048) 0:01:05.103 ******* ok: [managed_node1] => { "changed": false, "gid": 0, "group": "root", "mode": "0755", "owner": "root", "path": "/etc/containers/systemd", "secontext": "system_u:object_r:etc_t:s0", "size": 136, "state": "directory", "uid": 0 } TASK [fedora.linux_system_roles.podman : Ensure quadlet file is copied] ******** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:48 Saturday 24 August 2024 12:34:56 -0400 (0:00:00.480) 0:01:05.583 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Ensure quadlet file content is present] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:58 Saturday 24 August 2024 12:34:57 -0400 (0:00:00.037) 0:01:05.621 ******* changed: [managed_node1] => { "changed": true, "checksum": "998dccde0483b1654327a46ddd89cbaa47650370", "dest": "/etc/containers/systemd/quadlet-demo.yml", "gid": 0, "group": "root", "md5sum": "fd890594adfc24339cb9cdc5e7b19a66", "mode": "0644", "owner": "root", "secontext": "system_u:object_r:etc_t:s0", "size": 1605, "src": "/root/.ansible/tmp/ansible-tmp-1724517297.0551069-30667-219028600000855/source", "state": "file", "uid": 0 } TASK [fedora.linux_system_roles.podman : Ensure quadlet file is present] ******* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:70 Saturday 24 August 2024 12:34:57 -0400 (0:00:00.737) 0:01:06.359 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Reload systemctl] ********************* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:82 Saturday 24 August 2024 12:34:57 -0400 (0:00:00.058) 0:01:06.417 ******* ok: [managed_node1] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.podman : Start service] ************************ task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:110 Saturday 24 August 2024 12:34:58 -0400 (0:00:00.636) 0:01:07.054 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Restart service] ********************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:125 Saturday 24 August 2024 12:34:58 -0400 (0:00:00.058) 0:01:07.112 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14 Saturday 24 August 2024 12:34:58 -0400 (0:00:00.057) 0:01:07.170 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_file_src": "quadlet-demo.kube", "__podman_quadlet_spec": {}, "__podman_quadlet_str": "[Install]\nWantedBy=default.target\n\n[Unit]\nRequires=quadlet-demo-mysql.service\nAfter=quadlet-demo-mysql.service\n\n[Kube]\n# Point to the yaml file in the same directory\nYaml=quadlet-demo.yml\n# Use the quadlet-demo network\nNetwork=quadlet-demo.network\n# Publish the envoy proxy data port\nPublishPort=8000:8080\n# Publish the envoy proxy admin port\nPublishPort=9000:9901\n# Use the envoy proxy config map in the same directory\nConfigMap=envoy-proxy-configmap.yml", "__podman_quadlet_template_src": "" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25 Saturday 24 August 2024 12:34:58 -0400 (0:00:00.069) 0:01:07.239 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_pull_image": true, "__podman_state": "created", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] ***** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35 Saturday 24 August 2024 12:34:58 -0400 (0:00:00.059) 0:01:07.299 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48 Saturday 24 August 2024 12:34:58 -0400 (0:00:00.055) 0:01:07.355 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_name": "quadlet-demo", "__podman_quadlet_type": "kube", "__podman_rootless": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57 Saturday 24 August 2024 12:34:58 -0400 (0:00:00.071) 0:01:07.426 ******* included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Saturday 24 August 2024 12:34:58 -0400 (0:00:00.107) 0:01:07.534 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Saturday 24 August 2024 12:34:58 -0400 (0:00:00.059) 0:01:07.594 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Saturday 24 August 2024 12:34:59 -0400 (0:00:00.161) 0:01:07.756 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get group information] **************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:28 Saturday 24 August 2024 12:34:59 -0400 (0:00:00.071) 0:01:07.827 ******* ok: [managed_node1] => { "ansible_facts": { "getent_group": { "root": [ "x", "0", "" ] } }, "changed": false } TASK [fedora.linux_system_roles.podman : Set group name] *********************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:35 Saturday 24 August 2024 12:34:59 -0400 (0:00:00.405) 0:01:08.233 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_group_name": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:39 Saturday 24 August 2024 12:34:59 -0400 (0:00:00.040) 0:01:08.274 ******* ok: [managed_node1] => { "changed": false, "stat": { "atime": 1724516929.7168462, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "bb5b46ffbafcaa8c4021f3c8b3cb8594f48ef34b", "ctime": 1724516896.725572, "dev": 51713, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 6986657, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-sharedlib", "mode": "0755", "mtime": 1700557386.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 12640, "uid": 0, "version": "1255679238", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************ task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:50 Saturday 24 August 2024 12:35:00 -0400 (0:00:00.408) 0:01:08.682 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check group with getsubids] *********** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:55 Saturday 24 August 2024 12:35:00 -0400 (0:00:00.055) 0:01:08.738 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:60 Saturday 24 August 2024 12:35:00 -0400 (0:00:00.052) 0:01:08.790 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subuid file] ********************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:74 Saturday 24 August 2024 12:35:00 -0400 (0:00:00.055) 0:01:08.846 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subgid file] ********************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:79 Saturday 24 August 2024 12:35:00 -0400 (0:00:00.041) 0:01:08.887 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:84 Saturday 24 August 2024 12:35:00 -0400 (0:00:00.036) 0:01:08.924 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:94 Saturday 24 August 2024 12:35:00 -0400 (0:00:00.037) 0:01:08.962 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] ***** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:101 Saturday 24 August 2024 12:35:00 -0400 (0:00:00.043) 0:01:09.005 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62 Saturday 24 August 2024 12:35:00 -0400 (0:00:00.038) 0:01:09.043 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_activate_systemd_unit": true, "__podman_images_found": [], "__podman_kube_yamls_raw": [ "quadlet-demo.yml" ], "__podman_service_name": "quadlet-demo.service", "__podman_systemd_scope": "system", "__podman_user_home_dir": "/root", "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73 Saturday 24 August 2024 12:35:00 -0400 (0:00:00.057) 0:01:09.100 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_path": "/etc/containers/systemd" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get kube yaml contents] *************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77 Saturday 24 August 2024 12:35:00 -0400 (0:00:00.033) 0:01:09.134 ******* ok: [managed_node1] => { "changed": false, "content": "LS0tCmFwaVZlcnNpb246IHYxCmtpbmQ6IFBlcnNpc3RlbnRWb2x1bWVDbGFpbQptZXRhZGF0YToKICBuYW1lOiB3cC1wdi1jbGFpbQogIGxhYmVsczoKICAgIGFwcDogd29yZHByZXNzCnNwZWM6CiAgYWNjZXNzTW9kZXM6CiAgLSBSZWFkV3JpdGVPbmNlCiAgcmVzb3VyY2VzOgogICAgcmVxdWVzdHM6CiAgICAgIHN0b3JhZ2U6IDIwR2kKLS0tCmFwaVZlcnNpb246IHYxCmtpbmQ6IFBvZAptZXRhZGF0YToKICBuYW1lOiBxdWFkbGV0LWRlbW8Kc3BlYzoKICBjb250YWluZXJzOgogIC0gbmFtZTogd29yZHByZXNzCiAgICBpbWFnZTogcXVheS5pby9saW51eC1zeXN0ZW0tcm9sZXMvd29yZHByZXNzOjQuOC1hcGFjaGUKICAgIGVudjoKICAgIC0gbmFtZTogV09SRFBSRVNTX0RCX0hPU1QKICAgICAgdmFsdWU6IHF1YWRsZXQtZGVtby1teXNxbAogICAgLSBuYW1lOiBXT1JEUFJFU1NfREJfUEFTU1dPUkQKICAgICAgdmFsdWVGcm9tOgogICAgICAgIHNlY3JldEtleVJlZjoKICAgICAgICAgIG5hbWU6IG15c3FsLXJvb3QtcGFzc3dvcmQta3ViZQogICAgICAgICAga2V5OiBwYXNzd29yZAogICAgdm9sdW1lTW91bnRzOgogICAgLSBuYW1lOiB3b3JkcHJlc3MtcGVyc2lzdGVudC1zdG9yYWdlCiAgICAgIG1vdW50UGF0aDogL3Zhci93d3cvaHRtbAogICAgcmVzb3VyY2VzOgogICAgICByZXF1ZXN0czoKICAgICAgICBtZW1vcnk6ICI2NE1pIgogICAgICAgIGNwdTogIjI1MG0iCiAgICAgIGxpbWl0czoKICAgICAgICBtZW1vcnk6ICIxMjhNaSIKICAgICAgICBjcHU6ICI1MDBtIgogIC0gbmFtZTogZW52b3kKICAgIGltYWdlOiBxdWF5LmlvL2xpbnV4LXN5c3RlbS1yb2xlcy9lbnZveXByb3h5OnYxLjI1LjAKICAgIHZvbHVtZU1vdW50czoKICAgIC0gbmFtZTogY29uZmlnLXZvbHVtZQogICAgICBtb3VudFBhdGg6IC9ldGMvZW52b3kKICAgIC0gbmFtZTogY2VydGlmaWNhdGVzCiAgICAgIG1vdW50UGF0aDogL2V0Yy9lbnZveS1jZXJ0aWZpY2F0ZXMKICAgIGVudjoKICAgIC0gbmFtZTogRU5WT1lfVUlECiAgICAgIHZhbHVlOiAiMCIKICAgIHJlc291cmNlczoKICAgICAgcmVxdWVzdHM6CiAgICAgICAgbWVtb3J5OiAiNjRNaSIKICAgICAgICBjcHU6ICIyNTBtIgogICAgICBsaW1pdHM6CiAgICAgICAgbWVtb3J5OiAiMTI4TWkiCiAgICAgICAgY3B1OiAiNTAwbSIKICB2b2x1bWVzOgogIC0gbmFtZTogY29uZmlnLXZvbHVtZQogICAgY29uZmlnTWFwOgogICAgICBuYW1lOiBlbnZveS1wcm94eS1jb25maWcKICAtIG5hbWU6IGNlcnRpZmljYXRlcwogICAgc2VjcmV0OgogICAgICBzZWNyZXROYW1lOiBlbnZveS1jZXJ0aWZpY2F0ZXMKICAtIG5hbWU6IHdvcmRwcmVzcy1wZXJzaXN0ZW50LXN0b3JhZ2UKICAgIHBlcnNpc3RlbnRWb2x1bWVDbGFpbToKICAgICAgY2xhaW1OYW1lOiB3cC1wdi1jbGFpbQogIC0gbmFtZTogd3d3ICAjIG5vdCB1c2VkIC0gZm9yIHRlc3RpbmcgaG9zdHBhdGgKICAgIGhvc3RQYXRoOgogICAgICBwYXRoOiAvdG1wL2h0dHBkMwogIC0gbmFtZTogY3JlYXRlICAjIG5vdCB1c2VkIC0gZm9yIHRlc3RpbmcgaG9zdHBhdGgKICAgIGhvc3RQYXRoOgogICAgICBwYXRoOiAvdG1wL2h0dHBkMy1jcmVhdGUK", "encoding": "base64", "source": "/etc/containers/systemd/quadlet-demo.yml" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87 Saturday 24 August 2024 12:35:00 -0400 (0:00:00.429) 0:01:09.563 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_images": [ "quay.io/linux-system-roles/wordpress:4.8-apache", "quay.io/linux-system-roles/envoyproxy:v1.25.0" ], "__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo.kube", "__podman_volumes": [ "/tmp/httpd3-create", "/tmp/httpd3" ] }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 6] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:103 Saturday 24 August 2024 12:35:01 -0400 (0:00:00.195) 0:01:09.759 ******* ok: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Cleanup quadlets] ********************* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:110 Saturday 24 August 2024 12:35:01 -0400 (0:00:00.057) 0:01:09.817 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create and update quadlets] *********** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:114 Saturday 24 August 2024 12:35:01 -0400 (0:00:00.054) 0:01:09.871 ******* included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:2 Saturday 24 August 2024 12:35:01 -0400 (0:00:00.115) 0:01:09.986 ******* included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Saturday 24 August 2024 12:35:01 -0400 (0:00:00.095) 0:01:10.081 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Saturday 24 August 2024 12:35:01 -0400 (0:00:00.036) 0:01:10.118 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Saturday 24 August 2024 12:35:01 -0400 (0:00:00.041) 0:01:10.160 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create host directories] ************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:7 Saturday 24 August 2024 12:35:01 -0400 (0:00:00.042) 0:01:10.202 ******* [WARNING]: Using a variable for a task's 'args' is unsafe in some situations (see https://docs.ansible.com/ansible/devel/reference_appendices/faq.html#argsplat- unsafe) changed: [managed_node1] => (item=/tmp/httpd3-create) => { "ansible_loop_var": "item", "changed": true, "gid": 0, "group": "root", "item": "/tmp/httpd3-create", "mode": "0755", "owner": "root", "path": "/tmp/httpd3-create", "secontext": "unconfined_u:object_r:user_tmp_t:s0", "size": 6, "state": "directory", "uid": 0 } changed: [managed_node1] => (item=/tmp/httpd3) => { "ansible_loop_var": "item", "changed": true, "gid": 0, "group": "root", "item": "/tmp/httpd3", "mode": "0755", "owner": "root", "path": "/tmp/httpd3", "secontext": "unconfined_u:object_r:user_tmp_t:s0", "size": 6, "state": "directory", "uid": 0 } TASK [fedora.linux_system_roles.podman : Ensure container images are present] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:18 Saturday 24 August 2024 12:35:02 -0400 (0:00:00.733) 0:01:10.936 ******* changed: [managed_node1] => (item=None) => { "attempts": 1, "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": true } changed: [managed_node1] => (item=None) => { "attempts": 1, "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": true } changed: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": true } TASK [fedora.linux_system_roles.podman : Ensure the quadlet directory is present] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:39 Saturday 24 August 2024 12:35:20 -0400 (0:00:18.295) 0:01:29.231 ******* ok: [managed_node1] => { "changed": false, "gid": 0, "group": "root", "mode": "0755", "owner": "root", "path": "/etc/containers/systemd", "secontext": "system_u:object_r:etc_t:s0", "size": 160, "state": "directory", "uid": 0 } TASK [fedora.linux_system_roles.podman : Ensure quadlet file is copied] ******** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:48 Saturday 24 August 2024 12:35:20 -0400 (0:00:00.379) 0:01:29.611 ******* changed: [managed_node1] => { "changed": true, "checksum": "7a5c73a5d935a42431c87bcdbeb8a04ed0909dc7", "dest": "/etc/containers/systemd/quadlet-demo.kube", "gid": 0, "group": "root", "md5sum": "da53c88f92b68b0487aa209f795b6bb3", "mode": "0644", "owner": "root", "secontext": "system_u:object_r:etc_t:s0", "size": 456, "src": "/root/.ansible/tmp/ansible-tmp-1724517321.03834-31464-224102551766944/source", "state": "file", "uid": 0 } TASK [fedora.linux_system_roles.podman : Ensure quadlet file content is present] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:58 Saturday 24 August 2024 12:35:21 -0400 (0:00:00.691) 0:01:30.302 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Ensure quadlet file is present] ******* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:70 Saturday 24 August 2024 12:35:21 -0400 (0:00:00.047) 0:01:30.350 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Reload systemctl] ********************* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:82 Saturday 24 August 2024 12:35:21 -0400 (0:00:00.038) 0:01:30.388 ******* ok: [managed_node1] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.podman : Start service] ************************ task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:110 Saturday 24 August 2024 12:35:22 -0400 (0:00:00.599) 0:01:30.988 ******* changed: [managed_node1] => { "changed": true, "name": "quadlet-demo.service", "state": "started", "status": { "ActiveEnterTimestampMonotonic": "0", "ActiveExitTimestampMonotonic": "0", "ActiveState": "inactive", "After": "quadlet-demo-mysql.service basic.target -.mount sysinit.target system.slice quadlet-demo-network.service systemd-journald.socket", "AllowIsolate": "no", "AllowedCPUs": "", "AllowedMemoryNodes": "", "AmbientCapabilities": "", "AssertResult": "no", "AssertTimestampMonotonic": "0", "Before": "shutdown.target", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "CPUAccounting": "no", "CPUAffinity": "", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "[not set]", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "no", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf", "CollectMode": "inactive", "ConditionResult": "no", "ConditionTimestampMonotonic": "0", "ConfigurationDirectoryMode": "0755", "Conflicts": "shutdown.target", "ControlPID": "0", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "Delegate": "no", "Description": "quadlet-demo.service", "DevicePolicy": "auto", "DynamicUser": "no", "EffectiveCPUs": "", "EffectiveMemoryNodes": "", "Environment": "PODMAN_SYSTEMD_UNIT=quadlet-demo.service", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "0", "ExecMainStartTimestampMonotonic": "0", "ExecMainStatus": "0", "ExecStart": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman kube play --replace --service-container=true --network=systemd-quadlet-demo --configmap /etc/containers/systemd/envoy-proxy-configmap.yml --publish 8000:8080 --publish 9000:9901 /etc/containers/systemd/quadlet-demo.yml ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStopPost": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman kube down /etc/containers/systemd/quadlet-demo.yml ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FragmentPath": "/run/systemd/generator/quadlet-demo.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOSchedulingClass": "0", "IOSchedulingPriority": "0", "IOWeight": "[not set]", "IPAccounting": "no", "IPEgressBytes": "18446744073709551615", "IPEgressPackets": "18446744073709551615", "IPIngressBytes": "18446744073709551615", "IPIngressPackets": "18446744073709551615", "Id": "quadlet-demo.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestampMonotonic": "0", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "mixed", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "0", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "65536", "LimitMEMLOCKSoft": "65536", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "262144", "LimitNOFILESoft": "1024", "LimitNPROC": "14003", "LimitNPROCSoft": "14003", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "14003", "LimitSIGPENDINGSoft": "14003", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "0", "MemoryAccounting": "yes", "MemoryCurrent": "[not set]", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemorySwapMax": "infinity", "MountAPIVFS": "no", "MountFlags": "", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAMask": "", "NUMAPolicy": "n/a", "Names": "quadlet-demo.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "all", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "PermissionsStartOnly": "no", "Perpetual": "no", "PrivateDevices": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "RemainAfterExit": "no", "RemoveIPC": "no", "Requires": "quadlet-demo-network.service system.slice sysinit.target quadlet-demo-mysql.service -.mount", "RequiresMountsFor": "/run/containers", "Restart": "no", "RestartUSec": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "Slice": "system.slice", "SourcePath": "/etc/containers/systemd/quadlet-demo.kube", "StandardError": "inherit", "StandardInput": "null", "StandardInputData": "", "StandardOutput": "journal", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StateChangeTimestampMonotonic": "0", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "dead", "SuccessAction": "none", "SyslogFacility": "3", "SyslogIdentifier": "quadlet-demo", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "0", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "[not set]", "TasksMax": "22405", "TimeoutStartUSec": "1min 30s", "TimeoutStopUSec": "1min 30s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "notify", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "disabled", "UnitFileState": "generated", "UtmpMode": "init", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "0" } } TASK [fedora.linux_system_roles.podman : Restart service] ********************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:125 Saturday 24 August 2024 12:35:23 -0400 (0:00:01.479) 0:01:32.467 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Cancel linger] ************************ task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:167 Saturday 24 August 2024 12:35:23 -0400 (0:00:00.034) 0:01:32.501 ******* TASK [fedora.linux_system_roles.podman : Handle credential files - absent] ***** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:173 Saturday 24 August 2024 12:35:23 -0400 (0:00:00.029) 0:01:32.531 ******* skipping: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Handle certs.d files - absent] ******** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:182 Saturday 24 August 2024 12:35:23 -0400 (0:00:00.030) 0:01:32.561 ******* skipping: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [Check quadlet files] ***************************************************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_demo.yml:96 Saturday 24 August 2024 12:35:23 -0400 (0:00:00.029) 0:01:32.590 ******* ok: [managed_node1] => { "changed": false, "cmd": [ "ls", "-alrtF", "/etc/containers/systemd" ], "delta": "0:00:00.005668", "end": "2024-08-24 12:35:24.342396", "rc": 0, "start": "2024-08-24 12:35:24.336728" } STDOUT: total 24 drwxr-xr-x. 8 root root 182 Aug 24 12:30 ../ -rw-r--r--. 1 root root 74 Aug 24 12:34 quadlet-demo.network -rw-r--r--. 1 root root 9 Aug 24 12:34 quadlet-demo-mysql.volume -rw-r--r--. 1 root root 363 Aug 24 12:34 quadlet-demo-mysql.container -rw-r--r--. 1 root root 2102 Aug 24 12:34 envoy-proxy-configmap.yml -rw-r--r--. 1 root root 1605 Aug 24 12:34 quadlet-demo.yml -rw-r--r--. 1 root root 456 Aug 24 12:35 quadlet-demo.kube drwxr-xr-x. 2 root root 185 Aug 24 12:35 ./ TASK [Check containers] ******************************************************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_demo.yml:100 Saturday 24 August 2024 12:35:24 -0400 (0:00:00.441) 0:01:33.032 ******* ok: [managed_node1] => { "changed": false, "cmd": [ "podman", "ps", "-a" ], "delta": "0:00:00.051368", "end": "2024-08-24 12:35:24.808241", "failed_when_result": false, "rc": 0, "start": "2024-08-24 12:35:24.756873" } STDOUT: CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES 68250d5fc021 quay.io/linux-system-roles/mysql:5.6 mysqld 36 seconds ago Up 36 seconds (healthy) quadlet-demo-mysql a3a21fc2dc26 localhost/podman-pause:4.9.4-dev-1708535009 1 second ago Up 1 second a96f3a51b8d1-service c14ac2390f14 localhost/podman-pause:4.9.4-dev-1708535009 1 second ago Up 1 second 0.0.0.0:8000->8080/tcp, 0.0.0.0:9000->9901/tcp bbca104112c4-infra 3ecc5b442f00 quay.io/linux-system-roles/wordpress:4.8-apache apache2-foregroun... 1 second ago Up 1 second 0.0.0.0:8000->8080/tcp, 0.0.0.0:9000->9901/tcp quadlet-demo-wordpress 14fe7e5a1bc4 quay.io/linux-system-roles/envoyproxy:v1.25.0 envoy -c /etc/env... 1 second ago Up 1 second 0.0.0.0:8000->8080/tcp, 0.0.0.0:9000->9901/tcp quadlet-demo-envoy TASK [Check volumes] *********************************************************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_demo.yml:105 Saturday 24 August 2024 12:35:24 -0400 (0:00:00.463) 0:01:33.496 ******* ok: [managed_node1] => { "changed": false, "cmd": [ "podman", "volume", "ls" ], "delta": "0:00:00.028920", "end": "2024-08-24 12:35:25.231088", "failed_when_result": false, "rc": 0, "start": "2024-08-24 12:35:25.202168" } STDOUT: DRIVER VOLUME NAME local systemd-quadlet-demo-mysql local wp-pv-claim local envoy-proxy-config local envoy-certificates TASK [Check pods] ************************************************************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_demo.yml:110 Saturday 24 August 2024 12:35:25 -0400 (0:00:00.424) 0:01:33.920 ******* ok: [managed_node1] => { "changed": false, "cmd": [ "podman", "pod", "ps", "--ctr-ids", "--ctr-names", "--ctr-status" ], "delta": "0:00:00.037163", "end": "2024-08-24 12:35:25.657541", "failed_when_result": false, "rc": 0, "start": "2024-08-24 12:35:25.620378" } STDOUT: POD ID NAME STATUS CREATED INFRA ID IDS NAMES STATUS bbca104112c4 quadlet-demo Running 2 seconds ago c14ac2390f14 c14ac2390f14,3ecc5b442f00,14fe7e5a1bc4 bbca104112c4-infra,quadlet-demo-wordpress,quadlet-demo-envoy running,running,running TASK [Check systemd] *********************************************************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_demo.yml:115 Saturday 24 August 2024 12:35:25 -0400 (0:00:00.416) 0:01:34.337 ******* ok: [managed_node1] => { "changed": false, "cmd": "set -euo pipefail; systemctl list-units | grep quadlet", "delta": "0:00:00.011411", "end": "2024-08-24 12:35:26.031591", "failed_when_result": false, "rc": 0, "start": "2024-08-24 12:35:26.020180" } STDOUT: quadlet-demo-mysql-volume.service loaded active exited quadlet-demo-mysql-volume.service quadlet-demo-mysql.service loaded active running quadlet-demo-mysql.service quadlet-demo-network.service loaded active exited quadlet-demo-network.service quadlet-demo.service loaded active running quadlet-demo.service TASK [Check web] *************************************************************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_demo.yml:121 Saturday 24 August 2024 12:35:26 -0400 (0:00:00.373) 0:01:34.711 ******* changed: [managed_node1] => { "attempts": 1, "changed": true, "checksum_dest": null, "checksum_src": "9dbc8c154daaed74c52aa781ab9994c0246a012f", "dest": "/run/out", "elapsed": 0, "gid": 0, "group": "root", "md5sum": "f59bccb13786ba6aa549360a06d6b1b4", "mode": "0600", "owner": "root", "secontext": "system_u:object_r:var_run_t:s0", "size": 11666, "src": "/root/.ansible/tmp/ansible-tmp-1724517326.132896-31692-693845049303/tmps517tyl0", "state": "file", "status_code": 200, "uid": 0, "url": "https://localhost:8000" } MSG: OK (unknown bytes) TASK [Show web] **************************************************************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_demo.yml:132 Saturday 24 August 2024 12:35:27 -0400 (0:00:01.083) 0:01:35.794 ******* ok: [managed_node1] => { "changed": false, "cmd": [ "cat", "/run/out" ], "delta": "0:00:00.002511", "end": "2024-08-24 12:35:27.473382", "rc": 0, "start": "2024-08-24 12:35:27.470871" } STDOUT: WordPress › Installation

TASK [Error] ******************************************************************* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_demo.yml:137 Saturday 24 August 2024 12:35:27 -0400 (0:00:00.360) 0:01:36.155 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check] ******************************************************************* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_demo.yml:148 Saturday 24 August 2024 12:35:27 -0400 (0:00:00.030) 0:01:36.186 ******* ok: [managed_node1] => { "changed": false, "cmd": [ "podman", "ps", "-a" ], "delta": "0:00:00.036747", "end": "2024-08-24 12:35:27.893535", "rc": 0, "start": "2024-08-24 12:35:27.856788" } STDOUT: CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES 68250d5fc021 quay.io/linux-system-roles/mysql:5.6 mysqld 39 seconds ago Up 39 seconds (healthy) quadlet-demo-mysql a3a21fc2dc26 localhost/podman-pause:4.9.4-dev-1708535009 5 seconds ago Up 4 seconds a96f3a51b8d1-service c14ac2390f14 localhost/podman-pause:4.9.4-dev-1708535009 4 seconds ago Up 4 seconds 0.0.0.0:8000->8080/tcp, 0.0.0.0:9000->9901/tcp bbca104112c4-infra 3ecc5b442f00 quay.io/linux-system-roles/wordpress:4.8-apache apache2-foregroun... 4 seconds ago Up 4 seconds 0.0.0.0:8000->8080/tcp, 0.0.0.0:9000->9901/tcp quadlet-demo-wordpress 14fe7e5a1bc4 quay.io/linux-system-roles/envoyproxy:v1.25.0 envoy -c /etc/env... 4 seconds ago Up 4 seconds 0.0.0.0:8000->8080/tcp, 0.0.0.0:9000->9901/tcp quadlet-demo-envoy TASK [Check pods] ************************************************************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_demo.yml:152 Saturday 24 August 2024 12:35:27 -0400 (0:00:00.410) 0:01:36.597 ******* ok: [managed_node1] => { "changed": false, "cmd": [ "podman", "pod", "ps", "--ctr-ids", "--ctr-names", "--ctr-status" ], "delta": "0:00:00.036935", "end": "2024-08-24 12:35:28.336457", "failed_when_result": false, "rc": 0, "start": "2024-08-24 12:35:28.299522" } STDOUT: POD ID NAME STATUS CREATED INFRA ID IDS NAMES STATUS bbca104112c4 quadlet-demo Running 5 seconds ago c14ac2390f14 c14ac2390f14,3ecc5b442f00,14fe7e5a1bc4 bbca104112c4-infra,quadlet-demo-wordpress,quadlet-demo-envoy running,running,running TASK [Check systemd] *********************************************************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_demo.yml:157 Saturday 24 August 2024 12:35:28 -0400 (0:00:00.427) 0:01:37.025 ******* ok: [managed_node1] => { "changed": false, "cmd": "set -euo pipefail; systemctl list-units --all | grep quadlet", "delta": "0:00:00.010186", "end": "2024-08-24 12:35:28.707551", "failed_when_result": false, "rc": 0, "start": "2024-08-24 12:35:28.697365" } STDOUT: quadlet-demo-mysql-volume.service loaded active exited quadlet-demo-mysql-volume.service quadlet-demo-mysql.service loaded active running quadlet-demo-mysql.service quadlet-demo-network.service loaded active exited quadlet-demo-network.service quadlet-demo.service loaded active running quadlet-demo.service TASK [LS] ********************************************************************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_demo.yml:165 Saturday 24 August 2024 12:35:28 -0400 (0:00:00.379) 0:01:37.404 ******* ok: [managed_node1] => { "changed": false, "cmd": [ "ls", "-alrtF", "/etc/systemd/system" ], "delta": "0:00:00.003509", "end": "2024-08-24 12:35:29.082531", "failed_when_result": false, "rc": 0, "start": "2024-08-24 12:35:29.079022" } STDOUT: total 8 lrwxrwxrwx. 1 root root 9 May 11 2019 systemd-timedated.service -> /dev/null drwxr-xr-x. 4 root root 169 May 29 03:45 ../ lrwxrwxrwx. 1 root root 39 May 29 03:45 syslog.service -> /usr/lib/systemd/system/rsyslog.service drwxr-xr-x. 2 root root 32 May 29 03:45 getty.target.wants/ lrwxrwxrwx. 1 root root 37 May 29 03:45 ctrl-alt-del.target -> /usr/lib/systemd/system/reboot.target lrwxrwxrwx. 1 root root 57 May 29 03:45 dbus-org.freedesktop.nm-dispatcher.service -> /usr/lib/systemd/system/NetworkManager-dispatcher.service drwxr-xr-x. 2 root root 48 May 29 03:45 network-online.target.wants/ lrwxrwxrwx. 1 root root 41 May 29 03:45 dbus-org.freedesktop.timedate1.service -> /usr/lib/systemd/system/timedatex.service drwxr-xr-x. 2 root root 61 May 29 03:45 timers.target.wants/ drwxr-xr-x. 2 root root 31 May 29 03:45 basic.target.wants/ drwxr-xr-x. 2 root root 38 May 29 03:45 dev-virtio\x2dports-org.qemu.guest_agent.0.device.wants/ lrwxrwxrwx. 1 root root 41 May 29 03:47 default.target -> /usr/lib/systemd/system/multi-user.target drwxr-xr-x. 2 root root 51 May 29 03:55 sockets.target.wants/ drwxr-xr-x. 2 root root 31 May 29 03:55 remote-fs.target.wants/ drwxr-xr-x. 2 root root 59 May 29 03:55 sshd-keygen@.service.d/ drwxr-xr-x. 2 root root 119 May 29 03:55 cloud-init.target.wants/ drwxr-xr-x. 2 root root 181 May 29 03:56 sysinit.target.wants/ drwxr-xr-x. 2 root root 4096 Aug 24 12:34 multi-user.target.wants/ lrwxrwxrwx. 1 root root 41 Aug 24 12:34 dbus-org.fedoraproject.FirewallD1.service -> /usr/lib/systemd/system/firewalld.service drwxr-xr-x. 13 root root 4096 Aug 24 12:34 ./ TASK [Cleanup] ***************************************************************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_demo.yml:172 Saturday 24 August 2024 12:35:29 -0400 (0:00:00.373) 0:01:37.778 ******* TASK [fedora.linux_system_roles.podman : Set platform/version specific variables] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:3 Saturday 24 August 2024 12:35:29 -0400 (0:00:00.110) 0:01:37.888 ******* included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Ensure ansible_facts used by role] **** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:3 Saturday 24 August 2024 12:35:29 -0400 (0:00:00.089) 0:01:37.977 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if system is ostree] ************ task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:11 Saturday 24 August 2024 12:35:29 -0400 (0:00:00.063) 0:01:38.041 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set flag to indicate system is ostree] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:16 Saturday 24 August 2024 12:35:29 -0400 (0:00:00.053) 0:01:38.094 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if transactional-update exists in /sbin] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:23 Saturday 24 August 2024 12:35:29 -0400 (0:00:00.051) 0:01:38.146 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set flag if transactional-update exists] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:28 Saturday 24 August 2024 12:35:29 -0400 (0:00:00.052) 0:01:38.198 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set platform/version specific variables] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:32 Saturday 24 August 2024 12:35:29 -0400 (0:00:00.051) 0:01:38.250 ******* ok: [managed_node1] => (item=RedHat.yml) => { "ansible_facts": { "__podman_packages": [ "podman", "shadow-utils-subid" ] }, "ansible_included_var_files": [ "/tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/vars/RedHat.yml" ], "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml" } skipping: [managed_node1] => (item=CentOS.yml) => { "ansible_loop_var": "item", "changed": false, "item": "CentOS.yml", "skip_reason": "Conditional result was False" } ok: [managed_node1] => (item=CentOS_8.yml) => { "ansible_facts": { "__podman_packages": [ "crun", "podman", "podman-plugins", "shadow-utils-subid" ] }, "ansible_included_var_files": [ "/tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/vars/CentOS_8.yml" ], "ansible_loop_var": "item", "changed": false, "item": "CentOS_8.yml" } ok: [managed_node1] => (item=CentOS_8.yml) => { "ansible_facts": { "__podman_packages": [ "crun", "podman", "podman-plugins", "shadow-utils-subid" ] }, "ansible_included_var_files": [ "/tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/vars/CentOS_8.yml" ], "ansible_loop_var": "item", "changed": false, "item": "CentOS_8.yml" } TASK [fedora.linux_system_roles.podman : Gather the package facts] ************* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:6 Saturday 24 August 2024 12:35:29 -0400 (0:00:00.126) 0:01:38.376 ******* ok: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Enable copr if requested] ************* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:10 Saturday 24 August 2024 12:35:31 -0400 (0:00:01.508) 0:01:39.884 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Ensure required packages are installed] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:14 Saturday 24 August 2024 12:35:31 -0400 (0:00:00.036) 0:01:39.921 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Notify user that reboot is needed to apply changes] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:28 Saturday 24 August 2024 12:35:31 -0400 (0:00:00.046) 0:01:39.967 ******* skipping: [managed_node1] => {} TASK [fedora.linux_system_roles.podman : Reboot transactional update systems] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:33 Saturday 24 August 2024 12:35:31 -0400 (0:00:00.031) 0:01:39.998 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if reboot is needed and not set] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:38 Saturday 24 August 2024 12:35:31 -0400 (0:00:00.029) 0:01:40.028 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get podman version] ******************* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:46 Saturday 24 August 2024 12:35:31 -0400 (0:00:00.031) 0:01:40.059 ******* ok: [managed_node1] => { "changed": false, "cmd": [ "podman", "--version" ], "delta": "0:00:00.026652", "end": "2024-08-24 12:35:31.780172", "rc": 0, "start": "2024-08-24 12:35:31.753520" } STDOUT: podman version 4.9.4-dev TASK [fedora.linux_system_roles.podman : Set podman version] ******************* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:52 Saturday 24 August 2024 12:35:31 -0400 (0:00:00.403) 0:01:40.462 ******* ok: [managed_node1] => { "ansible_facts": { "podman_version": "4.9.4-dev" }, "changed": false } TASK [fedora.linux_system_roles.podman : Podman package version must be 4.2 or later] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:56 Saturday 24 August 2024 12:35:31 -0400 (0:00:00.033) 0:01:40.496 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Podman package version must be 4.4 or later for quadlet, secrets] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:63 Saturday 24 August 2024 12:35:31 -0400 (0:00:00.030) 0:01:40.527 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } META: end_host conditional evaluated to false, continuing execution for managed_node1 TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:80 Saturday 24 August 2024 12:35:31 -0400 (0:00:00.058) 0:01:40.585 ******* included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Saturday 24 August 2024 12:35:32 -0400 (0:00:00.126) 0:01:40.712 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Saturday 24 August 2024 12:35:32 -0400 (0:00:00.051) 0:01:40.764 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Saturday 24 August 2024 12:35:32 -0400 (0:00:00.053) 0:01:40.817 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get group information] **************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:28 Saturday 24 August 2024 12:35:32 -0400 (0:00:00.051) 0:01:40.868 ******* ok: [managed_node1] => { "ansible_facts": { "getent_group": { "root": [ "x", "0", "" ] } }, "changed": false } TASK [fedora.linux_system_roles.podman : Set group name] *********************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:35 Saturday 24 August 2024 12:35:32 -0400 (0:00:00.373) 0:01:41.241 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_group_name": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:39 Saturday 24 August 2024 12:35:32 -0400 (0:00:00.053) 0:01:41.295 ******* ok: [managed_node1] => { "changed": false, "stat": { "atime": 1724516929.7168462, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "bb5b46ffbafcaa8c4021f3c8b3cb8594f48ef34b", "ctime": 1724516896.725572, "dev": 51713, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 6986657, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-sharedlib", "mode": "0755", "mtime": 1700557386.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 12640, "uid": 0, "version": "1255679238", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************ task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:50 Saturday 24 August 2024 12:35:33 -0400 (0:00:00.374) 0:01:41.670 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check group with getsubids] *********** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:55 Saturday 24 August 2024 12:35:33 -0400 (0:00:00.039) 0:01:41.709 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:60 Saturday 24 August 2024 12:35:33 -0400 (0:00:00.051) 0:01:41.761 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subuid file] ********************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:74 Saturday 24 August 2024 12:35:33 -0400 (0:00:00.061) 0:01:41.823 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subgid file] ********************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:79 Saturday 24 August 2024 12:35:33 -0400 (0:00:00.052) 0:01:41.875 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:84 Saturday 24 August 2024 12:35:33 -0400 (0:00:00.049) 0:01:41.925 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:94 Saturday 24 August 2024 12:35:33 -0400 (0:00:00.051) 0:01:41.977 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] ***** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:101 Saturday 24 August 2024 12:35:33 -0400 (0:00:00.036) 0:01:42.013 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set config file paths] **************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:86 Saturday 24 August 2024 12:35:33 -0400 (0:00:00.037) 0:01:42.050 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_container_conf_file": "/etc/containers/containers.conf.d/50-systemroles.conf", "__podman_policy_json_file": "/etc/containers/policy.json", "__podman_registries_conf_file": "/etc/containers/registries.conf.d/50-systemroles.conf", "__podman_storage_conf_file": "/etc/containers/storage.conf" }, "changed": false } TASK [fedora.linux_system_roles.podman : Handle container.conf.d] ************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:95 Saturday 24 August 2024 12:35:33 -0400 (0:00:00.051) 0:01:42.102 ******* included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_container_conf_d.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Ensure containers.d exists] *********** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_container_conf_d.yml:5 Saturday 24 August 2024 12:35:33 -0400 (0:00:00.105) 0:01:42.208 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Update container config file] ********* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_container_conf_d.yml:13 Saturday 24 August 2024 12:35:33 -0400 (0:00:00.031) 0:01:42.239 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Handle registries.conf.d] ************* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:98 Saturday 24 August 2024 12:35:33 -0400 (0:00:00.032) 0:01:42.271 ******* included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_registries_conf_d.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Ensure registries.d exists] *********** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_registries_conf_d.yml:5 Saturday 24 August 2024 12:35:33 -0400 (0:00:00.060) 0:01:42.332 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Update registries config file] ******** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_registries_conf_d.yml:13 Saturday 24 August 2024 12:35:33 -0400 (0:00:00.037) 0:01:42.370 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Handle storage.conf] ****************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:101 Saturday 24 August 2024 12:35:33 -0400 (0:00:00.050) 0:01:42.421 ******* included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_storage_conf.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Ensure storage.conf parent dir exists] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_storage_conf.yml:5 Saturday 24 August 2024 12:35:33 -0400 (0:00:00.105) 0:01:42.526 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Update storage config file] *********** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_storage_conf.yml:13 Saturday 24 August 2024 12:35:33 -0400 (0:00:00.050) 0:01:42.577 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Handle policy.json] ******************* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:104 Saturday 24 August 2024 12:35:34 -0400 (0:00:00.057) 0:01:42.634 ******* included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Ensure policy.json parent dir exists] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:6 Saturday 24 August 2024 12:35:34 -0400 (0:00:00.104) 0:01:42.738 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Stat the policy.json file] ************ task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:14 Saturday 24 August 2024 12:35:34 -0400 (0:00:00.037) 0:01:42.776 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get the existing policy.json] ********* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:19 Saturday 24 August 2024 12:35:34 -0400 (0:00:00.045) 0:01:42.822 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Write new policy.json file] *********** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:25 Saturday 24 August 2024 12:35:34 -0400 (0:00:00.041) 0:01:42.863 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Manage firewall for specified ports] ************************************* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:110 Saturday 24 August 2024 12:35:34 -0400 (0:00:00.038) 0:01:42.901 ******* TASK [fedora.linux_system_roles.firewall : Setup firewalld] ******************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:2 Saturday 24 August 2024 12:35:34 -0400 (0:00:00.121) 0:01:43.023 ******* included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml for managed_node1 TASK [fedora.linux_system_roles.firewall : Ensure ansible_facts used by role] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:2 Saturday 24 August 2024 12:35:34 -0400 (0:00:00.057) 0:01:43.081 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Check if system is ostree] ********** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:10 Saturday 24 August 2024 12:35:34 -0400 (0:00:00.047) 0:01:43.128 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Set flag to indicate system is ostree] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:15 Saturday 24 August 2024 12:35:34 -0400 (0:00:00.050) 0:01:43.179 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Check if transactional-update exists in /sbin] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:22 Saturday 24 August 2024 12:35:34 -0400 (0:00:00.041) 0:01:43.220 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Set flag if transactional-update exists] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:27 Saturday 24 August 2024 12:35:34 -0400 (0:00:00.036) 0:01:43.257 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Install firewalld] ****************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:31 Saturday 24 August 2024 12:35:34 -0400 (0:00:00.036) 0:01:43.294 ******* ok: [managed_node1] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do lsrpackages: firewalld TASK [fedora.linux_system_roles.firewall : Notify user that reboot is needed to apply changes] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:43 Saturday 24 August 2024 12:35:37 -0400 (0:00:02.778) 0:01:46.073 ******* skipping: [managed_node1] => {} TASK [fedora.linux_system_roles.firewall : Reboot transactional update systems] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:48 Saturday 24 August 2024 12:35:37 -0400 (0:00:00.032) 0:01:46.105 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Fail if reboot is needed and not set] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:53 Saturday 24 August 2024 12:35:37 -0400 (0:00:00.032) 0:01:46.138 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Collect service facts] ************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:5 Saturday 24 August 2024 12:35:37 -0400 (0:00:00.032) 0:01:46.170 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Attempt to stop and disable conflicting services] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:9 Saturday 24 August 2024 12:35:37 -0400 (0:00:00.030) 0:01:46.201 ******* skipping: [managed_node1] => (item=nftables) => { "ansible_loop_var": "item", "changed": false, "item": "nftables", "skip_reason": "Conditional result was False" } skipping: [managed_node1] => (item=iptables) => { "ansible_loop_var": "item", "changed": false, "item": "iptables", "skip_reason": "Conditional result was False" } skipping: [managed_node1] => (item=ufw) => { "ansible_loop_var": "item", "changed": false, "item": "ufw", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Unmask firewalld service] *********** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:22 Saturday 24 August 2024 12:35:37 -0400 (0:00:00.041) 0:01:46.243 ******* ok: [managed_node1] => { "changed": false, "name": "firewalld", "status": { "ActiveEnterTimestamp": "Sat 2024-08-24 12:34:20 EDT", "ActiveEnterTimestampMonotonic": "663046301", "ActiveExitTimestampMonotonic": "0", "ActiveState": "active", "After": "dbus.service sysinit.target dbus.socket polkit.service system.slice basic.target", "AllowIsolate": "no", "AllowedCPUs": "", "AllowedMemoryNodes": "", "AmbientCapabilities": "", "AssertResult": "yes", "AssertTimestamp": "Sat 2024-08-24 12:34:19 EDT", "AssertTimestampMonotonic": "662544922", "Before": "shutdown.target multi-user.target network-pre.target", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "BusName": "org.fedoraproject.FirewallD1", "CPUAccounting": "no", "CPUAffinity": "", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "[not set]", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "yes", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf", "CollectMode": "inactive", "ConditionResult": "yes", "ConditionTimestamp": "Sat 2024-08-24 12:34:19 EDT", "ConditionTimestampMonotonic": "662544921", "ConfigurationDirectoryMode": "0755", "Conflicts": "shutdown.target ip6tables.service ipset.service ebtables.service iptables.service nftables.service", "ControlGroup": "/system.slice/firewalld.service", "ControlPID": "0", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "Delegate": "no", "Description": "firewalld - dynamic firewall daemon", "DevicePolicy": "auto", "Documentation": "man:firewalld(1)", "DynamicUser": "no", "EffectiveCPUs": "", "EffectiveMemoryNodes": "", "EnvironmentFiles": "/etc/sysconfig/firewalld (ignore_errors=yes)", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "80148", "ExecMainStartTimestamp": "Sat 2024-08-24 12:34:19 EDT", "ExecMainStartTimestampMonotonic": "662546362", "ExecMainStatus": "0", "ExecReload": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStart": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FragmentPath": "/usr/lib/systemd/system/firewalld.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOSchedulingClass": "0", "IOSchedulingPriority": "0", "IOWeight": "[not set]", "IPAccounting": "no", "IPEgressBytes": "18446744073709551615", "IPEgressPackets": "18446744073709551615", "IPIngressBytes": "18446744073709551615", "IPIngressPackets": "18446744073709551615", "Id": "firewalld.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestamp": "Sat 2024-08-24 12:34:19 EDT", "InactiveExitTimestampMonotonic": "662546400", "InvocationID": "4700a8e8b9b14f7194db3c9a057ff62b", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "mixed", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "0", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "65536", "LimitMEMLOCKSoft": "65536", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "262144", "LimitNOFILESoft": "1024", "LimitNPROC": "14003", "LimitNPROCSoft": "14003", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "14003", "LimitSIGPENDINGSoft": "14003", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "80148", "MemoryAccounting": "yes", "MemoryCurrent": "39976960", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemorySwapMax": "infinity", "MountAPIVFS": "no", "MountFlags": "", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAMask": "", "NUMAPolicy": "n/a", "Names": "firewalld.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "PermissionsStartOnly": "no", "Perpetual": "no", "PrivateDevices": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "RemainAfterExit": "no", "RemoveIPC": "no", "Requires": "dbus.socket system.slice sysinit.target", "Restart": "no", "RestartUSec": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "Slice": "system.slice", "StandardError": "null", "StandardInput": "null", "StandardInputData": "", "StandardOutput": "null", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StateChangeTimestamp": "Sat 2024-08-24 12:34:20 EDT", "StateChangeTimestampMonotonic": "663046301", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "running", "SuccessAction": "none", "SyslogFacility": "3", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "0", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "2", "TasksMax": "22405", "TimeoutStartUSec": "1min 30s", "TimeoutStopUSec": "1min 30s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "dbus", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "enabled", "UnitFileState": "enabled", "UtmpMode": "init", "WantedBy": "multi-user.target", "Wants": "network-pre.target", "WatchdogTimestamp": "Sat 2024-08-24 12:34:20 EDT", "WatchdogTimestampMonotonic": "663046298", "WatchdogUSec": "0" } } TASK [fedora.linux_system_roles.firewall : Enable and start firewalld service] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:28 Saturday 24 August 2024 12:35:38 -0400 (0:00:00.545) 0:01:46.788 ******* ok: [managed_node1] => { "changed": false, "enabled": true, "name": "firewalld", "state": "started", "status": { "ActiveEnterTimestamp": "Sat 2024-08-24 12:34:20 EDT", "ActiveEnterTimestampMonotonic": "663046301", "ActiveExitTimestampMonotonic": "0", "ActiveState": "active", "After": "dbus.service sysinit.target dbus.socket polkit.service system.slice basic.target", "AllowIsolate": "no", "AllowedCPUs": "", "AllowedMemoryNodes": "", "AmbientCapabilities": "", "AssertResult": "yes", "AssertTimestamp": "Sat 2024-08-24 12:34:19 EDT", "AssertTimestampMonotonic": "662544922", "Before": "shutdown.target multi-user.target network-pre.target", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "BusName": "org.fedoraproject.FirewallD1", "CPUAccounting": "no", "CPUAffinity": "", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "[not set]", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "yes", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf", "CollectMode": "inactive", "ConditionResult": "yes", "ConditionTimestamp": "Sat 2024-08-24 12:34:19 EDT", "ConditionTimestampMonotonic": "662544921", "ConfigurationDirectoryMode": "0755", "Conflicts": "shutdown.target ip6tables.service ipset.service ebtables.service iptables.service nftables.service", "ControlGroup": "/system.slice/firewalld.service", "ControlPID": "0", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "Delegate": "no", "Description": "firewalld - dynamic firewall daemon", "DevicePolicy": "auto", "Documentation": "man:firewalld(1)", "DynamicUser": "no", "EffectiveCPUs": "", "EffectiveMemoryNodes": "", "EnvironmentFiles": "/etc/sysconfig/firewalld (ignore_errors=yes)", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "80148", "ExecMainStartTimestamp": "Sat 2024-08-24 12:34:19 EDT", "ExecMainStartTimestampMonotonic": "662546362", "ExecMainStatus": "0", "ExecReload": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStart": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FragmentPath": "/usr/lib/systemd/system/firewalld.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOSchedulingClass": "0", "IOSchedulingPriority": "0", "IOWeight": "[not set]", "IPAccounting": "no", "IPEgressBytes": "18446744073709551615", "IPEgressPackets": "18446744073709551615", "IPIngressBytes": "18446744073709551615", "IPIngressPackets": "18446744073709551615", "Id": "firewalld.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestamp": "Sat 2024-08-24 12:34:19 EDT", "InactiveExitTimestampMonotonic": "662546400", "InvocationID": "4700a8e8b9b14f7194db3c9a057ff62b", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "mixed", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "0", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "65536", "LimitMEMLOCKSoft": "65536", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "262144", "LimitNOFILESoft": "1024", "LimitNPROC": "14003", "LimitNPROCSoft": "14003", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "14003", "LimitSIGPENDINGSoft": "14003", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "80148", "MemoryAccounting": "yes", "MemoryCurrent": "39976960", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemorySwapMax": "infinity", "MountAPIVFS": "no", "MountFlags": "", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAMask": "", "NUMAPolicy": "n/a", "Names": "firewalld.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "PermissionsStartOnly": "no", "Perpetual": "no", "PrivateDevices": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "RemainAfterExit": "no", "RemoveIPC": "no", "Requires": "dbus.socket system.slice sysinit.target", "Restart": "no", "RestartUSec": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "Slice": "system.slice", "StandardError": "null", "StandardInput": "null", "StandardInputData": "", "StandardOutput": "null", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StateChangeTimestamp": "Sat 2024-08-24 12:34:20 EDT", "StateChangeTimestampMonotonic": "663046301", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "running", "SuccessAction": "none", "SyslogFacility": "3", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "0", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "2", "TasksMax": "22405", "TimeoutStartUSec": "1min 30s", "TimeoutStopUSec": "1min 30s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "dbus", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "enabled", "UnitFileState": "enabled", "UtmpMode": "init", "WantedBy": "multi-user.target", "Wants": "network-pre.target", "WatchdogTimestamp": "Sat 2024-08-24 12:34:20 EDT", "WatchdogTimestampMonotonic": "663046298", "WatchdogUSec": "0" } } TASK [fedora.linux_system_roles.firewall : Check if previous replaced is defined] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:34 Saturday 24 August 2024 12:35:38 -0400 (0:00:00.518) 0:01:47.307 ******* ok: [managed_node1] => { "ansible_facts": { "__firewall_previous_replaced": false, "__firewall_python_cmd": "/usr/libexec/platform-python", "__firewall_report_changed": true }, "changed": false } TASK [fedora.linux_system_roles.firewall : Get config files, checksums before and remove] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:43 Saturday 24 August 2024 12:35:38 -0400 (0:00:00.044) 0:01:47.351 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Tell firewall module it is able to report changed] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:55 Saturday 24 August 2024 12:35:38 -0400 (0:00:00.039) 0:01:47.391 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Configure firewall] ***************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:71 Saturday 24 August 2024 12:35:38 -0400 (0:00:00.037) 0:01:47.429 ******* ok: [managed_node1] => (item={'port': '8000/tcp', 'state': 'enabled'}) => { "__firewall_changed": false, "ansible_loop_var": "item", "changed": false, "item": { "port": "8000/tcp", "state": "enabled" } } ok: [managed_node1] => (item={'port': '9000/tcp', 'state': 'enabled'}) => { "__firewall_changed": false, "ansible_loop_var": "item", "changed": false, "item": { "port": "9000/tcp", "state": "enabled" } } TASK [fedora.linux_system_roles.firewall : Gather firewall config information] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:120 Saturday 24 August 2024 12:35:39 -0400 (0:00:01.142) 0:01:48.571 ******* skipping: [managed_node1] => (item={'port': '8000/tcp', 'state': 'enabled'}) => { "ansible_loop_var": "item", "changed": false, "item": { "port": "8000/tcp", "state": "enabled" }, "skip_reason": "Conditional result was False" } skipping: [managed_node1] => (item={'port': '9000/tcp', 'state': 'enabled'}) => { "ansible_loop_var": "item", "changed": false, "item": { "port": "9000/tcp", "state": "enabled" }, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Update firewalld_config fact] ******* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:130 Saturday 24 August 2024 12:35:40 -0400 (0:00:00.050) 0:01:48.622 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Gather firewall config if no arguments] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:139 Saturday 24 August 2024 12:35:40 -0400 (0:00:00.034) 0:01:48.657 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Update firewalld_config fact] ******* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:144 Saturday 24 August 2024 12:35:40 -0400 (0:00:00.034) 0:01:48.691 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Get config files, checksums after] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:153 Saturday 24 August 2024 12:35:40 -0400 (0:00:00.033) 0:01:48.724 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Calculate what has changed] ********* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:163 Saturday 24 August 2024 12:35:40 -0400 (0:00:00.032) 0:01:48.756 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Show diffs] ************************* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:169 Saturday 24 August 2024 12:35:40 -0400 (0:00:00.031) 0:01:48.788 ******* skipping: [managed_node1] => {} TASK [Manage selinux for specified ports] ************************************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:117 Saturday 24 August 2024 12:35:40 -0400 (0:00:00.031) 0:01:48.819 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Keep track of users that need to cancel linger] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:124 Saturday 24 August 2024 12:35:40 -0400 (0:00:00.031) 0:01:48.850 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_cancel_user_linger": [] }, "changed": false } TASK [fedora.linux_system_roles.podman : Handle certs.d files - present] ******* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:128 Saturday 24 August 2024 12:35:40 -0400 (0:00:00.033) 0:01:48.884 ******* skipping: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Handle credential files - present] **** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:137 Saturday 24 August 2024 12:35:40 -0400 (0:00:00.029) 0:01:48.914 ******* skipping: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Handle secrets] *********************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:146 Saturday 24 August 2024 12:35:40 -0400 (0:00:00.028) 0:01:48.942 ******* included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml for managed_node1 included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml for managed_node1 included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Set variables part 1] ***************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:3 Saturday 24 August 2024 12:35:40 -0400 (0:00:00.131) 0:01:49.074 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set variables part 2] ***************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:7 Saturday 24 August 2024 12:35:40 -0400 (0:00:00.039) 0:01:49.113 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_rootless": false, "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:13 Saturday 24 August 2024 12:35:40 -0400 (0:00:00.120) 0:01:49.234 ******* included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Saturday 24 August 2024 12:35:40 -0400 (0:00:00.057) 0:01:49.291 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Saturday 24 August 2024 12:35:40 -0400 (0:00:00.032) 0:01:49.324 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Saturday 24 August 2024 12:35:40 -0400 (0:00:00.032) 0:01:49.356 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Stat XDG_RUNTIME_DIR] ***************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:18 Saturday 24 August 2024 12:35:40 -0400 (0:00:00.030) 0:01:49.387 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Manage each secret] ******************* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:34 Saturday 24 August 2024 12:35:40 -0400 (0:00:00.035) 0:01:49.422 ******* [WARNING]: Using a variable for a task's 'args' is unsafe in some situations (see https://docs.ansible.com/ansible/devel/reference_appendices/faq.html#argsplat- unsafe) changed: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": true } TASK [fedora.linux_system_roles.podman : Set variables part 1] ***************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:3 Saturday 24 August 2024 12:35:41 -0400 (0:00:00.424) 0:01:49.847 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set variables part 2] ***************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:7 Saturday 24 August 2024 12:35:41 -0400 (0:00:00.038) 0:01:49.885 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_rootless": false, "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:13 Saturday 24 August 2024 12:35:41 -0400 (0:00:00.044) 0:01:49.929 ******* included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Saturday 24 August 2024 12:35:41 -0400 (0:00:00.086) 0:01:50.016 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Saturday 24 August 2024 12:35:41 -0400 (0:00:00.049) 0:01:50.066 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Saturday 24 August 2024 12:35:41 -0400 (0:00:00.050) 0:01:50.117 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Stat XDG_RUNTIME_DIR] ***************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:18 Saturday 24 August 2024 12:35:41 -0400 (0:00:00.049) 0:01:50.167 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Manage each secret] ******************* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:34 Saturday 24 August 2024 12:35:41 -0400 (0:00:00.038) 0:01:50.206 ******* [WARNING]: Using a variable for a task's 'args' is unsafe in some situations (see https://docs.ansible.com/ansible/devel/reference_appendices/faq.html#argsplat- unsafe) changed: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": true } TASK [fedora.linux_system_roles.podman : Set variables part 1] ***************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:3 Saturday 24 August 2024 12:35:42 -0400 (0:00:00.414) 0:01:50.620 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set variables part 2] ***************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:7 Saturday 24 August 2024 12:35:42 -0400 (0:00:00.042) 0:01:50.662 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_rootless": false, "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:13 Saturday 24 August 2024 12:35:42 -0400 (0:00:00.055) 0:01:50.718 ******* included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Saturday 24 August 2024 12:35:42 -0400 (0:00:00.056) 0:01:50.774 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Saturday 24 August 2024 12:35:42 -0400 (0:00:00.031) 0:01:50.806 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Saturday 24 August 2024 12:35:42 -0400 (0:00:00.031) 0:01:50.837 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Stat XDG_RUNTIME_DIR] ***************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:18 Saturday 24 August 2024 12:35:42 -0400 (0:00:00.030) 0:01:50.868 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Manage each secret] ******************* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:34 Saturday 24 August 2024 12:35:42 -0400 (0:00:00.035) 0:01:50.904 ******* [WARNING]: Using a variable for a task's 'args' is unsafe in some situations (see https://docs.ansible.com/ansible/devel/reference_appendices/faq.html#argsplat- unsafe) changed: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": true } TASK [fedora.linux_system_roles.podman : Handle Kubernetes specifications] ***** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:153 Saturday 24 August 2024 12:35:42 -0400 (0:00:00.404) 0:01:51.308 ******* skipping: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Handle Quadlet specifications] ******** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:160 Saturday 24 August 2024 12:35:42 -0400 (0:00:00.030) 0:01:51.339 ******* included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for managed_node1 included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for managed_node1 included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for managed_node1 included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for managed_node1 included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for managed_node1 included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14 Saturday 24 August 2024 12:35:42 -0400 (0:00:00.136) 0:01:51.476 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_file_src": "quadlet-demo.kube", "__podman_quadlet_spec": {}, "__podman_quadlet_str": "[Install]\nWantedBy=default.target\n\n[Unit]\nRequires=quadlet-demo-mysql.service\nAfter=quadlet-demo-mysql.service\n\n[Kube]\n# Point to the yaml file in the same directory\nYaml=quadlet-demo.yml\n# Use the quadlet-demo network\nNetwork=quadlet-demo.network\n# Publish the envoy proxy data port\nPublishPort=8000:8080\n# Publish the envoy proxy admin port\nPublishPort=9000:9901\n# Use the envoy proxy config map in the same directory\nConfigMap=envoy-proxy-configmap.yml", "__podman_quadlet_template_src": "" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25 Saturday 24 August 2024 12:35:42 -0400 (0:00:00.045) 0:01:51.521 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_pull_image": true, "__podman_state": "absent", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] ***** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35 Saturday 24 August 2024 12:35:43 -0400 (0:00:00.099) 0:01:51.621 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48 Saturday 24 August 2024 12:35:43 -0400 (0:00:00.053) 0:01:51.674 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_name": "quadlet-demo", "__podman_quadlet_type": "kube", "__podman_rootless": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57 Saturday 24 August 2024 12:35:43 -0400 (0:00:00.074) 0:01:51.748 ******* included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Saturday 24 August 2024 12:35:43 -0400 (0:00:00.078) 0:01:51.826 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Saturday 24 August 2024 12:35:43 -0400 (0:00:00.048) 0:01:51.874 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Saturday 24 August 2024 12:35:43 -0400 (0:00:00.053) 0:01:51.928 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get group information] **************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:28 Saturday 24 August 2024 12:35:43 -0400 (0:00:00.048) 0:01:51.977 ******* ok: [managed_node1] => { "ansible_facts": { "getent_group": { "root": [ "x", "0", "" ] } }, "changed": false } TASK [fedora.linux_system_roles.podman : Set group name] *********************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:35 Saturday 24 August 2024 12:35:43 -0400 (0:00:00.389) 0:01:52.366 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_group_name": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:39 Saturday 24 August 2024 12:35:43 -0400 (0:00:00.066) 0:01:52.432 ******* ok: [managed_node1] => { "changed": false, "stat": { "atime": 1724516929.7168462, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "bb5b46ffbafcaa8c4021f3c8b3cb8594f48ef34b", "ctime": 1724516896.725572, "dev": 51713, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 6986657, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-sharedlib", "mode": "0755", "mtime": 1700557386.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 12640, "uid": 0, "version": "1255679238", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************ task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:50 Saturday 24 August 2024 12:35:44 -0400 (0:00:00.377) 0:01:52.810 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check group with getsubids] *********** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:55 Saturday 24 August 2024 12:35:44 -0400 (0:00:00.046) 0:01:52.857 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:60 Saturday 24 August 2024 12:35:44 -0400 (0:00:00.052) 0:01:52.910 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subuid file] ********************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:74 Saturday 24 August 2024 12:35:44 -0400 (0:00:00.043) 0:01:52.953 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subgid file] ********************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:79 Saturday 24 August 2024 12:35:44 -0400 (0:00:00.038) 0:01:52.992 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:84 Saturday 24 August 2024 12:35:44 -0400 (0:00:00.040) 0:01:53.032 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:94 Saturday 24 August 2024 12:35:44 -0400 (0:00:00.078) 0:01:53.111 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] ***** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:101 Saturday 24 August 2024 12:35:44 -0400 (0:00:00.032) 0:01:53.143 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62 Saturday 24 August 2024 12:35:44 -0400 (0:00:00.032) 0:01:53.176 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_activate_systemd_unit": true, "__podman_images_found": [], "__podman_kube_yamls_raw": [ "quadlet-demo.yml" ], "__podman_service_name": "quadlet-demo.service", "__podman_systemd_scope": "system", "__podman_user_home_dir": "/root", "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73 Saturday 24 August 2024 12:35:44 -0400 (0:00:00.055) 0:01:53.231 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_path": "/etc/containers/systemd" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get kube yaml contents] *************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77 Saturday 24 August 2024 12:35:44 -0400 (0:00:00.037) 0:01:53.268 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87 Saturday 24 August 2024 12:35:44 -0400 (0:00:00.050) 0:01:53.318 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_images": [], "__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo.kube", "__podman_volumes": [] }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 6] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:103 Saturday 24 August 2024 12:35:44 -0400 (0:00:00.088) 0:01:53.407 ******* ok: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Cleanup quadlets] ********************* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:110 Saturday 24 August 2024 12:35:44 -0400 (0:00:00.046) 0:01:53.453 ******* included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Stat XDG_RUNTIME_DIR] ***************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:4 Saturday 24 August 2024 12:35:44 -0400 (0:00:00.083) 0:01:53.537 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Stop and disable service] ************* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:12 Saturday 24 August 2024 12:35:44 -0400 (0:00:00.031) 0:01:53.568 ******* changed: [managed_node1] => { "changed": true, "enabled": false, "failed_when_result": false, "name": "quadlet-demo.service", "state": "stopped", "status": { "ActiveEnterTimestamp": "Sat 2024-08-24 12:35:23 EDT", "ActiveEnterTimestampMonotonic": "726499194", "ActiveExitTimestampMonotonic": "0", "ActiveState": "active", "After": "systemd-journald.socket basic.target -.mount quadlet-demo-network.service system.slice quadlet-demo-mysql.service sysinit.target", "AllowIsolate": "no", "AllowedCPUs": "", "AllowedMemoryNodes": "", "AmbientCapabilities": "", "AssertResult": "yes", "AssertTimestamp": "Sat 2024-08-24 12:35:22 EDT", "AssertTimestampMonotonic": "725568879", "Before": "shutdown.target", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "CPUAccounting": "no", "CPUAffinity": "", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "[not set]", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "no", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf", "CollectMode": "inactive", "ConditionResult": "yes", "ConditionTimestamp": "Sat 2024-08-24 12:35:22 EDT", "ConditionTimestampMonotonic": "725568878", "ConfigurationDirectoryMode": "0755", "Conflicts": "shutdown.target", "ControlGroup": "/system.slice/quadlet-demo.service", "ControlPID": "0", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "Delegate": "no", "Description": "quadlet-demo.service", "DevicePolicy": "auto", "DynamicUser": "no", "EffectiveCPUs": "", "EffectiveMemoryNodes": "", "Environment": "PODMAN_SYSTEMD_UNIT=quadlet-demo.service", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "87597", "ExecMainStartTimestamp": "Sat 2024-08-24 12:35:23 EDT", "ExecMainStartTimestampMonotonic": "726499171", "ExecMainStatus": "0", "ExecStart": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman kube play --replace --service-container=true --network=systemd-quadlet-demo --configmap /etc/containers/systemd/envoy-proxy-configmap.yml --publish 8000:8080 --publish 9000:9901 /etc/containers/systemd/quadlet-demo.yml ; ignore_errors=no ; start_time=[Sat 2024-08-24 12:35:22 EDT] ; stop_time=[n/a] ; pid=87537 ; code=(null) ; status=0/0 }", "ExecStopPost": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman kube down /etc/containers/systemd/quadlet-demo.yml ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FragmentPath": "/run/systemd/generator/quadlet-demo.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOSchedulingClass": "0", "IOSchedulingPriority": "0", "IOWeight": "[not set]", "IPAccounting": "no", "IPEgressBytes": "18446744073709551615", "IPEgressPackets": "18446744073709551615", "IPIngressBytes": "18446744073709551615", "IPIngressPackets": "18446744073709551615", "Id": "quadlet-demo.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestamp": "Sat 2024-08-24 12:35:22 EDT", "InactiveExitTimestampMonotonic": "725570197", "InvocationID": "67d7f03a2403424ab0f3e8a91570703c", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "mixed", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "0", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "65536", "LimitMEMLOCKSoft": "65536", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "262144", "LimitNOFILESoft": "1024", "LimitNPROC": "14003", "LimitNPROCSoft": "14003", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "14003", "LimitSIGPENDINGSoft": "14003", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "87597", "MemoryAccounting": "yes", "MemoryCurrent": "5255168", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemorySwapMax": "infinity", "MountAPIVFS": "no", "MountFlags": "", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAMask": "", "NUMAPolicy": "n/a", "Names": "quadlet-demo.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "all", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "PermissionsStartOnly": "no", "Perpetual": "no", "PrivateDevices": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "RemainAfterExit": "no", "RemoveIPC": "no", "Requires": "-.mount quadlet-demo-mysql.service sysinit.target quadlet-demo-network.service system.slice", "RequiresMountsFor": "/run/containers", "Restart": "no", "RestartUSec": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "Slice": "system.slice", "SourcePath": "/etc/containers/systemd/quadlet-demo.kube", "StandardError": "inherit", "StandardInput": "null", "StandardInputData": "", "StandardOutput": "journal", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StateChangeTimestamp": "Sat 2024-08-24 12:35:23 EDT", "StateChangeTimestampMonotonic": "726499194", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "running", "SuccessAction": "none", "SyslogFacility": "3", "SyslogIdentifier": "quadlet-demo", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "0", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "4", "TasksMax": "22405", "TimeoutStartUSec": "1min 30s", "TimeoutStopUSec": "1min 30s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "notify", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "disabled", "UnitFileState": "generated", "UtmpMode": "init", "WatchdogTimestamp": "Sat 2024-08-24 12:35:23 EDT", "WatchdogTimestampMonotonic": "726499191", "WatchdogUSec": "0" } } TASK [fedora.linux_system_roles.podman : See if quadlet file exists] *********** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:33 Saturday 24 August 2024 12:35:48 -0400 (0:00:03.207) 0:01:56.776 ******* ok: [managed_node1] => { "changed": false, "stat": { "atime": 1724517322.2370505, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 8, "charset": "us-ascii", "checksum": "7a5c73a5d935a42431c87bcdbeb8a04ed0909dc7", "ctime": 1724517321.5960455, "dev": 51713, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 497033403, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "text/plain", "mode": "0644", "mtime": 1724517321.3340435, "nlink": 1, "path": "/etc/containers/systemd/quadlet-demo.kube", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 456, "uid": 0, "version": "3427625058", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [fedora.linux_system_roles.podman : Parse quadlet file] ******************* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:38 Saturday 24 August 2024 12:35:48 -0400 (0:00:00.399) 0:01:57.175 ******* included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Slurp quadlet file] ******************* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml:6 Saturday 24 August 2024 12:35:48 -0400 (0:00:00.058) 0:01:57.234 ******* ok: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Parse quadlet file] ******************* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml:12 Saturday 24 August 2024 12:35:49 -0400 (0:00:00.388) 0:01:57.623 ******* ok: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Parse quadlet yaml file] ************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml:44 Saturday 24 August 2024 12:35:49 -0400 (0:00:00.075) 0:01:57.698 ******* skipping: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Reset raw variable] ******************* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml:52 Saturday 24 August 2024 12:35:49 -0400 (0:00:00.051) 0:01:57.750 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_raw": null }, "changed": false } TASK [fedora.linux_system_roles.podman : Remove quadlet file] ****************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:42 Saturday 24 August 2024 12:35:49 -0400 (0:00:00.051) 0:01:57.801 ******* changed: [managed_node1] => { "changed": true, "path": "/etc/containers/systemd/quadlet-demo.kube", "state": "absent" } TASK [fedora.linux_system_roles.podman : Refresh systemd] ********************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:48 Saturday 24 August 2024 12:35:49 -0400 (0:00:00.451) 0:01:58.252 ******* ok: [managed_node1] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.podman : Remove managed resource] ************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:58 Saturday 24 August 2024 12:35:50 -0400 (0:00:00.606) 0:01:58.859 ******* skipping: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Remove volumes] *********************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:95 Saturday 24 August 2024 12:35:50 -0400 (0:00:00.041) 0:01:58.900 ******* skipping: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Clear parsed podman variable] ********* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:112 Saturday 24 August 2024 12:35:50 -0400 (0:00:00.054) 0:01:58.955 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_parsed": null }, "changed": false } TASK [fedora.linux_system_roles.podman : Prune images no longer in use] ******** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:116 Saturday 24 August 2024 12:35:50 -0400 (0:00:00.037) 0:01:58.992 ******* changed: [managed_node1] => { "changed": true, "cmd": [ "podman", "image", "prune", "--all", "-f" ], "delta": "0:00:00.721014", "end": "2024-08-24 12:35:51.408908", "rc": 0, "start": "2024-08-24 12:35:50.687894" } STDOUT: 0030ba3d620c647159c935ee778991c68ef3e51a274703753b0bc530104ef5e5 9f9ec7f2fdef9168f74e9d057f307955db14d782cff22ded51d277d74798cb2f fc3855720f39e33996597ca2164a27fa28c6b03eb8774fdb5249a1c2244ef9fd fcf3e41b8864a14d75a6d0627d3d02154e28a153aa57e8baa392cd744ffa0d0b 5af2585e22ed1562885d9407efab74010090427be79048c2cd6a226517cc1e1d TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:127 Saturday 24 August 2024 12:35:51 -0400 (0:00:01.110) 0:02:00.103 ******* included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Saturday 24 August 2024 12:35:51 -0400 (0:00:00.083) 0:02:00.186 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Saturday 24 August 2024 12:35:51 -0400 (0:00:00.039) 0:02:00.225 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Saturday 24 August 2024 12:35:51 -0400 (0:00:00.041) 0:02:00.267 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : For testing and debugging - images] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:137 Saturday 24 August 2024 12:35:51 -0400 (0:00:00.035) 0:02:00.303 ******* ok: [managed_node1] => { "changed": false, "cmd": [ "podman", "images", "-n" ], "delta": "0:00:00.033505", "end": "2024-08-24 12:35:52.020143", "rc": 0, "start": "2024-08-24 12:35:51.986638" } STDOUT: quay.io/linux-system-roles/mysql 5.6 dd3b2a5dcb48 2 years ago 308 MB TASK [fedora.linux_system_roles.podman : For testing and debugging - volumes] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:146 Saturday 24 August 2024 12:35:52 -0400 (0:00:00.403) 0:02:00.707 ******* ok: [managed_node1] => { "changed": false, "cmd": [ "podman", "volume", "ls", "-n" ], "delta": "0:00:00.030150", "end": "2024-08-24 12:35:52.451392", "rc": 0, "start": "2024-08-24 12:35:52.421242" } STDOUT: local systemd-quadlet-demo-mysql local wp-pv-claim local envoy-proxy-config local envoy-certificates TASK [fedora.linux_system_roles.podman : For testing and debugging - containers] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:155 Saturday 24 August 2024 12:35:52 -0400 (0:00:00.425) 0:02:01.132 ******* ok: [managed_node1] => { "changed": false, "cmd": [ "podman", "ps", "--noheading" ], "delta": "0:00:00.035408", "end": "2024-08-24 12:35:52.888353", "rc": 0, "start": "2024-08-24 12:35:52.852945" } STDOUT: 68250d5fc021 quay.io/linux-system-roles/mysql:5.6 mysqld About a minute ago Up About a minute (healthy) quadlet-demo-mysql TASK [fedora.linux_system_roles.podman : For testing and debugging - networks] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:164 Saturday 24 August 2024 12:35:52 -0400 (0:00:00.436) 0:02:01.569 ******* ok: [managed_node1] => { "changed": false, "cmd": [ "podman", "network", "ls", "-n", "-q" ], "delta": "0:00:00.055656", "end": "2024-08-24 12:35:53.333009", "rc": 0, "start": "2024-08-24 12:35:53.277353" } STDOUT: podman podman-default-kube-network systemd-quadlet-demo TASK [fedora.linux_system_roles.podman : For testing and debugging - secrets] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:173 Saturday 24 August 2024 12:35:53 -0400 (0:00:00.454) 0:02:02.024 ******* ok: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : For testing and debugging - services] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:183 Saturday 24 August 2024 12:35:53 -0400 (0:00:00.421) 0:02:02.445 ******* ok: [managed_node1] => { "ansible_facts": { "services": { "68250d5fc0218bc6db3ab2...5ebe15c37a833d3345ad2e2df2c1916dd6cc3.service": { "name": "68250d5fc0218bc6db3ab2...5ebe15c37a833d3345ad2e2df2c1916dd6cc3.service", "source": "systemd", "state": "inactive", "status": "transient" }, "68250d5fc0218bc6db3ab2ece605ebe15c37a833d3345ad2e2df2c1916dd6cc3.service": { "name": "68250d5fc0218bc6db3ab2ece605ebe15c37a833d3345ad2e2df2c1916dd6cc3.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "NetworkManager-dispatcher.service": { "name": "NetworkManager-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "NetworkManager-wait-online.service": { "name": "NetworkManager-wait-online.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "NetworkManager.service": { "name": "NetworkManager.service", "source": "systemd", "state": "running", "status": "enabled" }, "auditd.service": { "name": "auditd.service", "source": "systemd", "state": "running", "status": "enabled" }, "auth-rpcgss-module.service": { "name": "auth-rpcgss-module.service", "source": "systemd", "state": "stopped", "status": "static" }, "autovt@.service": { "name": "autovt@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "certmonger.service": { "name": "certmonger.service", "source": "systemd", "state": "running", "status": "enabled" }, "chrony-dnssrv@.service": { "name": "chrony-dnssrv@.service", "source": "systemd", "state": "unknown", "status": "static" }, "chrony-wait.service": { "name": "chrony-wait.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd.service": { "name": "chronyd.service", "source": "systemd", "state": "running", "status": "enabled" }, "cloud-config.service": { "name": "cloud-config.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-final.service": { "name": "cloud-final.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init-hotplugd.service": { "name": "cloud-init-hotplugd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cloud-init-local.service": { "name": "cloud-init-local.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init.service": { "name": "cloud-init.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cni-dhcp.service": { "name": "cni-dhcp.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "console-getty.service": { "name": "console-getty.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "container-getty@.service": { "name": "container-getty@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cpupower.service": { "name": "cpupower.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "crond.service": { "name": "crond.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-org.fedoraproject.FirewallD1.service": { "name": "dbus-org.fedoraproject.FirewallD1.service", "source": "systemd", "state": "active", "status": "enabled" }, "dbus-org.freedesktop.hostname1.service": { "name": "dbus-org.freedesktop.hostname1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.locale1.service": { "name": "dbus-org.freedesktop.locale1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.login1.service": { "name": "dbus-org.freedesktop.login1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.nm-dispatcher.service": { "name": "dbus-org.freedesktop.nm-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "dbus-org.freedesktop.portable1.service": { "name": "dbus-org.freedesktop.portable1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.timedate1.service": { "name": "dbus-org.freedesktop.timedate1.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "dbus.service": { "name": "dbus.service", "source": "systemd", "state": "running", "status": "static" }, "debug-shell.service": { "name": "debug-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dnf-makecache.service": { "name": "dnf-makecache.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-system-upgrade-cleanup.service": { "name": "dnf-system-upgrade-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "dnf-system-upgrade.service": { "name": "dnf-system-upgrade.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dnsmasq.service": { "name": "dnsmasq.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dracut-cmdline.service": { "name": "dracut-cmdline.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-initqueue.service": { "name": "dracut-initqueue.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-mount.service": { "name": "dracut-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-mount.service": { "name": "dracut-pre-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-pivot.service": { "name": "dracut-pre-pivot.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-trigger.service": { "name": "dracut-pre-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-udev.service": { "name": "dracut-pre-udev.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown-onfailure.service": { "name": "dracut-shutdown-onfailure.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown.service": { "name": "dracut-shutdown.service", "source": "systemd", "state": "stopped", "status": "static" }, "ebtables.service": { "name": "ebtables.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "emergency.service": { "name": "emergency.service", "source": "systemd", "state": "stopped", "status": "static" }, "firewalld.service": { "name": "firewalld.service", "source": "systemd", "state": "running", "status": "enabled" }, "fstrim.service": { "name": "fstrim.service", "source": "systemd", "state": "inactive", "status": "static" }, "getty@.service": { "name": "getty@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "getty@tty1.service": { "name": "getty@tty1.service", "source": "systemd", "state": "running", "status": "unknown" }, "grub-boot-indeterminate.service": { "name": "grub-boot-indeterminate.service", "source": "systemd", "state": "inactive", "status": "static" }, "gssproxy.service": { "name": "gssproxy.service", "source": "systemd", "state": "running", "status": "disabled" }, "halt-local.service": { "name": "halt-local.service", "source": "systemd", "state": "inactive", "status": "static" }, "import-state.service": { "name": "import-state.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "initrd-cleanup.service": { "name": "initrd-cleanup.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-parse-etc.service": { "name": "initrd-parse-etc.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-switch-root.service": { "name": "initrd-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-udevadm-cleanup-db.service": { "name": "initrd-udevadm-cleanup-db.service", "source": "systemd", "state": "stopped", "status": "static" }, "iprdump.service": { "name": "iprdump.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "iprinit.service": { "name": "iprinit.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "iprupdate.service": { "name": "iprupdate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "irqbalance.service": { "name": "irqbalance.service", "source": "systemd", "state": "running", "status": "enabled" }, "kdump.service": { "name": "kdump.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "kmod-static-nodes.service": { "name": "kmod-static-nodes.service", "source": "systemd", "state": "stopped", "status": "static" }, "kvm_stat.service": { "name": "kvm_stat.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "ldconfig.service": { "name": "ldconfig.service", "source": "systemd", "state": "stopped", "status": "static" }, "loadmodules.service": { "name": "loadmodules.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "man-db-cache-update.service": { "name": "man-db-cache-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "man-db-restart-cache-update.service": { "name": "man-db-restart-cache-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "messagebus.service": { "name": "messagebus.service", "source": "systemd", "state": "active", "status": "static" }, "microcode.service": { "name": "microcode.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "nfs-blkmap.service": { "name": "nfs-blkmap.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-convert.service": { "name": "nfs-convert.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-idmapd.service": { "name": "nfs-idmapd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-mountd.service": { "name": "nfs-mountd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-server.service": { "name": "nfs-server.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-utils.service": { "name": "nfs-utils.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfsdcld.service": { "name": "nfsdcld.service", "source": "systemd", "state": "stopped", "status": "static" }, "nftables.service": { "name": "nftables.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nis-domainname.service": { "name": "nis-domainname.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "oddjobd.service": { "name": "oddjobd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "plymouth-halt.service": { "name": "plymouth-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-kexec.service": { "name": "plymouth-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-poweroff.service": { "name": "plymouth-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-quit-wait.service": { "name": "plymouth-quit-wait.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-quit.service": { "name": "plymouth-quit.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-read-write.service": { "name": "plymouth-read-write.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-reboot.service": { "name": "plymouth-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-start.service": { "name": "plymouth-start.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-switch-root-initramfs.service": { "name": "plymouth-switch-root-initramfs.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-switch-root.service": { "name": "plymouth-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "podman-auto-update.service": { "name": "podman-auto-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "podman-clean-transient.service": { "name": "podman-clean-transient.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "podman-kube@.service": { "name": "podman-kube@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "podman-restart.service": { "name": "podman-restart.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "podman.service": { "name": "podman.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "polkit.service": { "name": "polkit.service", "source": "systemd", "state": "running", "status": "static" }, "qemu-guest-agent.service": { "name": "qemu-guest-agent.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "quadlet-demo-mysql-volume.service": { "name": "quadlet-demo-mysql-volume.service", "source": "systemd", "state": "stopped", "status": "generated" }, "quadlet-demo-mysql.service": { "name": "quadlet-demo-mysql.service", "source": "systemd", "state": "running", "status": "generated" }, "quadlet-demo-network.service": { "name": "quadlet-demo-network.service", "source": "systemd", "state": "stopped", "status": "generated" }, "quotaon.service": { "name": "quotaon.service", "source": "systemd", "state": "inactive", "status": "static" }, "rc-local.service": { "name": "rc-local.service", "source": "systemd", "state": "stopped", "status": "static" }, "rdisc.service": { "name": "rdisc.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rescue.service": { "name": "rescue.service", "source": "systemd", "state": "stopped", "status": "static" }, "restraintd.service": { "name": "restraintd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rngd.service": { "name": "rngd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpc-gssd.service": { "name": "rpc-gssd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd-notify.service": { "name": "rpc-statd-notify.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd.service": { "name": "rpc-statd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpcbind.service": { "name": "rpcbind.service", "source": "systemd", "state": "running", "status": "enabled" }, "rsyslog.service": { "name": "rsyslog.service", "source": "systemd", "state": "running", "status": "enabled" }, "selinux-autorelabel-mark.service": { "name": "selinux-autorelabel-mark.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "selinux-autorelabel.service": { "name": "selinux-autorelabel.service", "source": "systemd", "state": "inactive", "status": "static" }, "serial-getty@.service": { "name": "serial-getty@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@.service": { "name": "sshd-keygen@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@ecdsa.service": { "name": "sshd-keygen@ecdsa.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd-keygen@ed25519.service": { "name": "sshd-keygen@ed25519.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd-keygen@rsa.service": { "name": "sshd-keygen@rsa.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd.service": { "name": "sshd.service", "source": "systemd", "state": "running", "status": "enabled" }, "sshd@.service": { "name": "sshd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "sssd-autofs.service": { "name": "sssd-autofs.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-kcm.service": { "name": "sssd-kcm.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "sssd-nss.service": { "name": "sssd-nss.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pac.service": { "name": "sssd-pac.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pam.service": { "name": "sssd-pam.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-ssh.service": { "name": "sssd-ssh.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-sudo.service": { "name": "sssd-sudo.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd.service": { "name": "sssd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "syslog.service": { "name": "syslog.service", "source": "systemd", "state": "active", "status": "enabled" }, "system-update-cleanup.service": { "name": "system-update-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-ask-password-console.service": { "name": "systemd-ask-password-console.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-plymouth.service": { "name": "systemd-ask-password-plymouth.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-wall.service": { "name": "systemd-ask-password-wall.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-backlight@.service": { "name": "systemd-backlight@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-binfmt.service": { "name": "systemd-binfmt.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-coredump@.service": { "name": "systemd-coredump@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-exit.service": { "name": "systemd-exit.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-firstboot.service": { "name": "systemd-firstboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck-root.service": { "name": "systemd-fsck-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck@.service": { "name": "systemd-fsck@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-halt.service": { "name": "systemd-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hibernate-resume@.service": { "name": "systemd-hibernate-resume@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-hibernate.service": { "name": "systemd-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hostnamed.service": { "name": "systemd-hostnamed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hwdb-update.service": { "name": "systemd-hwdb-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-hybrid-sleep.service": { "name": "systemd-hybrid-sleep.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-initctl.service": { "name": "systemd-initctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-catalog-update.service": { "name": "systemd-journal-catalog-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-flush.service": { "name": "systemd-journal-flush.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journald.service": { "name": "systemd-journald.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-kexec.service": { "name": "systemd-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-localed.service": { "name": "systemd-localed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-logind.service": { "name": "systemd-logind.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-machine-id-commit.service": { "name": "systemd-machine-id-commit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-modules-load.service": { "name": "systemd-modules-load.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-portabled.service": { "name": "systemd-portabled.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-poweroff.service": { "name": "systemd-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pstore.service": { "name": "systemd-pstore.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-quotacheck.service": { "name": "systemd-quotacheck.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-random-seed.service": { "name": "systemd-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-reboot.service": { "name": "systemd-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-remount-fs.service": { "name": "systemd-remount-fs.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-resolved.service": { "name": "systemd-resolved.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-rfkill.service": { "name": "systemd-rfkill.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend-then-hibernate.service": { "name": "systemd-suspend-then-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend.service": { "name": "systemd-suspend.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-sysctl.service": { "name": "systemd-sysctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-sysusers.service": { "name": "systemd-sysusers.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-timedated.service": { "name": "systemd-timedated.service", "source": "systemd", "state": "inactive", "status": "masked" }, "systemd-tmpfiles-clean.service": { "name": "systemd-tmpfiles-clean.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev.service": { "name": "systemd-tmpfiles-setup-dev.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup.service": { "name": "systemd-tmpfiles-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-settle.service": { "name": "systemd-udev-settle.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-udev-trigger.service": { "name": "systemd-udev-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udevd.service": { "name": "systemd-udevd.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-update-done.service": { "name": "systemd-update-done.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp-runlevel.service": { "name": "systemd-update-utmp-runlevel.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp.service": { "name": "systemd-update-utmp.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-user-sessions.service": { "name": "systemd-user-sessions.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-vconsole-setup.service": { "name": "systemd-vconsole-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-volatile-root.service": { "name": "systemd-volatile-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "tcsd.service": { "name": "tcsd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "teamd@.service": { "name": "teamd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "timedatex.service": { "name": "timedatex.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "tuned.service": { "name": "tuned.service", "source": "systemd", "state": "running", "status": "enabled" }, "unbound-anchor.service": { "name": "unbound-anchor.service", "source": "systemd", "state": "stopped", "status": "static" }, "user-runtime-dir@.service": { "name": "user-runtime-dir@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user-runtime-dir@0.service": { "name": "user-runtime-dir@0.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "user@.service": { "name": "user@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user@0.service": { "name": "user@0.service", "source": "systemd", "state": "running", "status": "unknown" } } }, "changed": false } TASK [fedora.linux_system_roles.podman : Create and update quadlets] *********** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:114 Saturday 24 August 2024 12:35:55 -0400 (0:00:01.994) 0:02:04.439 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14 Saturday 24 August 2024 12:35:55 -0400 (0:00:00.033) 0:02:04.473 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_file_src": "", "__podman_quadlet_spec": {}, "__podman_quadlet_str": "---\napiVersion: v1\nkind: PersistentVolumeClaim\nmetadata:\n name: wp-pv-claim\n labels:\n app: wordpress\nspec:\n accessModes:\n - ReadWriteOnce\n resources:\n requests:\n storage: 20Gi\n---\napiVersion: v1\nkind: Pod\nmetadata:\n name: quadlet-demo\nspec:\n containers:\n - name: wordpress\n image: quay.io/linux-system-roles/wordpress:4.8-apache\n env:\n - name: WORDPRESS_DB_HOST\n value: quadlet-demo-mysql\n - name: WORDPRESS_DB_PASSWORD\n valueFrom:\n secretKeyRef:\n name: mysql-root-password-kube\n key: password\n volumeMounts:\n - name: wordpress-persistent-storage\n mountPath: /var/www/html\n resources:\n requests:\n memory: \"64Mi\"\n cpu: \"250m\"\n limits:\n memory: \"128Mi\"\n cpu: \"500m\"\n - name: envoy\n image: quay.io/linux-system-roles/envoyproxy:v1.25.0\n volumeMounts:\n - name: config-volume\n mountPath: /etc/envoy\n - name: certificates\n mountPath: /etc/envoy-certificates\n env:\n - name: ENVOY_UID\n value: \"0\"\n resources:\n requests:\n memory: \"64Mi\"\n cpu: \"250m\"\n limits:\n memory: \"128Mi\"\n cpu: \"500m\"\n volumes:\n - name: config-volume\n configMap:\n name: envoy-proxy-config\n - name: certificates\n secret:\n secretName: envoy-certificates\n - name: wordpress-persistent-storage\n persistentVolumeClaim:\n claimName: wp-pv-claim\n - name: www # not used - for testing hostpath\n hostPath:\n path: /tmp/httpd3\n - name: create # not used - for testing hostpath\n hostPath:\n path: /tmp/httpd3-create\n", "__podman_quadlet_template_src": "quadlet-demo.yml.j2" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25 Saturday 24 August 2024 12:35:55 -0400 (0:00:00.111) 0:02:04.584 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_pull_image": true, "__podman_state": "absent", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] ***** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35 Saturday 24 August 2024 12:35:56 -0400 (0:00:00.042) 0:02:04.626 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48 Saturday 24 August 2024 12:35:56 -0400 (0:00:00.036) 0:02:04.663 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_name": "quadlet-demo", "__podman_quadlet_type": "yml", "__podman_rootless": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57 Saturday 24 August 2024 12:35:56 -0400 (0:00:00.046) 0:02:04.709 ******* included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Saturday 24 August 2024 12:35:56 -0400 (0:00:00.064) 0:02:04.774 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Saturday 24 August 2024 12:35:56 -0400 (0:00:00.041) 0:02:04.816 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Saturday 24 August 2024 12:35:56 -0400 (0:00:00.044) 0:02:04.860 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get group information] **************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:28 Saturday 24 August 2024 12:35:56 -0400 (0:00:00.046) 0:02:04.907 ******* ok: [managed_node1] => { "ansible_facts": { "getent_group": { "root": [ "x", "0", "" ] } }, "changed": false } TASK [fedora.linux_system_roles.podman : Set group name] *********************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:35 Saturday 24 August 2024 12:35:56 -0400 (0:00:00.369) 0:02:05.276 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_group_name": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:39 Saturday 24 August 2024 12:35:56 -0400 (0:00:00.043) 0:02:05.320 ******* ok: [managed_node1] => { "changed": false, "stat": { "atime": 1724516929.7168462, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "bb5b46ffbafcaa8c4021f3c8b3cb8594f48ef34b", "ctime": 1724516896.725572, "dev": 51713, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 6986657, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-sharedlib", "mode": "0755", "mtime": 1700557386.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 12640, "uid": 0, "version": "1255679238", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************ task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:50 Saturday 24 August 2024 12:35:57 -0400 (0:00:00.460) 0:02:05.781 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check group with getsubids] *********** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:55 Saturday 24 August 2024 12:35:57 -0400 (0:00:00.055) 0:02:05.837 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:60 Saturday 24 August 2024 12:35:57 -0400 (0:00:00.039) 0:02:05.877 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subuid file] ********************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:74 Saturday 24 August 2024 12:35:57 -0400 (0:00:00.033) 0:02:05.910 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subgid file] ********************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:79 Saturday 24 August 2024 12:35:57 -0400 (0:00:00.034) 0:02:05.945 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:84 Saturday 24 August 2024 12:35:57 -0400 (0:00:00.033) 0:02:05.978 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:94 Saturday 24 August 2024 12:35:57 -0400 (0:00:00.032) 0:02:06.011 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] ***** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:101 Saturday 24 August 2024 12:35:57 -0400 (0:00:00.036) 0:02:06.047 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62 Saturday 24 August 2024 12:35:57 -0400 (0:00:00.033) 0:02:06.080 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_activate_systemd_unit": true, "__podman_images_found": [], "__podman_kube_yamls_raw": "", "__podman_service_name": "", "__podman_systemd_scope": "system", "__podman_user_home_dir": "/root", "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73 Saturday 24 August 2024 12:35:57 -0400 (0:00:00.057) 0:02:06.138 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_path": "/etc/containers/systemd" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get kube yaml contents] *************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77 Saturday 24 August 2024 12:35:57 -0400 (0:00:00.033) 0:02:06.171 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87 Saturday 24 August 2024 12:35:57 -0400 (0:00:00.032) 0:02:06.204 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_images": [], "__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo.yml", "__podman_volumes": [] }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 6] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:103 Saturday 24 August 2024 12:35:57 -0400 (0:00:00.071) 0:02:06.275 ******* ok: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Cleanup quadlets] ********************* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:110 Saturday 24 August 2024 12:35:57 -0400 (0:00:00.037) 0:02:06.313 ******* included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Stat XDG_RUNTIME_DIR] ***************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:4 Saturday 24 August 2024 12:35:57 -0400 (0:00:00.074) 0:02:06.388 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Stop and disable service] ************* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:12 Saturday 24 August 2024 12:35:57 -0400 (0:00:00.072) 0:02:06.460 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : See if quadlet file exists] *********** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:33 Saturday 24 August 2024 12:35:57 -0400 (0:00:00.035) 0:02:06.495 ******* ok: [managed_node1] => { "changed": false, "stat": { "atime": 1724517300.8158789, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 8, "charset": "us-ascii", "checksum": "998dccde0483b1654327a46ddd89cbaa47650370", "ctime": 1724517297.6538534, "dev": 51713, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 473956741, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "text/plain", "mode": "0644", "mtime": 1724517297.3598511, "nlink": 1, "path": "/etc/containers/systemd/quadlet-demo.yml", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 1605, "uid": 0, "version": "336872421", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [fedora.linux_system_roles.podman : Parse quadlet file] ******************* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:38 Saturday 24 August 2024 12:35:58 -0400 (0:00:00.365) 0:02:06.861 ******* included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Slurp quadlet file] ******************* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml:6 Saturday 24 August 2024 12:35:58 -0400 (0:00:00.062) 0:02:06.923 ******* ok: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Parse quadlet file] ******************* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml:12 Saturday 24 August 2024 12:35:58 -0400 (0:00:00.357) 0:02:07.281 ******* skipping: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Parse quadlet yaml file] ************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml:44 Saturday 24 August 2024 12:35:58 -0400 (0:00:00.033) 0:02:07.314 ******* ok: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Reset raw variable] ******************* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml:52 Saturday 24 August 2024 12:35:58 -0400 (0:00:00.039) 0:02:07.354 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_raw": null }, "changed": false } TASK [fedora.linux_system_roles.podman : Remove quadlet file] ****************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:42 Saturday 24 August 2024 12:35:58 -0400 (0:00:00.032) 0:02:07.387 ******* changed: [managed_node1] => { "changed": true, "path": "/etc/containers/systemd/quadlet-demo.yml", "state": "absent" } TASK [fedora.linux_system_roles.podman : Refresh systemd] ********************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:48 Saturday 24 August 2024 12:35:59 -0400 (0:00:00.358) 0:02:07.745 ******* ok: [managed_node1] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.podman : Remove managed resource] ************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:58 Saturday 24 August 2024 12:35:59 -0400 (0:00:00.597) 0:02:08.342 ******* skipping: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Remove volumes] *********************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:95 Saturday 24 August 2024 12:35:59 -0400 (0:00:00.071) 0:02:08.413 ******* skipping: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Clear parsed podman variable] ********* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:112 Saturday 24 August 2024 12:35:59 -0400 (0:00:00.047) 0:02:08.461 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_parsed": null }, "changed": false } TASK [fedora.linux_system_roles.podman : Prune images no longer in use] ******** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:116 Saturday 24 August 2024 12:35:59 -0400 (0:00:00.031) 0:02:08.493 ******* changed: [managed_node1] => { "changed": true, "cmd": [ "podman", "image", "prune", "--all", "-f" ], "delta": "0:00:00.030382", "end": "2024-08-24 12:36:00.203613", "rc": 0, "start": "2024-08-24 12:36:00.173231" } TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:127 Saturday 24 August 2024 12:36:00 -0400 (0:00:00.393) 0:02:08.886 ******* included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Saturday 24 August 2024 12:36:00 -0400 (0:00:00.060) 0:02:08.947 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Saturday 24 August 2024 12:36:00 -0400 (0:00:00.033) 0:02:08.980 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Saturday 24 August 2024 12:36:00 -0400 (0:00:00.034) 0:02:09.015 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : For testing and debugging - images] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:137 Saturday 24 August 2024 12:36:00 -0400 (0:00:00.033) 0:02:09.049 ******* ok: [managed_node1] => { "changed": false, "cmd": [ "podman", "images", "-n" ], "delta": "0:00:00.033089", "end": "2024-08-24 12:36:00.767456", "rc": 0, "start": "2024-08-24 12:36:00.734367" } STDOUT: quay.io/linux-system-roles/mysql 5.6 dd3b2a5dcb48 2 years ago 308 MB TASK [fedora.linux_system_roles.podman : For testing and debugging - volumes] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:146 Saturday 24 August 2024 12:36:00 -0400 (0:00:00.401) 0:02:09.450 ******* ok: [managed_node1] => { "changed": false, "cmd": [ "podman", "volume", "ls", "-n" ], "delta": "0:00:00.030605", "end": "2024-08-24 12:36:01.162074", "rc": 0, "start": "2024-08-24 12:36:01.131469" } STDOUT: local systemd-quadlet-demo-mysql local wp-pv-claim local envoy-proxy-config local envoy-certificates TASK [fedora.linux_system_roles.podman : For testing and debugging - containers] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:155 Saturday 24 August 2024 12:36:01 -0400 (0:00:00.394) 0:02:09.845 ******* ok: [managed_node1] => { "changed": false, "cmd": [ "podman", "ps", "--noheading" ], "delta": "0:00:00.035869", "end": "2024-08-24 12:36:01.565909", "rc": 0, "start": "2024-08-24 12:36:01.530040" } STDOUT: 68250d5fc021 quay.io/linux-system-roles/mysql:5.6 mysqld About a minute ago Up About a minute (healthy) quadlet-demo-mysql TASK [fedora.linux_system_roles.podman : For testing and debugging - networks] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:164 Saturday 24 August 2024 12:36:01 -0400 (0:00:00.403) 0:02:10.248 ******* ok: [managed_node1] => { "changed": false, "cmd": [ "podman", "network", "ls", "-n", "-q" ], "delta": "0:00:00.056603", "end": "2024-08-24 12:36:01.990366", "rc": 0, "start": "2024-08-24 12:36:01.933763" } STDOUT: podman podman-default-kube-network systemd-quadlet-demo TASK [fedora.linux_system_roles.podman : For testing and debugging - secrets] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:173 Saturday 24 August 2024 12:36:02 -0400 (0:00:00.424) 0:02:10.673 ******* ok: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : For testing and debugging - services] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:183 Saturday 24 August 2024 12:36:02 -0400 (0:00:00.393) 0:02:11.066 ******* ok: [managed_node1] => { "ansible_facts": { "services": { "68250d5fc0218bc6db3ab2...5ebe15c37a833d3345ad2e2df2c1916dd6cc3.service": { "name": "68250d5fc0218bc6db3ab2...5ebe15c37a833d3345ad2e2df2c1916dd6cc3.service", "source": "systemd", "state": "inactive", "status": "transient" }, "68250d5fc0218bc6db3ab2ece605ebe15c37a833d3345ad2e2df2c1916dd6cc3.service": { "name": "68250d5fc0218bc6db3ab2ece605ebe15c37a833d3345ad2e2df2c1916dd6cc3.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "NetworkManager-dispatcher.service": { "name": "NetworkManager-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "NetworkManager-wait-online.service": { "name": "NetworkManager-wait-online.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "NetworkManager.service": { "name": "NetworkManager.service", "source": "systemd", "state": "running", "status": "enabled" }, "auditd.service": { "name": "auditd.service", "source": "systemd", "state": "running", "status": "enabled" }, "auth-rpcgss-module.service": { "name": "auth-rpcgss-module.service", "source": "systemd", "state": "stopped", "status": "static" }, "autovt@.service": { "name": "autovt@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "certmonger.service": { "name": "certmonger.service", "source": "systemd", "state": "running", "status": "enabled" }, "chrony-dnssrv@.service": { "name": "chrony-dnssrv@.service", "source": "systemd", "state": "unknown", "status": "static" }, "chrony-wait.service": { "name": "chrony-wait.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd.service": { "name": "chronyd.service", "source": "systemd", "state": "running", "status": "enabled" }, "cloud-config.service": { "name": "cloud-config.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-final.service": { "name": "cloud-final.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init-hotplugd.service": { "name": "cloud-init-hotplugd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cloud-init-local.service": { "name": "cloud-init-local.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init.service": { "name": "cloud-init.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cni-dhcp.service": { "name": "cni-dhcp.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "console-getty.service": { "name": "console-getty.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "container-getty@.service": { "name": "container-getty@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cpupower.service": { "name": "cpupower.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "crond.service": { "name": "crond.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-org.fedoraproject.FirewallD1.service": { "name": "dbus-org.fedoraproject.FirewallD1.service", "source": "systemd", "state": "active", "status": "enabled" }, "dbus-org.freedesktop.hostname1.service": { "name": "dbus-org.freedesktop.hostname1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.locale1.service": { "name": "dbus-org.freedesktop.locale1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.login1.service": { "name": "dbus-org.freedesktop.login1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.nm-dispatcher.service": { "name": "dbus-org.freedesktop.nm-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "dbus-org.freedesktop.portable1.service": { "name": "dbus-org.freedesktop.portable1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.timedate1.service": { "name": "dbus-org.freedesktop.timedate1.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "dbus.service": { "name": "dbus.service", "source": "systemd", "state": "running", "status": "static" }, "debug-shell.service": { "name": "debug-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dnf-makecache.service": { "name": "dnf-makecache.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-system-upgrade-cleanup.service": { "name": "dnf-system-upgrade-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "dnf-system-upgrade.service": { "name": "dnf-system-upgrade.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dnsmasq.service": { "name": "dnsmasq.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dracut-cmdline.service": { "name": "dracut-cmdline.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-initqueue.service": { "name": "dracut-initqueue.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-mount.service": { "name": "dracut-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-mount.service": { "name": "dracut-pre-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-pivot.service": { "name": "dracut-pre-pivot.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-trigger.service": { "name": "dracut-pre-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-udev.service": { "name": "dracut-pre-udev.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown-onfailure.service": { "name": "dracut-shutdown-onfailure.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown.service": { "name": "dracut-shutdown.service", "source": "systemd", "state": "stopped", "status": "static" }, "ebtables.service": { "name": "ebtables.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "emergency.service": { "name": "emergency.service", "source": "systemd", "state": "stopped", "status": "static" }, "firewalld.service": { "name": "firewalld.service", "source": "systemd", "state": "running", "status": "enabled" }, "fstrim.service": { "name": "fstrim.service", "source": "systemd", "state": "inactive", "status": "static" }, "getty@.service": { "name": "getty@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "getty@tty1.service": { "name": "getty@tty1.service", "source": "systemd", "state": "running", "status": "unknown" }, "grub-boot-indeterminate.service": { "name": "grub-boot-indeterminate.service", "source": "systemd", "state": "inactive", "status": "static" }, "gssproxy.service": { "name": "gssproxy.service", "source": "systemd", "state": "running", "status": "disabled" }, "halt-local.service": { "name": "halt-local.service", "source": "systemd", "state": "inactive", "status": "static" }, "import-state.service": { "name": "import-state.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "initrd-cleanup.service": { "name": "initrd-cleanup.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-parse-etc.service": { "name": "initrd-parse-etc.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-switch-root.service": { "name": "initrd-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-udevadm-cleanup-db.service": { "name": "initrd-udevadm-cleanup-db.service", "source": "systemd", "state": "stopped", "status": "static" }, "iprdump.service": { "name": "iprdump.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "iprinit.service": { "name": "iprinit.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "iprupdate.service": { "name": "iprupdate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "irqbalance.service": { "name": "irqbalance.service", "source": "systemd", "state": "running", "status": "enabled" }, "kdump.service": { "name": "kdump.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "kmod-static-nodes.service": { "name": "kmod-static-nodes.service", "source": "systemd", "state": "stopped", "status": "static" }, "kvm_stat.service": { "name": "kvm_stat.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "ldconfig.service": { "name": "ldconfig.service", "source": "systemd", "state": "stopped", "status": "static" }, "loadmodules.service": { "name": "loadmodules.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "man-db-cache-update.service": { "name": "man-db-cache-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "man-db-restart-cache-update.service": { "name": "man-db-restart-cache-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "messagebus.service": { "name": "messagebus.service", "source": "systemd", "state": "active", "status": "static" }, "microcode.service": { "name": "microcode.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "nfs-blkmap.service": { "name": "nfs-blkmap.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-convert.service": { "name": "nfs-convert.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-idmapd.service": { "name": "nfs-idmapd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-mountd.service": { "name": "nfs-mountd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-server.service": { "name": "nfs-server.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-utils.service": { "name": "nfs-utils.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfsdcld.service": { "name": "nfsdcld.service", "source": "systemd", "state": "stopped", "status": "static" }, "nftables.service": { "name": "nftables.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nis-domainname.service": { "name": "nis-domainname.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "oddjobd.service": { "name": "oddjobd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "plymouth-halt.service": { "name": "plymouth-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-kexec.service": { "name": "plymouth-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-poweroff.service": { "name": "plymouth-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-quit-wait.service": { "name": "plymouth-quit-wait.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-quit.service": { "name": "plymouth-quit.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-read-write.service": { "name": "plymouth-read-write.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-reboot.service": { "name": "plymouth-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-start.service": { "name": "plymouth-start.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-switch-root-initramfs.service": { "name": "plymouth-switch-root-initramfs.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-switch-root.service": { "name": "plymouth-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "podman-auto-update.service": { "name": "podman-auto-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "podman-clean-transient.service": { "name": "podman-clean-transient.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "podman-kube@.service": { "name": "podman-kube@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "podman-restart.service": { "name": "podman-restart.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "podman.service": { "name": "podman.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "polkit.service": { "name": "polkit.service", "source": "systemd", "state": "running", "status": "static" }, "qemu-guest-agent.service": { "name": "qemu-guest-agent.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "quadlet-demo-mysql-volume.service": { "name": "quadlet-demo-mysql-volume.service", "source": "systemd", "state": "stopped", "status": "generated" }, "quadlet-demo-mysql.service": { "name": "quadlet-demo-mysql.service", "source": "systemd", "state": "running", "status": "generated" }, "quadlet-demo-network.service": { "name": "quadlet-demo-network.service", "source": "systemd", "state": "stopped", "status": "generated" }, "quotaon.service": { "name": "quotaon.service", "source": "systemd", "state": "inactive", "status": "static" }, "rc-local.service": { "name": "rc-local.service", "source": "systemd", "state": "stopped", "status": "static" }, "rdisc.service": { "name": "rdisc.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rescue.service": { "name": "rescue.service", "source": "systemd", "state": "stopped", "status": "static" }, "restraintd.service": { "name": "restraintd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rngd.service": { "name": "rngd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpc-gssd.service": { "name": "rpc-gssd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd-notify.service": { "name": "rpc-statd-notify.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd.service": { "name": "rpc-statd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpcbind.service": { "name": "rpcbind.service", "source": "systemd", "state": "running", "status": "enabled" }, "rsyslog.service": { "name": "rsyslog.service", "source": "systemd", "state": "running", "status": "enabled" }, "selinux-autorelabel-mark.service": { "name": "selinux-autorelabel-mark.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "selinux-autorelabel.service": { "name": "selinux-autorelabel.service", "source": "systemd", "state": "inactive", "status": "static" }, "serial-getty@.service": { "name": "serial-getty@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@.service": { "name": "sshd-keygen@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@ecdsa.service": { "name": "sshd-keygen@ecdsa.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd-keygen@ed25519.service": { "name": "sshd-keygen@ed25519.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd-keygen@rsa.service": { "name": "sshd-keygen@rsa.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd.service": { "name": "sshd.service", "source": "systemd", "state": "running", "status": "enabled" }, "sshd@.service": { "name": "sshd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "sssd-autofs.service": { "name": "sssd-autofs.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-kcm.service": { "name": "sssd-kcm.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "sssd-nss.service": { "name": "sssd-nss.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pac.service": { "name": "sssd-pac.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pam.service": { "name": "sssd-pam.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-ssh.service": { "name": "sssd-ssh.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-sudo.service": { "name": "sssd-sudo.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd.service": { "name": "sssd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "syslog.service": { "name": "syslog.service", "source": "systemd", "state": "active", "status": "enabled" }, "system-update-cleanup.service": { "name": "system-update-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-ask-password-console.service": { "name": "systemd-ask-password-console.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-plymouth.service": { "name": "systemd-ask-password-plymouth.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-wall.service": { "name": "systemd-ask-password-wall.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-backlight@.service": { "name": "systemd-backlight@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-binfmt.service": { "name": "systemd-binfmt.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-coredump@.service": { "name": "systemd-coredump@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-exit.service": { "name": "systemd-exit.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-firstboot.service": { "name": "systemd-firstboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck-root.service": { "name": "systemd-fsck-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck@.service": { "name": "systemd-fsck@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-halt.service": { "name": "systemd-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hibernate-resume@.service": { "name": "systemd-hibernate-resume@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-hibernate.service": { "name": "systemd-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hostnamed.service": { "name": "systemd-hostnamed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hwdb-update.service": { "name": "systemd-hwdb-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-hybrid-sleep.service": { "name": "systemd-hybrid-sleep.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-initctl.service": { "name": "systemd-initctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-catalog-update.service": { "name": "systemd-journal-catalog-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-flush.service": { "name": "systemd-journal-flush.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journald.service": { "name": "systemd-journald.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-kexec.service": { "name": "systemd-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-localed.service": { "name": "systemd-localed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-logind.service": { "name": "systemd-logind.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-machine-id-commit.service": { "name": "systemd-machine-id-commit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-modules-load.service": { "name": "systemd-modules-load.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-portabled.service": { "name": "systemd-portabled.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-poweroff.service": { "name": "systemd-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pstore.service": { "name": "systemd-pstore.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-quotacheck.service": { "name": "systemd-quotacheck.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-random-seed.service": { "name": "systemd-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-reboot.service": { "name": "systemd-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-remount-fs.service": { "name": "systemd-remount-fs.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-resolved.service": { "name": "systemd-resolved.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-rfkill.service": { "name": "systemd-rfkill.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend-then-hibernate.service": { "name": "systemd-suspend-then-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend.service": { "name": "systemd-suspend.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-sysctl.service": { "name": "systemd-sysctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-sysusers.service": { "name": "systemd-sysusers.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-timedated.service": { "name": "systemd-timedated.service", "source": "systemd", "state": "inactive", "status": "masked" }, "systemd-tmpfiles-clean.service": { "name": "systemd-tmpfiles-clean.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev.service": { "name": "systemd-tmpfiles-setup-dev.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup.service": { "name": "systemd-tmpfiles-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-settle.service": { "name": "systemd-udev-settle.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-udev-trigger.service": { "name": "systemd-udev-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udevd.service": { "name": "systemd-udevd.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-update-done.service": { "name": "systemd-update-done.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp-runlevel.service": { "name": "systemd-update-utmp-runlevel.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp.service": { "name": "systemd-update-utmp.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-user-sessions.service": { "name": "systemd-user-sessions.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-vconsole-setup.service": { "name": "systemd-vconsole-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-volatile-root.service": { "name": "systemd-volatile-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "tcsd.service": { "name": "tcsd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "teamd@.service": { "name": "teamd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "timedatex.service": { "name": "timedatex.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "tuned.service": { "name": "tuned.service", "source": "systemd", "state": "running", "status": "enabled" }, "unbound-anchor.service": { "name": "unbound-anchor.service", "source": "systemd", "state": "stopped", "status": "static" }, "user-runtime-dir@.service": { "name": "user-runtime-dir@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user-runtime-dir@0.service": { "name": "user-runtime-dir@0.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "user@.service": { "name": "user@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user@0.service": { "name": "user@0.service", "source": "systemd", "state": "running", "status": "unknown" } } }, "changed": false } TASK [fedora.linux_system_roles.podman : Create and update quadlets] *********** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:114 Saturday 24 August 2024 12:36:04 -0400 (0:00:01.589) 0:02:12.656 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14 Saturday 24 August 2024 12:36:04 -0400 (0:00:00.034) 0:02:12.691 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_file_src": "envoy-proxy-configmap.yml", "__podman_quadlet_spec": {}, "__podman_quadlet_str": "---\napiVersion: v1\nkind: ConfigMap\nmetadata:\n name: envoy-proxy-config\ndata:\n envoy.yaml: |\n admin:\n address:\n socket_address:\n address: 0.0.0.0\n port_value: 9901\n\n static_resources:\n listeners:\n - name: listener_0\n address:\n socket_address:\n address: 0.0.0.0\n port_value: 8080\n filter_chains:\n - filters:\n - name: envoy.filters.network.http_connection_manager\n typed_config:\n \"@type\": type.googleapis.com/envoy.extensions.filters.network.http_connection_manager.v3.HttpConnectionManager\n stat_prefix: ingress_http\n codec_type: AUTO\n route_config:\n name: local_route\n virtual_hosts:\n - name: local_service\n domains: [\"*\"]\n routes:\n - match:\n prefix: \"/\"\n route:\n cluster: backend\n http_filters:\n - name: envoy.filters.http.router\n typed_config:\n \"@type\": type.googleapis.com/envoy.extensions.filters.http.router.v3.Router\n transport_socket:\n name: envoy.transport_sockets.tls\n typed_config:\n \"@type\": type.googleapis.com/envoy.extensions.transport_sockets.tls.v3.DownstreamTlsContext\n common_tls_context:\n tls_certificates:\n - certificate_chain:\n filename: /etc/envoy-certificates/certificate.pem\n private_key:\n filename: /etc/envoy-certificates/certificate.key\n clusters:\n - name: backend\n connect_timeout: 5s\n type: STATIC\n dns_refresh_rate: 1800s\n lb_policy: ROUND_ROBIN\n load_assignment:\n cluster_name: backend\n endpoints:\n - lb_endpoints:\n - endpoint:\n address:\n socket_address:\n address: 127.0.0.1\n port_value: 80", "__podman_quadlet_template_src": "" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25 Saturday 24 August 2024 12:36:04 -0400 (0:00:00.087) 0:02:12.779 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_pull_image": true, "__podman_state": "absent", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] ***** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35 Saturday 24 August 2024 12:36:04 -0400 (0:00:00.041) 0:02:12.821 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48 Saturday 24 August 2024 12:36:04 -0400 (0:00:00.037) 0:02:12.858 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_name": "envoy-proxy-configmap", "__podman_quadlet_type": "yml", "__podman_rootless": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57 Saturday 24 August 2024 12:36:04 -0400 (0:00:00.048) 0:02:12.906 ******* included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Saturday 24 August 2024 12:36:04 -0400 (0:00:00.063) 0:02:12.969 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Saturday 24 August 2024 12:36:04 -0400 (0:00:00.041) 0:02:13.011 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Saturday 24 August 2024 12:36:04 -0400 (0:00:00.041) 0:02:13.052 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get group information] **************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:28 Saturday 24 August 2024 12:36:04 -0400 (0:00:00.048) 0:02:13.101 ******* ok: [managed_node1] => { "ansible_facts": { "getent_group": { "root": [ "x", "0", "" ] } }, "changed": false } TASK [fedora.linux_system_roles.podman : Set group name] *********************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:35 Saturday 24 August 2024 12:36:04 -0400 (0:00:00.370) 0:02:13.472 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_group_name": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:39 Saturday 24 August 2024 12:36:04 -0400 (0:00:00.042) 0:02:13.514 ******* ok: [managed_node1] => { "changed": false, "stat": { "atime": 1724516929.7168462, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "bb5b46ffbafcaa8c4021f3c8b3cb8594f48ef34b", "ctime": 1724516896.725572, "dev": 51713, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 6986657, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-sharedlib", "mode": "0755", "mtime": 1700557386.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 12640, "uid": 0, "version": "1255679238", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************ task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:50 Saturday 24 August 2024 12:36:05 -0400 (0:00:00.357) 0:02:13.871 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check group with getsubids] *********** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:55 Saturday 24 August 2024 12:36:05 -0400 (0:00:00.033) 0:02:13.905 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:60 Saturday 24 August 2024 12:36:05 -0400 (0:00:00.032) 0:02:13.937 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subuid file] ********************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:74 Saturday 24 August 2024 12:36:05 -0400 (0:00:00.071) 0:02:14.008 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subgid file] ********************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:79 Saturday 24 August 2024 12:36:05 -0400 (0:00:00.034) 0:02:14.042 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:84 Saturday 24 August 2024 12:36:05 -0400 (0:00:00.033) 0:02:14.076 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:94 Saturday 24 August 2024 12:36:05 -0400 (0:00:00.039) 0:02:14.115 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] ***** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:101 Saturday 24 August 2024 12:36:05 -0400 (0:00:00.034) 0:02:14.149 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62 Saturday 24 August 2024 12:36:05 -0400 (0:00:00.032) 0:02:14.182 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_activate_systemd_unit": true, "__podman_images_found": [], "__podman_kube_yamls_raw": "", "__podman_service_name": "", "__podman_systemd_scope": "system", "__podman_user_home_dir": "/root", "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73 Saturday 24 August 2024 12:36:05 -0400 (0:00:00.060) 0:02:14.243 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_path": "/etc/containers/systemd" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get kube yaml contents] *************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77 Saturday 24 August 2024 12:36:05 -0400 (0:00:00.036) 0:02:14.279 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87 Saturday 24 August 2024 12:36:05 -0400 (0:00:00.032) 0:02:14.311 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_images": [], "__podman_quadlet_file": "/etc/containers/systemd/envoy-proxy-configmap.yml", "__podman_volumes": [] }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 6] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:103 Saturday 24 August 2024 12:36:05 -0400 (0:00:00.074) 0:02:14.386 ******* ok: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Cleanup quadlets] ********************* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:110 Saturday 24 August 2024 12:36:05 -0400 (0:00:00.039) 0:02:14.425 ******* included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Stat XDG_RUNTIME_DIR] ***************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:4 Saturday 24 August 2024 12:36:05 -0400 (0:00:00.076) 0:02:14.501 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Stop and disable service] ************* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:12 Saturday 24 August 2024 12:36:05 -0400 (0:00:00.033) 0:02:14.535 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : See if quadlet file exists] *********** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:33 Saturday 24 August 2024 12:36:05 -0400 (0:00:00.035) 0:02:14.570 ******* ok: [managed_node1] => { "changed": false, "stat": { "atime": 1724517322.898056, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 8, "charset": "us-ascii", "checksum": "d681c7d56f912150d041873e880818b22a90c188", "ctime": 1724517292.8068147, "dev": 51713, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 448790699, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "text/plain", "mode": "0644", "mtime": 1724517292.5298123, "nlink": 1, "path": "/etc/containers/systemd/envoy-proxy-configmap.yml", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 2102, "uid": 0, "version": "2763400557", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [fedora.linux_system_roles.podman : Parse quadlet file] ******************* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:38 Saturday 24 August 2024 12:36:06 -0400 (0:00:00.384) 0:02:14.955 ******* included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Slurp quadlet file] ******************* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml:6 Saturday 24 August 2024 12:36:06 -0400 (0:00:00.061) 0:02:15.016 ******* ok: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Parse quadlet file] ******************* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml:12 Saturday 24 August 2024 12:36:06 -0400 (0:00:00.348) 0:02:15.365 ******* skipping: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Parse quadlet yaml file] ************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml:44 Saturday 24 August 2024 12:36:06 -0400 (0:00:00.033) 0:02:15.399 ******* ok: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Reset raw variable] ******************* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml:52 Saturday 24 August 2024 12:36:06 -0400 (0:00:00.038) 0:02:15.438 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_raw": null }, "changed": false } TASK [fedora.linux_system_roles.podman : Remove quadlet file] ****************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:42 Saturday 24 August 2024 12:36:06 -0400 (0:00:00.033) 0:02:15.471 ******* changed: [managed_node1] => { "changed": true, "path": "/etc/containers/systemd/envoy-proxy-configmap.yml", "state": "absent" } TASK [fedora.linux_system_roles.podman : Refresh systemd] ********************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:48 Saturday 24 August 2024 12:36:07 -0400 (0:00:00.360) 0:02:15.832 ******* ok: [managed_node1] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.podman : Remove managed resource] ************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:58 Saturday 24 August 2024 12:36:07 -0400 (0:00:00.597) 0:02:16.429 ******* skipping: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Remove volumes] *********************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:95 Saturday 24 August 2024 12:36:07 -0400 (0:00:00.036) 0:02:16.465 ******* skipping: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Clear parsed podman variable] ********* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:112 Saturday 24 August 2024 12:36:07 -0400 (0:00:00.046) 0:02:16.511 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_parsed": null }, "changed": false } TASK [fedora.linux_system_roles.podman : Prune images no longer in use] ******** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:116 Saturday 24 August 2024 12:36:07 -0400 (0:00:00.069) 0:02:16.581 ******* changed: [managed_node1] => { "changed": true, "cmd": [ "podman", "image", "prune", "--all", "-f" ], "delta": "0:00:00.030343", "end": "2024-08-24 12:36:08.291158", "rc": 0, "start": "2024-08-24 12:36:08.260815" } TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:127 Saturday 24 August 2024 12:36:08 -0400 (0:00:00.392) 0:02:16.973 ******* included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Saturday 24 August 2024 12:36:08 -0400 (0:00:00.058) 0:02:17.032 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Saturday 24 August 2024 12:36:08 -0400 (0:00:00.033) 0:02:17.066 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Saturday 24 August 2024 12:36:08 -0400 (0:00:00.032) 0:02:17.098 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : For testing and debugging - images] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:137 Saturday 24 August 2024 12:36:08 -0400 (0:00:00.038) 0:02:17.136 ******* ok: [managed_node1] => { "changed": false, "cmd": [ "podman", "images", "-n" ], "delta": "0:00:00.031808", "end": "2024-08-24 12:36:08.852358", "rc": 0, "start": "2024-08-24 12:36:08.820550" } STDOUT: quay.io/linux-system-roles/mysql 5.6 dd3b2a5dcb48 2 years ago 308 MB TASK [fedora.linux_system_roles.podman : For testing and debugging - volumes] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:146 Saturday 24 August 2024 12:36:08 -0400 (0:00:00.399) 0:02:17.536 ******* ok: [managed_node1] => { "changed": false, "cmd": [ "podman", "volume", "ls", "-n" ], "delta": "0:00:00.029144", "end": "2024-08-24 12:36:09.242928", "rc": 0, "start": "2024-08-24 12:36:09.213784" } STDOUT: local systemd-quadlet-demo-mysql local wp-pv-claim local envoy-proxy-config local envoy-certificates TASK [fedora.linux_system_roles.podman : For testing and debugging - containers] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:155 Saturday 24 August 2024 12:36:09 -0400 (0:00:00.390) 0:02:17.926 ******* ok: [managed_node1] => { "changed": false, "cmd": [ "podman", "ps", "--noheading" ], "delta": "0:00:00.035472", "end": "2024-08-24 12:36:09.639561", "rc": 0, "start": "2024-08-24 12:36:09.604089" } STDOUT: 68250d5fc021 quay.io/linux-system-roles/mysql:5.6 mysqld About a minute ago Up About a minute (healthy) quadlet-demo-mysql TASK [fedora.linux_system_roles.podman : For testing and debugging - networks] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:164 Saturday 24 August 2024 12:36:09 -0400 (0:00:00.396) 0:02:18.323 ******* ok: [managed_node1] => { "changed": false, "cmd": [ "podman", "network", "ls", "-n", "-q" ], "delta": "0:00:00.059485", "end": "2024-08-24 12:36:10.065687", "rc": 0, "start": "2024-08-24 12:36:10.006202" } STDOUT: podman podman-default-kube-network systemd-quadlet-demo TASK [fedora.linux_system_roles.podman : For testing and debugging - secrets] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:173 Saturday 24 August 2024 12:36:10 -0400 (0:00:00.426) 0:02:18.749 ******* ok: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : For testing and debugging - services] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:183 Saturday 24 August 2024 12:36:10 -0400 (0:00:00.396) 0:02:19.145 ******* ok: [managed_node1] => { "ansible_facts": { "services": { "68250d5fc0218bc6db3ab2...5ebe15c37a833d3345ad2e2df2c1916dd6cc3.service": { "name": "68250d5fc0218bc6db3ab2...5ebe15c37a833d3345ad2e2df2c1916dd6cc3.service", "source": "systemd", "state": "inactive", "status": "transient" }, "68250d5fc0218bc6db3ab2ece605ebe15c37a833d3345ad2e2df2c1916dd6cc3.service": { "name": "68250d5fc0218bc6db3ab2ece605ebe15c37a833d3345ad2e2df2c1916dd6cc3.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "NetworkManager-dispatcher.service": { "name": "NetworkManager-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "NetworkManager-wait-online.service": { "name": "NetworkManager-wait-online.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "NetworkManager.service": { "name": "NetworkManager.service", "source": "systemd", "state": "running", "status": "enabled" }, "auditd.service": { "name": "auditd.service", "source": "systemd", "state": "running", "status": "enabled" }, "auth-rpcgss-module.service": { "name": "auth-rpcgss-module.service", "source": "systemd", "state": "stopped", "status": "static" }, "autovt@.service": { "name": "autovt@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "certmonger.service": { "name": "certmonger.service", "source": "systemd", "state": "running", "status": "enabled" }, "chrony-dnssrv@.service": { "name": "chrony-dnssrv@.service", "source": "systemd", "state": "unknown", "status": "static" }, "chrony-wait.service": { "name": "chrony-wait.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd.service": { "name": "chronyd.service", "source": "systemd", "state": "running", "status": "enabled" }, "cloud-config.service": { "name": "cloud-config.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-final.service": { "name": "cloud-final.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init-hotplugd.service": { "name": "cloud-init-hotplugd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cloud-init-local.service": { "name": "cloud-init-local.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init.service": { "name": "cloud-init.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cni-dhcp.service": { "name": "cni-dhcp.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "console-getty.service": { "name": "console-getty.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "container-getty@.service": { "name": "container-getty@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cpupower.service": { "name": "cpupower.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "crond.service": { "name": "crond.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-org.fedoraproject.FirewallD1.service": { "name": "dbus-org.fedoraproject.FirewallD1.service", "source": "systemd", "state": "active", "status": "enabled" }, "dbus-org.freedesktop.hostname1.service": { "name": "dbus-org.freedesktop.hostname1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.locale1.service": { "name": "dbus-org.freedesktop.locale1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.login1.service": { "name": "dbus-org.freedesktop.login1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.nm-dispatcher.service": { "name": "dbus-org.freedesktop.nm-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "dbus-org.freedesktop.portable1.service": { "name": "dbus-org.freedesktop.portable1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.timedate1.service": { "name": "dbus-org.freedesktop.timedate1.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "dbus.service": { "name": "dbus.service", "source": "systemd", "state": "running", "status": "static" }, "debug-shell.service": { "name": "debug-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dnf-makecache.service": { "name": "dnf-makecache.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-system-upgrade-cleanup.service": { "name": "dnf-system-upgrade-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "dnf-system-upgrade.service": { "name": "dnf-system-upgrade.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dnsmasq.service": { "name": "dnsmasq.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dracut-cmdline.service": { "name": "dracut-cmdline.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-initqueue.service": { "name": "dracut-initqueue.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-mount.service": { "name": "dracut-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-mount.service": { "name": "dracut-pre-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-pivot.service": { "name": "dracut-pre-pivot.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-trigger.service": { "name": "dracut-pre-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-udev.service": { "name": "dracut-pre-udev.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown-onfailure.service": { "name": "dracut-shutdown-onfailure.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown.service": { "name": "dracut-shutdown.service", "source": "systemd", "state": "stopped", "status": "static" }, "ebtables.service": { "name": "ebtables.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "emergency.service": { "name": "emergency.service", "source": "systemd", "state": "stopped", "status": "static" }, "firewalld.service": { "name": "firewalld.service", "source": "systemd", "state": "running", "status": "enabled" }, "fstrim.service": { "name": "fstrim.service", "source": "systemd", "state": "inactive", "status": "static" }, "getty@.service": { "name": "getty@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "getty@tty1.service": { "name": "getty@tty1.service", "source": "systemd", "state": "running", "status": "unknown" }, "grub-boot-indeterminate.service": { "name": "grub-boot-indeterminate.service", "source": "systemd", "state": "inactive", "status": "static" }, "gssproxy.service": { "name": "gssproxy.service", "source": "systemd", "state": "running", "status": "disabled" }, "halt-local.service": { "name": "halt-local.service", "source": "systemd", "state": "inactive", "status": "static" }, "import-state.service": { "name": "import-state.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "initrd-cleanup.service": { "name": "initrd-cleanup.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-parse-etc.service": { "name": "initrd-parse-etc.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-switch-root.service": { "name": "initrd-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-udevadm-cleanup-db.service": { "name": "initrd-udevadm-cleanup-db.service", "source": "systemd", "state": "stopped", "status": "static" }, "iprdump.service": { "name": "iprdump.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "iprinit.service": { "name": "iprinit.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "iprupdate.service": { "name": "iprupdate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "irqbalance.service": { "name": "irqbalance.service", "source": "systemd", "state": "running", "status": "enabled" }, "kdump.service": { "name": "kdump.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "kmod-static-nodes.service": { "name": "kmod-static-nodes.service", "source": "systemd", "state": "stopped", "status": "static" }, "kvm_stat.service": { "name": "kvm_stat.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "ldconfig.service": { "name": "ldconfig.service", "source": "systemd", "state": "stopped", "status": "static" }, "loadmodules.service": { "name": "loadmodules.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "man-db-cache-update.service": { "name": "man-db-cache-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "man-db-restart-cache-update.service": { "name": "man-db-restart-cache-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "messagebus.service": { "name": "messagebus.service", "source": "systemd", "state": "active", "status": "static" }, "microcode.service": { "name": "microcode.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "nfs-blkmap.service": { "name": "nfs-blkmap.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-convert.service": { "name": "nfs-convert.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-idmapd.service": { "name": "nfs-idmapd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-mountd.service": { "name": "nfs-mountd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-server.service": { "name": "nfs-server.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-utils.service": { "name": "nfs-utils.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfsdcld.service": { "name": "nfsdcld.service", "source": "systemd", "state": "stopped", "status": "static" }, "nftables.service": { "name": "nftables.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nis-domainname.service": { "name": "nis-domainname.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "oddjobd.service": { "name": "oddjobd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "plymouth-halt.service": { "name": "plymouth-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-kexec.service": { "name": "plymouth-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-poweroff.service": { "name": "plymouth-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-quit-wait.service": { "name": "plymouth-quit-wait.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-quit.service": { "name": "plymouth-quit.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-read-write.service": { "name": "plymouth-read-write.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-reboot.service": { "name": "plymouth-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-start.service": { "name": "plymouth-start.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-switch-root-initramfs.service": { "name": "plymouth-switch-root-initramfs.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-switch-root.service": { "name": "plymouth-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "podman-auto-update.service": { "name": "podman-auto-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "podman-clean-transient.service": { "name": "podman-clean-transient.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "podman-kube@.service": { "name": "podman-kube@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "podman-restart.service": { "name": "podman-restart.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "podman.service": { "name": "podman.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "polkit.service": { "name": "polkit.service", "source": "systemd", "state": "running", "status": "static" }, "qemu-guest-agent.service": { "name": "qemu-guest-agent.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "quadlet-demo-mysql-volume.service": { "name": "quadlet-demo-mysql-volume.service", "source": "systemd", "state": "stopped", "status": "generated" }, "quadlet-demo-mysql.service": { "name": "quadlet-demo-mysql.service", "source": "systemd", "state": "running", "status": "generated" }, "quadlet-demo-network.service": { "name": "quadlet-demo-network.service", "source": "systemd", "state": "stopped", "status": "generated" }, "quotaon.service": { "name": "quotaon.service", "source": "systemd", "state": "inactive", "status": "static" }, "rc-local.service": { "name": "rc-local.service", "source": "systemd", "state": "stopped", "status": "static" }, "rdisc.service": { "name": "rdisc.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rescue.service": { "name": "rescue.service", "source": "systemd", "state": "stopped", "status": "static" }, "restraintd.service": { "name": "restraintd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rngd.service": { "name": "rngd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpc-gssd.service": { "name": "rpc-gssd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd-notify.service": { "name": "rpc-statd-notify.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd.service": { "name": "rpc-statd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpcbind.service": { "name": "rpcbind.service", "source": "systemd", "state": "running", "status": "enabled" }, "rsyslog.service": { "name": "rsyslog.service", "source": "systemd", "state": "running", "status": "enabled" }, "selinux-autorelabel-mark.service": { "name": "selinux-autorelabel-mark.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "selinux-autorelabel.service": { "name": "selinux-autorelabel.service", "source": "systemd", "state": "inactive", "status": "static" }, "serial-getty@.service": { "name": "serial-getty@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@.service": { "name": "sshd-keygen@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@ecdsa.service": { "name": "sshd-keygen@ecdsa.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd-keygen@ed25519.service": { "name": "sshd-keygen@ed25519.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd-keygen@rsa.service": { "name": "sshd-keygen@rsa.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd.service": { "name": "sshd.service", "source": "systemd", "state": "running", "status": "enabled" }, "sshd@.service": { "name": "sshd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "sssd-autofs.service": { "name": "sssd-autofs.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-kcm.service": { "name": "sssd-kcm.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "sssd-nss.service": { "name": "sssd-nss.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pac.service": { "name": "sssd-pac.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pam.service": { "name": "sssd-pam.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-ssh.service": { "name": "sssd-ssh.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-sudo.service": { "name": "sssd-sudo.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd.service": { "name": "sssd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "syslog.service": { "name": "syslog.service", "source": "systemd", "state": "active", "status": "enabled" }, "system-update-cleanup.service": { "name": "system-update-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-ask-password-console.service": { "name": "systemd-ask-password-console.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-plymouth.service": { "name": "systemd-ask-password-plymouth.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-wall.service": { "name": "systemd-ask-password-wall.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-backlight@.service": { "name": "systemd-backlight@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-binfmt.service": { "name": "systemd-binfmt.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-coredump@.service": { "name": "systemd-coredump@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-exit.service": { "name": "systemd-exit.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-firstboot.service": { "name": "systemd-firstboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck-root.service": { "name": "systemd-fsck-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck@.service": { "name": "systemd-fsck@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-halt.service": { "name": "systemd-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hibernate-resume@.service": { "name": "systemd-hibernate-resume@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-hibernate.service": { "name": "systemd-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hostnamed.service": { "name": "systemd-hostnamed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hwdb-update.service": { "name": "systemd-hwdb-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-hybrid-sleep.service": { "name": "systemd-hybrid-sleep.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-initctl.service": { "name": "systemd-initctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-catalog-update.service": { "name": "systemd-journal-catalog-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-flush.service": { "name": "systemd-journal-flush.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journald.service": { "name": "systemd-journald.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-kexec.service": { "name": "systemd-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-localed.service": { "name": "systemd-localed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-logind.service": { "name": "systemd-logind.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-machine-id-commit.service": { "name": "systemd-machine-id-commit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-modules-load.service": { "name": "systemd-modules-load.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-portabled.service": { "name": "systemd-portabled.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-poweroff.service": { "name": "systemd-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pstore.service": { "name": "systemd-pstore.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-quotacheck.service": { "name": "systemd-quotacheck.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-random-seed.service": { "name": "systemd-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-reboot.service": { "name": "systemd-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-remount-fs.service": { "name": "systemd-remount-fs.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-resolved.service": { "name": "systemd-resolved.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-rfkill.service": { "name": "systemd-rfkill.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend-then-hibernate.service": { "name": "systemd-suspend-then-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend.service": { "name": "systemd-suspend.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-sysctl.service": { "name": "systemd-sysctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-sysusers.service": { "name": "systemd-sysusers.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-timedated.service": { "name": "systemd-timedated.service", "source": "systemd", "state": "inactive", "status": "masked" }, "systemd-tmpfiles-clean.service": { "name": "systemd-tmpfiles-clean.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev.service": { "name": "systemd-tmpfiles-setup-dev.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup.service": { "name": "systemd-tmpfiles-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-settle.service": { "name": "systemd-udev-settle.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-udev-trigger.service": { "name": "systemd-udev-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udevd.service": { "name": "systemd-udevd.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-update-done.service": { "name": "systemd-update-done.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp-runlevel.service": { "name": "systemd-update-utmp-runlevel.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp.service": { "name": "systemd-update-utmp.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-user-sessions.service": { "name": "systemd-user-sessions.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-vconsole-setup.service": { "name": "systemd-vconsole-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-volatile-root.service": { "name": "systemd-volatile-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "tcsd.service": { "name": "tcsd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "teamd@.service": { "name": "teamd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "timedatex.service": { "name": "timedatex.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "tuned.service": { "name": "tuned.service", "source": "systemd", "state": "running", "status": "enabled" }, "unbound-anchor.service": { "name": "unbound-anchor.service", "source": "systemd", "state": "stopped", "status": "static" }, "user-runtime-dir@.service": { "name": "user-runtime-dir@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user-runtime-dir@0.service": { "name": "user-runtime-dir@0.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "user@.service": { "name": "user@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user@0.service": { "name": "user@0.service", "source": "systemd", "state": "running", "status": "unknown" } } }, "changed": false } TASK [fedora.linux_system_roles.podman : Create and update quadlets] *********** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:114 Saturday 24 August 2024 12:36:12 -0400 (0:00:01.601) 0:02:20.747 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14 Saturday 24 August 2024 12:36:12 -0400 (0:00:00.033) 0:02:20.780 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_file_src": "", "__podman_quadlet_spec": {}, "__podman_quadlet_str": "[Install]\nWantedBy=default.target\n\n[Container]\nImage=quay.io/linux-system-roles/mysql:5.6\nContainerName=quadlet-demo-mysql\nVolume=quadlet-demo-mysql.volume:/var/lib/mysql\nVolume=/tmp/quadlet_demo:/var/lib/quadlet_demo:Z\nNetwork=quadlet-demo.network\nSecret=mysql-root-password-container,type=env,target=MYSQL_ROOT_PASSWORD\nHealthCmd=/bin/true\nHealthOnFailure=kill\n", "__podman_quadlet_template_src": "quadlet-demo-mysql.container.j2" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25 Saturday 24 August 2024 12:36:12 -0400 (0:00:00.107) 0:02:20.888 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_pull_image": true, "__podman_state": "absent", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] ***** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35 Saturday 24 August 2024 12:36:12 -0400 (0:00:00.042) 0:02:20.930 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48 Saturday 24 August 2024 12:36:12 -0400 (0:00:00.035) 0:02:20.965 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_name": "quadlet-demo-mysql", "__podman_quadlet_type": "container", "__podman_rootless": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57 Saturday 24 August 2024 12:36:12 -0400 (0:00:00.081) 0:02:21.047 ******* included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Saturday 24 August 2024 12:36:12 -0400 (0:00:00.063) 0:02:21.110 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Saturday 24 August 2024 12:36:12 -0400 (0:00:00.040) 0:02:21.151 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Saturday 24 August 2024 12:36:12 -0400 (0:00:00.045) 0:02:21.197 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get group information] **************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:28 Saturday 24 August 2024 12:36:12 -0400 (0:00:00.045) 0:02:21.242 ******* ok: [managed_node1] => { "ansible_facts": { "getent_group": { "root": [ "x", "0", "" ] } }, "changed": false } TASK [fedora.linux_system_roles.podman : Set group name] *********************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:35 Saturday 24 August 2024 12:36:12 -0400 (0:00:00.369) 0:02:21.611 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_group_name": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:39 Saturday 24 August 2024 12:36:13 -0400 (0:00:00.043) 0:02:21.655 ******* ok: [managed_node1] => { "changed": false, "stat": { "atime": 1724516929.7168462, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "bb5b46ffbafcaa8c4021f3c8b3cb8594f48ef34b", "ctime": 1724516896.725572, "dev": 51713, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 6986657, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-sharedlib", "mode": "0755", "mtime": 1700557386.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 12640, "uid": 0, "version": "1255679238", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************ task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:50 Saturday 24 August 2024 12:36:13 -0400 (0:00:00.363) 0:02:22.018 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check group with getsubids] *********** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:55 Saturday 24 August 2024 12:36:13 -0400 (0:00:00.032) 0:02:22.051 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:60 Saturday 24 August 2024 12:36:13 -0400 (0:00:00.033) 0:02:22.084 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subuid file] ********************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:74 Saturday 24 August 2024 12:36:13 -0400 (0:00:00.033) 0:02:22.117 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subgid file] ********************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:79 Saturday 24 August 2024 12:36:13 -0400 (0:00:00.032) 0:02:22.150 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:84 Saturday 24 August 2024 12:36:13 -0400 (0:00:00.077) 0:02:22.227 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:94 Saturday 24 August 2024 12:36:13 -0400 (0:00:00.033) 0:02:22.261 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] ***** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:101 Saturday 24 August 2024 12:36:13 -0400 (0:00:00.035) 0:02:22.296 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62 Saturday 24 August 2024 12:36:13 -0400 (0:00:00.034) 0:02:22.330 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_activate_systemd_unit": true, "__podman_images_found": [ "quay.io/linux-system-roles/mysql:5.6" ], "__podman_kube_yamls_raw": "", "__podman_service_name": "quadlet-demo-mysql.service", "__podman_systemd_scope": "system", "__podman_user_home_dir": "/root", "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73 Saturday 24 August 2024 12:36:13 -0400 (0:00:00.060) 0:02:22.391 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_path": "/etc/containers/systemd" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get kube yaml contents] *************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77 Saturday 24 August 2024 12:36:13 -0400 (0:00:00.035) 0:02:22.426 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87 Saturday 24 August 2024 12:36:13 -0400 (0:00:00.034) 0:02:22.461 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_images": [ "quay.io/linux-system-roles/mysql:5.6" ], "__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo-mysql.container", "__podman_volumes": [ "/tmp/quadlet_demo" ] }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 6] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:103 Saturday 24 August 2024 12:36:13 -0400 (0:00:00.074) 0:02:22.535 ******* ok: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Cleanup quadlets] ********************* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:110 Saturday 24 August 2024 12:36:13 -0400 (0:00:00.039) 0:02:22.574 ******* included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Stat XDG_RUNTIME_DIR] ***************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:4 Saturday 24 August 2024 12:36:14 -0400 (0:00:00.078) 0:02:22.653 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Stop and disable service] ************* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:12 Saturday 24 August 2024 12:36:14 -0400 (0:00:00.034) 0:02:22.687 ******* changed: [managed_node1] => { "changed": true, "enabled": false, "failed_when_result": false, "name": "quadlet-demo-mysql.service", "state": "stopped", "status": { "ActiveEnterTimestamp": "Sat 2024-08-24 12:34:48 EDT", "ActiveEnterTimestampMonotonic": "691670672", "ActiveExitTimestampMonotonic": "0", "ActiveState": "active", "After": "basic.target system.slice quadlet-demo-network.service quadlet-demo-mysql-volume.service systemd-journald.socket tmp.mount -.mount sysinit.target", "AllowIsolate": "no", "AllowedCPUs": "", "AllowedMemoryNodes": "", "AmbientCapabilities": "", "AssertResult": "yes", "AssertTimestamp": "Sat 2024-08-24 12:34:48 EDT", "AssertTimestampMonotonic": "691284549", "Before": "shutdown.target", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "CPUAccounting": "no", "CPUAffinity": "", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "[not set]", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "no", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf", "CollectMode": "inactive", "ConditionResult": "yes", "ConditionTimestamp": "Sat 2024-08-24 12:34:48 EDT", "ConditionTimestampMonotonic": "691284547", "ConfigurationDirectoryMode": "0755", "Conflicts": "shutdown.target", "ControlGroup": "/system.slice/quadlet-demo-mysql.service", "ControlPID": "0", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "Delegate": "yes", "DelegateControllers": "cpu cpuacct cpuset io blkio memory devices pids", "Description": "quadlet-demo-mysql.service", "DevicePolicy": "auto", "DynamicUser": "no", "EffectiveCPUs": "", "EffectiveMemoryNodes": "", "Environment": "PODMAN_SYSTEMD_UNIT=quadlet-demo-mysql.service", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "84084", "ExecMainStartTimestamp": "Sat 2024-08-24 12:34:48 EDT", "ExecMainStartTimestampMonotonic": "691670652", "ExecMainStatus": "0", "ExecStart": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman run --name=quadlet-demo-mysql --cidfile=/run/quadlet-demo-mysql.cid --replace --rm --cgroups=split --network=systemd-quadlet-demo --sdnotify=conmon -d -v systemd-quadlet-demo-mysql:/var/lib/mysql -v /tmp/quadlet_demo:/var/lib/quadlet_demo:Z --secret mysql-root-password-container,type=env,target=MYSQL_ROOT_PASSWORD --health-cmd /bin/true --health-on-failure kill quay.io/linux-system-roles/mysql:5.6 ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStop": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman rm -v -f -i --cidfile=/run/quadlet-demo-mysql.cid ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStopPost": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman rm -v -f -i --cidfile=/run/quadlet-demo-mysql.cid ; ignore_errors=yes ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FragmentPath": "/run/systemd/generator/quadlet-demo-mysql.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOSchedulingClass": "0", "IOSchedulingPriority": "0", "IOWeight": "[not set]", "IPAccounting": "no", "IPEgressBytes": "18446744073709551615", "IPEgressPackets": "18446744073709551615", "IPIngressBytes": "18446744073709551615", "IPIngressPackets": "18446744073709551615", "Id": "quadlet-demo-mysql.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestamp": "Sat 2024-08-24 12:34:48 EDT", "InactiveExitTimestampMonotonic": "691285998", "InvocationID": "f636d1c329514adfa01cbd36661564f5", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "mixed", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "0", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "65536", "LimitMEMLOCKSoft": "65536", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "262144", "LimitNOFILESoft": "1024", "LimitNPROC": "14003", "LimitNPROCSoft": "14003", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "14003", "LimitSIGPENDINGSoft": "14003", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "84084", "MemoryAccounting": "yes", "MemoryCurrent": "604258304", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemorySwapMax": "infinity", "MountAPIVFS": "no", "MountFlags": "", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAMask": "", "NUMAPolicy": "n/a", "Names": "quadlet-demo-mysql.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "all", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "PermissionsStartOnly": "no", "Perpetual": "no", "PrivateDevices": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "RemainAfterExit": "no", "RemoveIPC": "no", "Requires": "quadlet-demo-mysql-volume.service sysinit.target system.slice -.mount quadlet-demo-network.service", "RequiresMountsFor": "/tmp/quadlet_demo /run/containers", "Restart": "no", "RestartUSec": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "Slice": "system.slice", "SourcePath": "/etc/containers/systemd/quadlet-demo-mysql.container", "StandardError": "inherit", "StandardInput": "null", "StandardInputData": "", "StandardOutput": "journal", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StateChangeTimestamp": "Sat 2024-08-24 12:34:48 EDT", "StateChangeTimestampMonotonic": "691670672", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "running", "SuccessAction": "none", "SyslogFacility": "3", "SyslogIdentifier": "quadlet-demo-mysql", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "0", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "24", "TasksMax": "22405", "TimeoutStartUSec": "1min 30s", "TimeoutStopUSec": "1min 30s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "notify", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "disabled", "UnitFileState": "generated", "UtmpMode": "init", "WatchdogTimestamp": "Sat 2024-08-24 12:34:48 EDT", "WatchdogTimestampMonotonic": "691670669", "WatchdogUSec": "0" } } TASK [fedora.linux_system_roles.podman : See if quadlet file exists] *********** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:33 Saturday 24 August 2024 12:36:16 -0400 (0:00:01.942) 0:02:24.630 ******* ok: [managed_node1] => { "changed": false, "stat": { "atime": 1724517287.8987753, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 8, "charset": "us-ascii", "checksum": "ca62b2ad3cc9afb5b5371ebbf797b9bc4fd7edd4", "ctime": 1724517287.3017704, "dev": 51713, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 369098918, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "text/plain", "mode": "0644", "mtime": 1724517287.0177681, "nlink": 1, "path": "/etc/containers/systemd/quadlet-demo-mysql.container", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 363, "uid": 0, "version": "4287155550", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [fedora.linux_system_roles.podman : Parse quadlet file] ******************* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:38 Saturday 24 August 2024 12:36:16 -0400 (0:00:00.365) 0:02:24.996 ******* included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Slurp quadlet file] ******************* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml:6 Saturday 24 August 2024 12:36:16 -0400 (0:00:00.061) 0:02:25.057 ******* ok: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Parse quadlet file] ******************* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml:12 Saturday 24 August 2024 12:36:16 -0400 (0:00:00.352) 0:02:25.410 ******* ok: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Parse quadlet yaml file] ************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml:44 Saturday 24 August 2024 12:36:16 -0400 (0:00:00.047) 0:02:25.457 ******* skipping: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Reset raw variable] ******************* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml:52 Saturday 24 August 2024 12:36:16 -0400 (0:00:00.031) 0:02:25.489 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_raw": null }, "changed": false } TASK [fedora.linux_system_roles.podman : Remove quadlet file] ****************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:42 Saturday 24 August 2024 12:36:16 -0400 (0:00:00.032) 0:02:25.521 ******* changed: [managed_node1] => { "changed": true, "path": "/etc/containers/systemd/quadlet-demo-mysql.container", "state": "absent" } TASK [fedora.linux_system_roles.podman : Refresh systemd] ********************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:48 Saturday 24 August 2024 12:36:17 -0400 (0:00:00.360) 0:02:25.882 ******* ok: [managed_node1] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.podman : Remove managed resource] ************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:58 Saturday 24 August 2024 12:36:17 -0400 (0:00:00.656) 0:02:26.538 ******* ok: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Remove volumes] *********************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:95 Saturday 24 August 2024 12:36:18 -0400 (0:00:00.415) 0:02:26.954 ******* skipping: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Clear parsed podman variable] ********* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:112 Saturday 24 August 2024 12:36:18 -0400 (0:00:00.048) 0:02:27.002 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_parsed": null }, "changed": false } TASK [fedora.linux_system_roles.podman : Prune images no longer in use] ******** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:116 Saturday 24 August 2024 12:36:18 -0400 (0:00:00.032) 0:02:27.034 ******* changed: [managed_node1] => { "changed": true, "cmd": [ "podman", "image", "prune", "--all", "-f" ], "delta": "0:00:00.234770", "end": "2024-08-24 12:36:18.949782", "rc": 0, "start": "2024-08-24 12:36:18.715012" } STDOUT: dd3b2a5dcb48ff61113592ed5ddd762581be4387c7bc552375a2159422aa6bf5 TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:127 Saturday 24 August 2024 12:36:19 -0400 (0:00:00.598) 0:02:27.632 ******* included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Saturday 24 August 2024 12:36:19 -0400 (0:00:00.060) 0:02:27.692 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Saturday 24 August 2024 12:36:19 -0400 (0:00:00.032) 0:02:27.724 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Saturday 24 August 2024 12:36:19 -0400 (0:00:00.032) 0:02:27.757 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : For testing and debugging - images] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:137 Saturday 24 August 2024 12:36:19 -0400 (0:00:00.032) 0:02:27.789 ******* ok: [managed_node1] => { "changed": false, "cmd": [ "podman", "images", "-n" ], "delta": "0:00:00.036357", "end": "2024-08-24 12:36:19.506720", "rc": 0, "start": "2024-08-24 12:36:19.470363" } TASK [fedora.linux_system_roles.podman : For testing and debugging - volumes] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:146 Saturday 24 August 2024 12:36:19 -0400 (0:00:00.447) 0:02:28.237 ******* ok: [managed_node1] => { "changed": false, "cmd": [ "podman", "volume", "ls", "-n" ], "delta": "0:00:00.036451", "end": "2024-08-24 12:36:19.956747", "rc": 0, "start": "2024-08-24 12:36:19.920296" } STDOUT: local systemd-quadlet-demo-mysql local wp-pv-claim local envoy-proxy-config local envoy-certificates TASK [fedora.linux_system_roles.podman : For testing and debugging - containers] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:155 Saturday 24 August 2024 12:36:20 -0400 (0:00:00.401) 0:02:28.639 ******* ok: [managed_node1] => { "changed": false, "cmd": [ "podman", "ps", "--noheading" ], "delta": "0:00:00.031885", "end": "2024-08-24 12:36:20.352641", "rc": 0, "start": "2024-08-24 12:36:20.320756" } TASK [fedora.linux_system_roles.podman : For testing and debugging - networks] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:164 Saturday 24 August 2024 12:36:20 -0400 (0:00:00.396) 0:02:29.035 ******* ok: [managed_node1] => { "changed": false, "cmd": [ "podman", "network", "ls", "-n", "-q" ], "delta": "0:00:00.057821", "end": "2024-08-24 12:36:20.775500", "rc": 0, "start": "2024-08-24 12:36:20.717679" } STDOUT: podman podman-default-kube-network systemd-quadlet-demo TASK [fedora.linux_system_roles.podman : For testing and debugging - secrets] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:173 Saturday 24 August 2024 12:36:20 -0400 (0:00:00.422) 0:02:29.458 ******* ok: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : For testing and debugging - services] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:183 Saturday 24 August 2024 12:36:21 -0400 (0:00:00.399) 0:02:29.858 ******* ok: [managed_node1] => { "ansible_facts": { "services": { "NetworkManager-dispatcher.service": { "name": "NetworkManager-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "NetworkManager-wait-online.service": { "name": "NetworkManager-wait-online.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "NetworkManager.service": { "name": "NetworkManager.service", "source": "systemd", "state": "running", "status": "enabled" }, "auditd.service": { "name": "auditd.service", "source": "systemd", "state": "running", "status": "enabled" }, "auth-rpcgss-module.service": { "name": "auth-rpcgss-module.service", "source": "systemd", "state": "stopped", "status": "static" }, "autovt@.service": { "name": "autovt@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "certmonger.service": { "name": "certmonger.service", "source": "systemd", "state": "running", "status": "enabled" }, "chrony-dnssrv@.service": { "name": "chrony-dnssrv@.service", "source": "systemd", "state": "unknown", "status": "static" }, "chrony-wait.service": { "name": "chrony-wait.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd.service": { "name": "chronyd.service", "source": "systemd", "state": "running", "status": "enabled" }, "cloud-config.service": { "name": "cloud-config.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-final.service": { "name": "cloud-final.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init-hotplugd.service": { "name": "cloud-init-hotplugd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cloud-init-local.service": { "name": "cloud-init-local.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init.service": { "name": "cloud-init.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cni-dhcp.service": { "name": "cni-dhcp.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "console-getty.service": { "name": "console-getty.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "container-getty@.service": { "name": "container-getty@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cpupower.service": { "name": "cpupower.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "crond.service": { "name": "crond.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-org.fedoraproject.FirewallD1.service": { "name": "dbus-org.fedoraproject.FirewallD1.service", "source": "systemd", "state": "active", "status": "enabled" }, "dbus-org.freedesktop.hostname1.service": { "name": "dbus-org.freedesktop.hostname1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.locale1.service": { "name": "dbus-org.freedesktop.locale1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.login1.service": { "name": "dbus-org.freedesktop.login1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.nm-dispatcher.service": { "name": "dbus-org.freedesktop.nm-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "dbus-org.freedesktop.portable1.service": { "name": "dbus-org.freedesktop.portable1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.timedate1.service": { "name": "dbus-org.freedesktop.timedate1.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "dbus.service": { "name": "dbus.service", "source": "systemd", "state": "running", "status": "static" }, "debug-shell.service": { "name": "debug-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dnf-makecache.service": { "name": "dnf-makecache.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-system-upgrade-cleanup.service": { "name": "dnf-system-upgrade-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "dnf-system-upgrade.service": { "name": "dnf-system-upgrade.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dnsmasq.service": { "name": "dnsmasq.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dracut-cmdline.service": { "name": "dracut-cmdline.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-initqueue.service": { "name": "dracut-initqueue.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-mount.service": { "name": "dracut-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-mount.service": { "name": "dracut-pre-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-pivot.service": { "name": "dracut-pre-pivot.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-trigger.service": { "name": "dracut-pre-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-udev.service": { "name": "dracut-pre-udev.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown-onfailure.service": { "name": "dracut-shutdown-onfailure.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown.service": { "name": "dracut-shutdown.service", "source": "systemd", "state": "stopped", "status": "static" }, "ebtables.service": { "name": "ebtables.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "emergency.service": { "name": "emergency.service", "source": "systemd", "state": "stopped", "status": "static" }, "firewalld.service": { "name": "firewalld.service", "source": "systemd", "state": "running", "status": "enabled" }, "fstrim.service": { "name": "fstrim.service", "source": "systemd", "state": "inactive", "status": "static" }, "getty@.service": { "name": "getty@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "getty@tty1.service": { "name": "getty@tty1.service", "source": "systemd", "state": "running", "status": "unknown" }, "grub-boot-indeterminate.service": { "name": "grub-boot-indeterminate.service", "source": "systemd", "state": "inactive", "status": "static" }, "gssproxy.service": { "name": "gssproxy.service", "source": "systemd", "state": "running", "status": "disabled" }, "halt-local.service": { "name": "halt-local.service", "source": "systemd", "state": "inactive", "status": "static" }, "import-state.service": { "name": "import-state.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "initrd-cleanup.service": { "name": "initrd-cleanup.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-parse-etc.service": { "name": "initrd-parse-etc.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-switch-root.service": { "name": "initrd-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-udevadm-cleanup-db.service": { "name": "initrd-udevadm-cleanup-db.service", "source": "systemd", "state": "stopped", "status": "static" }, "iprdump.service": { "name": "iprdump.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "iprinit.service": { "name": "iprinit.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "iprupdate.service": { "name": "iprupdate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "irqbalance.service": { "name": "irqbalance.service", "source": "systemd", "state": "running", "status": "enabled" }, "kdump.service": { "name": "kdump.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "kmod-static-nodes.service": { "name": "kmod-static-nodes.service", "source": "systemd", "state": "stopped", "status": "static" }, "kvm_stat.service": { "name": "kvm_stat.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "ldconfig.service": { "name": "ldconfig.service", "source": "systemd", "state": "stopped", "status": "static" }, "loadmodules.service": { "name": "loadmodules.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "man-db-cache-update.service": { "name": "man-db-cache-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "man-db-restart-cache-update.service": { "name": "man-db-restart-cache-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "messagebus.service": { "name": "messagebus.service", "source": "systemd", "state": "active", "status": "static" }, "microcode.service": { "name": "microcode.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "nfs-blkmap.service": { "name": "nfs-blkmap.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-convert.service": { "name": "nfs-convert.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-idmapd.service": { "name": "nfs-idmapd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-mountd.service": { "name": "nfs-mountd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-server.service": { "name": "nfs-server.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-utils.service": { "name": "nfs-utils.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfsdcld.service": { "name": "nfsdcld.service", "source": "systemd", "state": "stopped", "status": "static" }, "nftables.service": { "name": "nftables.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nis-domainname.service": { "name": "nis-domainname.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "oddjobd.service": { "name": "oddjobd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "plymouth-halt.service": { "name": "plymouth-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-kexec.service": { "name": "plymouth-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-poweroff.service": { "name": "plymouth-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-quit-wait.service": { "name": "plymouth-quit-wait.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-quit.service": { "name": "plymouth-quit.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-read-write.service": { "name": "plymouth-read-write.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-reboot.service": { "name": "plymouth-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-start.service": { "name": "plymouth-start.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-switch-root-initramfs.service": { "name": "plymouth-switch-root-initramfs.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-switch-root.service": { "name": "plymouth-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "podman-auto-update.service": { "name": "podman-auto-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "podman-clean-transient.service": { "name": "podman-clean-transient.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "podman-kube@.service": { "name": "podman-kube@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "podman-restart.service": { "name": "podman-restart.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "podman.service": { "name": "podman.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "polkit.service": { "name": "polkit.service", "source": "systemd", "state": "running", "status": "static" }, "qemu-guest-agent.service": { "name": "qemu-guest-agent.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "quadlet-demo-mysql-volume.service": { "name": "quadlet-demo-mysql-volume.service", "source": "systemd", "state": "stopped", "status": "generated" }, "quadlet-demo-network.service": { "name": "quadlet-demo-network.service", "source": "systemd", "state": "stopped", "status": "generated" }, "quotaon.service": { "name": "quotaon.service", "source": "systemd", "state": "inactive", "status": "static" }, "rc-local.service": { "name": "rc-local.service", "source": "systemd", "state": "stopped", "status": "static" }, "rdisc.service": { "name": "rdisc.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rescue.service": { "name": "rescue.service", "source": "systemd", "state": "stopped", "status": "static" }, "restraintd.service": { "name": "restraintd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rngd.service": { "name": "rngd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpc-gssd.service": { "name": "rpc-gssd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd-notify.service": { "name": "rpc-statd-notify.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd.service": { "name": "rpc-statd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpcbind.service": { "name": "rpcbind.service", "source": "systemd", "state": "running", "status": "enabled" }, "rsyslog.service": { "name": "rsyslog.service", "source": "systemd", "state": "running", "status": "enabled" }, "selinux-autorelabel-mark.service": { "name": "selinux-autorelabel-mark.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "selinux-autorelabel.service": { "name": "selinux-autorelabel.service", "source": "systemd", "state": "inactive", "status": "static" }, "serial-getty@.service": { "name": "serial-getty@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@.service": { "name": "sshd-keygen@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@ecdsa.service": { "name": "sshd-keygen@ecdsa.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd-keygen@ed25519.service": { "name": "sshd-keygen@ed25519.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd-keygen@rsa.service": { "name": "sshd-keygen@rsa.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd.service": { "name": "sshd.service", "source": "systemd", "state": "running", "status": "enabled" }, "sshd@.service": { "name": "sshd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "sssd-autofs.service": { "name": "sssd-autofs.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-kcm.service": { "name": "sssd-kcm.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "sssd-nss.service": { "name": "sssd-nss.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pac.service": { "name": "sssd-pac.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pam.service": { "name": "sssd-pam.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-ssh.service": { "name": "sssd-ssh.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-sudo.service": { "name": "sssd-sudo.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd.service": { "name": "sssd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "syslog.service": { "name": "syslog.service", "source": "systemd", "state": "active", "status": "enabled" }, "system-update-cleanup.service": { "name": "system-update-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-ask-password-console.service": { "name": "systemd-ask-password-console.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-plymouth.service": { "name": "systemd-ask-password-plymouth.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-wall.service": { "name": "systemd-ask-password-wall.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-backlight@.service": { "name": "systemd-backlight@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-binfmt.service": { "name": "systemd-binfmt.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-coredump@.service": { "name": "systemd-coredump@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-exit.service": { "name": "systemd-exit.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-firstboot.service": { "name": "systemd-firstboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck-root.service": { "name": "systemd-fsck-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck@.service": { "name": "systemd-fsck@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-halt.service": { "name": "systemd-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hibernate-resume@.service": { "name": "systemd-hibernate-resume@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-hibernate.service": { "name": "systemd-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hostnamed.service": { "name": "systemd-hostnamed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hwdb-update.service": { "name": "systemd-hwdb-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-hybrid-sleep.service": { "name": "systemd-hybrid-sleep.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-initctl.service": { "name": "systemd-initctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-catalog-update.service": { "name": "systemd-journal-catalog-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-flush.service": { "name": "systemd-journal-flush.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journald.service": { "name": "systemd-journald.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-kexec.service": { "name": "systemd-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-localed.service": { "name": "systemd-localed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-logind.service": { "name": "systemd-logind.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-machine-id-commit.service": { "name": "systemd-machine-id-commit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-modules-load.service": { "name": "systemd-modules-load.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-portabled.service": { "name": "systemd-portabled.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-poweroff.service": { "name": "systemd-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pstore.service": { "name": "systemd-pstore.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-quotacheck.service": { "name": "systemd-quotacheck.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-random-seed.service": { "name": "systemd-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-reboot.service": { "name": "systemd-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-remount-fs.service": { "name": "systemd-remount-fs.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-resolved.service": { "name": "systemd-resolved.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-rfkill.service": { "name": "systemd-rfkill.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend-then-hibernate.service": { "name": "systemd-suspend-then-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend.service": { "name": "systemd-suspend.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-sysctl.service": { "name": "systemd-sysctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-sysusers.service": { "name": "systemd-sysusers.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-timedated.service": { "name": "systemd-timedated.service", "source": "systemd", "state": "inactive", "status": "masked" }, "systemd-tmpfiles-clean.service": { "name": "systemd-tmpfiles-clean.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev.service": { "name": "systemd-tmpfiles-setup-dev.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup.service": { "name": "systemd-tmpfiles-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-settle.service": { "name": "systemd-udev-settle.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-udev-trigger.service": { "name": "systemd-udev-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udevd.service": { "name": "systemd-udevd.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-update-done.service": { "name": "systemd-update-done.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp-runlevel.service": { "name": "systemd-update-utmp-runlevel.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp.service": { "name": "systemd-update-utmp.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-user-sessions.service": { "name": "systemd-user-sessions.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-vconsole-setup.service": { "name": "systemd-vconsole-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-volatile-root.service": { "name": "systemd-volatile-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "tcsd.service": { "name": "tcsd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "teamd@.service": { "name": "teamd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "timedatex.service": { "name": "timedatex.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "tuned.service": { "name": "tuned.service", "source": "systemd", "state": "running", "status": "enabled" }, "unbound-anchor.service": { "name": "unbound-anchor.service", "source": "systemd", "state": "stopped", "status": "static" }, "user-runtime-dir@.service": { "name": "user-runtime-dir@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user-runtime-dir@0.service": { "name": "user-runtime-dir@0.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "user@.service": { "name": "user@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user@0.service": { "name": "user@0.service", "source": "systemd", "state": "running", "status": "unknown" } } }, "changed": false } TASK [fedora.linux_system_roles.podman : Create and update quadlets] *********** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:114 Saturday 24 August 2024 12:36:22 -0400 (0:00:01.575) 0:02:31.433 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14 Saturday 24 August 2024 12:36:22 -0400 (0:00:00.067) 0:02:31.500 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_file_src": "quadlet-demo-mysql.volume", "__podman_quadlet_spec": {}, "__podman_quadlet_str": "[Volume]", "__podman_quadlet_template_src": "" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25 Saturday 24 August 2024 12:36:22 -0400 (0:00:00.045) 0:02:31.545 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_pull_image": true, "__podman_state": "absent", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] ***** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35 Saturday 24 August 2024 12:36:22 -0400 (0:00:00.040) 0:02:31.586 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48 Saturday 24 August 2024 12:36:23 -0400 (0:00:00.035) 0:02:31.621 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_name": "quadlet-demo-mysql", "__podman_quadlet_type": "volume", "__podman_rootless": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57 Saturday 24 August 2024 12:36:23 -0400 (0:00:00.046) 0:02:31.667 ******* included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Saturday 24 August 2024 12:36:23 -0400 (0:00:00.061) 0:02:31.728 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Saturday 24 August 2024 12:36:23 -0400 (0:00:00.041) 0:02:31.770 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Saturday 24 August 2024 12:36:23 -0400 (0:00:00.040) 0:02:31.810 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get group information] **************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:28 Saturday 24 August 2024 12:36:23 -0400 (0:00:00.047) 0:02:31.857 ******* ok: [managed_node1] => { "ansible_facts": { "getent_group": { "root": [ "x", "0", "" ] } }, "changed": false } TASK [fedora.linux_system_roles.podman : Set group name] *********************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:35 Saturday 24 August 2024 12:36:23 -0400 (0:00:00.366) 0:02:32.223 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_group_name": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:39 Saturday 24 August 2024 12:36:23 -0400 (0:00:00.046) 0:02:32.270 ******* ok: [managed_node1] => { "changed": false, "stat": { "atime": 1724516929.7168462, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "bb5b46ffbafcaa8c4021f3c8b3cb8594f48ef34b", "ctime": 1724516896.725572, "dev": 51713, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 6986657, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-sharedlib", "mode": "0755", "mtime": 1700557386.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 12640, "uid": 0, "version": "1255679238", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************ task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:50 Saturday 24 August 2024 12:36:24 -0400 (0:00:00.363) 0:02:32.634 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check group with getsubids] *********** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:55 Saturday 24 August 2024 12:36:24 -0400 (0:00:00.081) 0:02:32.715 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:60 Saturday 24 August 2024 12:36:24 -0400 (0:00:00.034) 0:02:32.750 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subuid file] ********************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:74 Saturday 24 August 2024 12:36:24 -0400 (0:00:00.035) 0:02:32.786 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subgid file] ********************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:79 Saturday 24 August 2024 12:36:24 -0400 (0:00:00.035) 0:02:32.822 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:84 Saturday 24 August 2024 12:36:24 -0400 (0:00:00.034) 0:02:32.856 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:94 Saturday 24 August 2024 12:36:24 -0400 (0:00:00.035) 0:02:32.892 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] ***** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:101 Saturday 24 August 2024 12:36:24 -0400 (0:00:00.035) 0:02:32.928 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62 Saturday 24 August 2024 12:36:24 -0400 (0:00:00.034) 0:02:32.962 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_activate_systemd_unit": true, "__podman_images_found": [], "__podman_kube_yamls_raw": "", "__podman_service_name": "quadlet-demo-mysql-volume.service", "__podman_systemd_scope": "system", "__podman_user_home_dir": "/root", "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73 Saturday 24 August 2024 12:36:24 -0400 (0:00:00.062) 0:02:33.025 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_path": "/etc/containers/systemd" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get kube yaml contents] *************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77 Saturday 24 August 2024 12:36:24 -0400 (0:00:00.037) 0:02:33.062 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87 Saturday 24 August 2024 12:36:24 -0400 (0:00:00.033) 0:02:33.095 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_images": [], "__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo-mysql.volume", "__podman_volumes": [] }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 6] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:103 Saturday 24 August 2024 12:36:24 -0400 (0:00:00.076) 0:02:33.172 ******* ok: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Cleanup quadlets] ********************* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:110 Saturday 24 August 2024 12:36:24 -0400 (0:00:00.039) 0:02:33.211 ******* included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Stat XDG_RUNTIME_DIR] ***************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:4 Saturday 24 August 2024 12:36:24 -0400 (0:00:00.081) 0:02:33.293 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Stop and disable service] ************* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:12 Saturday 24 August 2024 12:36:24 -0400 (0:00:00.034) 0:02:33.327 ******* changed: [managed_node1] => { "changed": true, "enabled": false, "failed_when_result": false, "name": "quadlet-demo-mysql-volume.service", "state": "stopped", "status": { "ActiveEnterTimestamp": "Sat 2024-08-24 12:34:35 EDT", "ActiveEnterTimestampMonotonic": "678725800", "ActiveExitTimestampMonotonic": "0", "ActiveState": "active", "After": "basic.target -.mount sysinit.target systemd-journald.socket system.slice", "AllowIsolate": "no", "AllowedCPUs": "", "AllowedMemoryNodes": "", "AmbientCapabilities": "", "AssertResult": "yes", "AssertTimestamp": "Sat 2024-08-24 12:34:35 EDT", "AssertTimestampMonotonic": "678677120", "Before": "shutdown.target", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "CPUAccounting": "no", "CPUAffinity": "", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "[not set]", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "no", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf", "CollectMode": "inactive", "ConditionResult": "yes", "ConditionTimestamp": "Sat 2024-08-24 12:34:35 EDT", "ConditionTimestampMonotonic": "678677118", "ConfigurationDirectoryMode": "0755", "Conflicts": "shutdown.target", "ControlGroup": "/system.slice/quadlet-demo-mysql-volume.service", "ControlPID": "0", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "Delegate": "no", "Description": "quadlet-demo-mysql-volume.service", "DevicePolicy": "auto", "DynamicUser": "no", "EffectiveCPUs": "", "EffectiveMemoryNodes": "", "ExecMainCode": "1", "ExecMainExitTimestamp": "Sat 2024-08-24 12:34:35 EDT", "ExecMainExitTimestampMonotonic": "678725428", "ExecMainPID": "82689", "ExecMainStartTimestamp": "Sat 2024-08-24 12:34:35 EDT", "ExecMainStartTimestampMonotonic": "678678139", "ExecMainStatus": "0", "ExecStart": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman volume create --ignore systemd-quadlet-demo-mysql ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FragmentPath": "/run/systemd/generator/quadlet-demo-mysql-volume.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOSchedulingClass": "0", "IOSchedulingPriority": "0", "IOWeight": "[not set]", "IPAccounting": "no", "IPEgressBytes": "18446744073709551615", "IPEgressPackets": "18446744073709551615", "IPIngressBytes": "18446744073709551615", "IPIngressPackets": "18446744073709551615", "Id": "quadlet-demo-mysql-volume.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestamp": "Sat 2024-08-24 12:34:35 EDT", "InactiveExitTimestampMonotonic": "678678193", "InvocationID": "4947f49535a04e7d84bd4de87f0ddea2", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "control-group", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "0", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "65536", "LimitMEMLOCKSoft": "65536", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "262144", "LimitNOFILESoft": "1024", "LimitNPROC": "14003", "LimitNPROCSoft": "14003", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "14003", "LimitSIGPENDINGSoft": "14003", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "0", "MemoryAccounting": "yes", "MemoryCurrent": "0", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemorySwapMax": "infinity", "MountAPIVFS": "no", "MountFlags": "", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAMask": "", "NUMAPolicy": "n/a", "Names": "quadlet-demo-mysql-volume.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "PermissionsStartOnly": "no", "Perpetual": "no", "PrivateDevices": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "RemainAfterExit": "yes", "RemoveIPC": "no", "Requires": "-.mount sysinit.target system.slice", "RequiresMountsFor": "/run/containers", "Restart": "no", "RestartUSec": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "Slice": "system.slice", "StandardError": "inherit", "StandardInput": "null", "StandardInputData": "", "StandardOutput": "journal", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StateChangeTimestamp": "Sat 2024-08-24 12:34:35 EDT", "StateChangeTimestampMonotonic": "678725800", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "exited", "SuccessAction": "none", "SyslogFacility": "3", "SyslogIdentifier": "quadlet-demo-mysql-volume", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "0", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "0", "TasksMax": "22405", "TimeoutStartUSec": "infinity", "TimeoutStopUSec": "1min 30s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "oneshot", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "disabled", "UnitFileState": "generated", "UtmpMode": "init", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "0" } } TASK [fedora.linux_system_roles.podman : See if quadlet file exists] *********** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:33 Saturday 24 August 2024 12:36:25 -0400 (0:00:00.643) 0:02:33.970 ******* ok: [managed_node1] => { "changed": false, "stat": { "atime": 1724517275.3136742, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 8, "charset": "us-ascii", "checksum": "585f8cbdf0ec73000f9227dcffbef71e9552ea4a", "ctime": 1724517274.5916681, "dev": 51713, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 213909769, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "text/plain", "mode": "0644", "mtime": 1724517274.2876654, "nlink": 1, "path": "/etc/containers/systemd/quadlet-demo-mysql.volume", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 9, "uid": 0, "version": "147345147", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [fedora.linux_system_roles.podman : Parse quadlet file] ******************* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:38 Saturday 24 August 2024 12:36:25 -0400 (0:00:00.374) 0:02:34.344 ******* included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Slurp quadlet file] ******************* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml:6 Saturday 24 August 2024 12:36:25 -0400 (0:00:00.061) 0:02:34.406 ******* ok: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Parse quadlet file] ******************* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml:12 Saturday 24 August 2024 12:36:26 -0400 (0:00:00.351) 0:02:34.758 ******* ok: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Parse quadlet yaml file] ************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml:44 Saturday 24 August 2024 12:36:26 -0400 (0:00:00.050) 0:02:34.808 ******* skipping: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Reset raw variable] ******************* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml:52 Saturday 24 August 2024 12:36:26 -0400 (0:00:00.033) 0:02:34.841 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_raw": null }, "changed": false } TASK [fedora.linux_system_roles.podman : Remove quadlet file] ****************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:42 Saturday 24 August 2024 12:36:26 -0400 (0:00:00.033) 0:02:34.875 ******* changed: [managed_node1] => { "changed": true, "path": "/etc/containers/systemd/quadlet-demo-mysql.volume", "state": "absent" } TASK [fedora.linux_system_roles.podman : Refresh systemd] ********************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:48 Saturday 24 August 2024 12:36:26 -0400 (0:00:00.363) 0:02:35.238 ******* ok: [managed_node1] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.podman : Remove managed resource] ************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:58 Saturday 24 August 2024 12:36:27 -0400 (0:00:00.592) 0:02:35.831 ******* changed: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": true } TASK [fedora.linux_system_roles.podman : Remove volumes] *********************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:95 Saturday 24 August 2024 12:36:27 -0400 (0:00:00.430) 0:02:36.261 ******* skipping: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Clear parsed podman variable] ********* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:112 Saturday 24 August 2024 12:36:27 -0400 (0:00:00.050) 0:02:36.312 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_parsed": null }, "changed": false } TASK [fedora.linux_system_roles.podman : Prune images no longer in use] ******** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:116 Saturday 24 August 2024 12:36:27 -0400 (0:00:00.033) 0:02:36.345 ******* changed: [managed_node1] => { "changed": true, "cmd": [ "podman", "image", "prune", "--all", "-f" ], "delta": "0:00:00.033667", "end": "2024-08-24 12:36:28.057630", "rc": 0, "start": "2024-08-24 12:36:28.023963" } TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:127 Saturday 24 August 2024 12:36:28 -0400 (0:00:00.394) 0:02:36.739 ******* included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Saturday 24 August 2024 12:36:28 -0400 (0:00:00.059) 0:02:36.799 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Saturday 24 August 2024 12:36:28 -0400 (0:00:00.032) 0:02:36.831 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Saturday 24 August 2024 12:36:28 -0400 (0:00:00.033) 0:02:36.864 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : For testing and debugging - images] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:137 Saturday 24 August 2024 12:36:28 -0400 (0:00:00.032) 0:02:36.897 ******* ok: [managed_node1] => { "changed": false, "cmd": [ "podman", "images", "-n" ], "delta": "0:00:00.033438", "end": "2024-08-24 12:36:28.608598", "rc": 0, "start": "2024-08-24 12:36:28.575160" } TASK [fedora.linux_system_roles.podman : For testing and debugging - volumes] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:146 Saturday 24 August 2024 12:36:28 -0400 (0:00:00.394) 0:02:37.292 ******* ok: [managed_node1] => { "changed": false, "cmd": [ "podman", "volume", "ls", "-n" ], "delta": "0:00:00.035375", "end": "2024-08-24 12:36:29.020418", "rc": 0, "start": "2024-08-24 12:36:28.985043" } STDOUT: local wp-pv-claim local envoy-proxy-config local envoy-certificates TASK [fedora.linux_system_roles.podman : For testing and debugging - containers] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:155 Saturday 24 August 2024 12:36:29 -0400 (0:00:00.412) 0:02:37.704 ******* ok: [managed_node1] => { "changed": false, "cmd": [ "podman", "ps", "--noheading" ], "delta": "0:00:00.031622", "end": "2024-08-24 12:36:29.419551", "rc": 0, "start": "2024-08-24 12:36:29.387929" } TASK [fedora.linux_system_roles.podman : For testing and debugging - networks] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:164 Saturday 24 August 2024 12:36:29 -0400 (0:00:00.398) 0:02:38.103 ******* ok: [managed_node1] => { "changed": false, "cmd": [ "podman", "network", "ls", "-n", "-q" ], "delta": "0:00:00.058888", "end": "2024-08-24 12:36:29.844427", "rc": 0, "start": "2024-08-24 12:36:29.785539" } STDOUT: podman podman-default-kube-network systemd-quadlet-demo TASK [fedora.linux_system_roles.podman : For testing and debugging - secrets] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:173 Saturday 24 August 2024 12:36:29 -0400 (0:00:00.424) 0:02:38.527 ******* ok: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : For testing and debugging - services] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:183 Saturday 24 August 2024 12:36:30 -0400 (0:00:00.397) 0:02:38.925 ******* ok: [managed_node1] => { "ansible_facts": { "services": { "NetworkManager-dispatcher.service": { "name": "NetworkManager-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "NetworkManager-wait-online.service": { "name": "NetworkManager-wait-online.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "NetworkManager.service": { "name": "NetworkManager.service", "source": "systemd", "state": "running", "status": "enabled" }, "auditd.service": { "name": "auditd.service", "source": "systemd", "state": "running", "status": "enabled" }, "auth-rpcgss-module.service": { "name": "auth-rpcgss-module.service", "source": "systemd", "state": "stopped", "status": "static" }, "autovt@.service": { "name": "autovt@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "certmonger.service": { "name": "certmonger.service", "source": "systemd", "state": "running", "status": "enabled" }, "chrony-dnssrv@.service": { "name": "chrony-dnssrv@.service", "source": "systemd", "state": "unknown", "status": "static" }, "chrony-wait.service": { "name": "chrony-wait.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd.service": { "name": "chronyd.service", "source": "systemd", "state": "running", "status": "enabled" }, "cloud-config.service": { "name": "cloud-config.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-final.service": { "name": "cloud-final.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init-hotplugd.service": { "name": "cloud-init-hotplugd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cloud-init-local.service": { "name": "cloud-init-local.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init.service": { "name": "cloud-init.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cni-dhcp.service": { "name": "cni-dhcp.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "console-getty.service": { "name": "console-getty.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "container-getty@.service": { "name": "container-getty@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cpupower.service": { "name": "cpupower.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "crond.service": { "name": "crond.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-org.fedoraproject.FirewallD1.service": { "name": "dbus-org.fedoraproject.FirewallD1.service", "source": "systemd", "state": "active", "status": "enabled" }, "dbus-org.freedesktop.hostname1.service": { "name": "dbus-org.freedesktop.hostname1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.locale1.service": { "name": "dbus-org.freedesktop.locale1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.login1.service": { "name": "dbus-org.freedesktop.login1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.nm-dispatcher.service": { "name": "dbus-org.freedesktop.nm-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "dbus-org.freedesktop.portable1.service": { "name": "dbus-org.freedesktop.portable1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.timedate1.service": { "name": "dbus-org.freedesktop.timedate1.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "dbus.service": { "name": "dbus.service", "source": "systemd", "state": "running", "status": "static" }, "debug-shell.service": { "name": "debug-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dnf-makecache.service": { "name": "dnf-makecache.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-system-upgrade-cleanup.service": { "name": "dnf-system-upgrade-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "dnf-system-upgrade.service": { "name": "dnf-system-upgrade.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dnsmasq.service": { "name": "dnsmasq.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dracut-cmdline.service": { "name": "dracut-cmdline.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-initqueue.service": { "name": "dracut-initqueue.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-mount.service": { "name": "dracut-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-mount.service": { "name": "dracut-pre-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-pivot.service": { "name": "dracut-pre-pivot.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-trigger.service": { "name": "dracut-pre-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-udev.service": { "name": "dracut-pre-udev.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown-onfailure.service": { "name": "dracut-shutdown-onfailure.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown.service": { "name": "dracut-shutdown.service", "source": "systemd", "state": "stopped", "status": "static" }, "ebtables.service": { "name": "ebtables.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "emergency.service": { "name": "emergency.service", "source": "systemd", "state": "stopped", "status": "static" }, "firewalld.service": { "name": "firewalld.service", "source": "systemd", "state": "running", "status": "enabled" }, "fstrim.service": { "name": "fstrim.service", "source": "systemd", "state": "inactive", "status": "static" }, "getty@.service": { "name": "getty@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "getty@tty1.service": { "name": "getty@tty1.service", "source": "systemd", "state": "running", "status": "unknown" }, "grub-boot-indeterminate.service": { "name": "grub-boot-indeterminate.service", "source": "systemd", "state": "inactive", "status": "static" }, "gssproxy.service": { "name": "gssproxy.service", "source": "systemd", "state": "running", "status": "disabled" }, "halt-local.service": { "name": "halt-local.service", "source": "systemd", "state": "inactive", "status": "static" }, "import-state.service": { "name": "import-state.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "initrd-cleanup.service": { "name": "initrd-cleanup.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-parse-etc.service": { "name": "initrd-parse-etc.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-switch-root.service": { "name": "initrd-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-udevadm-cleanup-db.service": { "name": "initrd-udevadm-cleanup-db.service", "source": "systemd", "state": "stopped", "status": "static" }, "iprdump.service": { "name": "iprdump.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "iprinit.service": { "name": "iprinit.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "iprupdate.service": { "name": "iprupdate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "irqbalance.service": { "name": "irqbalance.service", "source": "systemd", "state": "running", "status": "enabled" }, "kdump.service": { "name": "kdump.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "kmod-static-nodes.service": { "name": "kmod-static-nodes.service", "source": "systemd", "state": "stopped", "status": "static" }, "kvm_stat.service": { "name": "kvm_stat.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "ldconfig.service": { "name": "ldconfig.service", "source": "systemd", "state": "stopped", "status": "static" }, "loadmodules.service": { "name": "loadmodules.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "man-db-cache-update.service": { "name": "man-db-cache-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "man-db-restart-cache-update.service": { "name": "man-db-restart-cache-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "messagebus.service": { "name": "messagebus.service", "source": "systemd", "state": "active", "status": "static" }, "microcode.service": { "name": "microcode.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "nfs-blkmap.service": { "name": "nfs-blkmap.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-convert.service": { "name": "nfs-convert.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-idmapd.service": { "name": "nfs-idmapd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-mountd.service": { "name": "nfs-mountd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-server.service": { "name": "nfs-server.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-utils.service": { "name": "nfs-utils.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfsdcld.service": { "name": "nfsdcld.service", "source": "systemd", "state": "stopped", "status": "static" }, "nftables.service": { "name": "nftables.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nis-domainname.service": { "name": "nis-domainname.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "oddjobd.service": { "name": "oddjobd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "plymouth-halt.service": { "name": "plymouth-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-kexec.service": { "name": "plymouth-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-poweroff.service": { "name": "plymouth-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-quit-wait.service": { "name": "plymouth-quit-wait.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-quit.service": { "name": "plymouth-quit.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-read-write.service": { "name": "plymouth-read-write.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-reboot.service": { "name": "plymouth-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-start.service": { "name": "plymouth-start.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-switch-root-initramfs.service": { "name": "plymouth-switch-root-initramfs.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-switch-root.service": { "name": "plymouth-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "podman-auto-update.service": { "name": "podman-auto-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "podman-clean-transient.service": { "name": "podman-clean-transient.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "podman-kube@.service": { "name": "podman-kube@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "podman-restart.service": { "name": "podman-restart.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "podman.service": { "name": "podman.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "polkit.service": { "name": "polkit.service", "source": "systemd", "state": "running", "status": "static" }, "qemu-guest-agent.service": { "name": "qemu-guest-agent.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "quadlet-demo-network.service": { "name": "quadlet-demo-network.service", "source": "systemd", "state": "stopped", "status": "generated" }, "quotaon.service": { "name": "quotaon.service", "source": "systemd", "state": "inactive", "status": "static" }, "rc-local.service": { "name": "rc-local.service", "source": "systemd", "state": "stopped", "status": "static" }, "rdisc.service": { "name": "rdisc.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rescue.service": { "name": "rescue.service", "source": "systemd", "state": "stopped", "status": "static" }, "restraintd.service": { "name": "restraintd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rngd.service": { "name": "rngd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpc-gssd.service": { "name": "rpc-gssd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd-notify.service": { "name": "rpc-statd-notify.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd.service": { "name": "rpc-statd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpcbind.service": { "name": "rpcbind.service", "source": "systemd", "state": "running", "status": "enabled" }, "rsyslog.service": { "name": "rsyslog.service", "source": "systemd", "state": "running", "status": "enabled" }, "selinux-autorelabel-mark.service": { "name": "selinux-autorelabel-mark.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "selinux-autorelabel.service": { "name": "selinux-autorelabel.service", "source": "systemd", "state": "inactive", "status": "static" }, "serial-getty@.service": { "name": "serial-getty@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@.service": { "name": "sshd-keygen@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@ecdsa.service": { "name": "sshd-keygen@ecdsa.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd-keygen@ed25519.service": { "name": "sshd-keygen@ed25519.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd-keygen@rsa.service": { "name": "sshd-keygen@rsa.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd.service": { "name": "sshd.service", "source": "systemd", "state": "running", "status": "enabled" }, "sshd@.service": { "name": "sshd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "sssd-autofs.service": { "name": "sssd-autofs.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-kcm.service": { "name": "sssd-kcm.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "sssd-nss.service": { "name": "sssd-nss.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pac.service": { "name": "sssd-pac.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pam.service": { "name": "sssd-pam.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-ssh.service": { "name": "sssd-ssh.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-sudo.service": { "name": "sssd-sudo.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd.service": { "name": "sssd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "syslog.service": { "name": "syslog.service", "source": "systemd", "state": "active", "status": "enabled" }, "system-update-cleanup.service": { "name": "system-update-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-ask-password-console.service": { "name": "systemd-ask-password-console.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-plymouth.service": { "name": "systemd-ask-password-plymouth.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-wall.service": { "name": "systemd-ask-password-wall.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-backlight@.service": { "name": "systemd-backlight@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-binfmt.service": { "name": "systemd-binfmt.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-coredump@.service": { "name": "systemd-coredump@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-exit.service": { "name": "systemd-exit.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-firstboot.service": { "name": "systemd-firstboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck-root.service": { "name": "systemd-fsck-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck@.service": { "name": "systemd-fsck@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-halt.service": { "name": "systemd-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hibernate-resume@.service": { "name": "systemd-hibernate-resume@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-hibernate.service": { "name": "systemd-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hostnamed.service": { "name": "systemd-hostnamed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hwdb-update.service": { "name": "systemd-hwdb-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-hybrid-sleep.service": { "name": "systemd-hybrid-sleep.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-initctl.service": { "name": "systemd-initctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-catalog-update.service": { "name": "systemd-journal-catalog-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-flush.service": { "name": "systemd-journal-flush.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journald.service": { "name": "systemd-journald.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-kexec.service": { "name": "systemd-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-localed.service": { "name": "systemd-localed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-logind.service": { "name": "systemd-logind.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-machine-id-commit.service": { "name": "systemd-machine-id-commit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-modules-load.service": { "name": "systemd-modules-load.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-portabled.service": { "name": "systemd-portabled.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-poweroff.service": { "name": "systemd-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pstore.service": { "name": "systemd-pstore.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-quotacheck.service": { "name": "systemd-quotacheck.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-random-seed.service": { "name": "systemd-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-reboot.service": { "name": "systemd-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-remount-fs.service": { "name": "systemd-remount-fs.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-resolved.service": { "name": "systemd-resolved.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-rfkill.service": { "name": "systemd-rfkill.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend-then-hibernate.service": { "name": "systemd-suspend-then-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend.service": { "name": "systemd-suspend.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-sysctl.service": { "name": "systemd-sysctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-sysusers.service": { "name": "systemd-sysusers.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-timedated.service": { "name": "systemd-timedated.service", "source": "systemd", "state": "inactive", "status": "masked" }, "systemd-tmpfiles-clean.service": { "name": "systemd-tmpfiles-clean.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev.service": { "name": "systemd-tmpfiles-setup-dev.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup.service": { "name": "systemd-tmpfiles-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-settle.service": { "name": "systemd-udev-settle.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-udev-trigger.service": { "name": "systemd-udev-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udevd.service": { "name": "systemd-udevd.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-update-done.service": { "name": "systemd-update-done.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp-runlevel.service": { "name": "systemd-update-utmp-runlevel.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp.service": { "name": "systemd-update-utmp.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-user-sessions.service": { "name": "systemd-user-sessions.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-vconsole-setup.service": { "name": "systemd-vconsole-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-volatile-root.service": { "name": "systemd-volatile-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "tcsd.service": { "name": "tcsd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "teamd@.service": { "name": "teamd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "timedatex.service": { "name": "timedatex.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "tuned.service": { "name": "tuned.service", "source": "systemd", "state": "running", "status": "enabled" }, "unbound-anchor.service": { "name": "unbound-anchor.service", "source": "systemd", "state": "stopped", "status": "static" }, "user-runtime-dir@.service": { "name": "user-runtime-dir@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user-runtime-dir@0.service": { "name": "user-runtime-dir@0.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "user@.service": { "name": "user@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user@0.service": { "name": "user@0.service", "source": "systemd", "state": "running", "status": "unknown" } } }, "changed": false } TASK [fedora.linux_system_roles.podman : Create and update quadlets] *********** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:114 Saturday 24 August 2024 12:36:31 -0400 (0:00:01.613) 0:02:40.538 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14 Saturday 24 August 2024 12:36:31 -0400 (0:00:00.065) 0:02:40.603 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_file_src": "quadlet-demo.network", "__podman_quadlet_spec": {}, "__podman_quadlet_str": "[Network]\nSubnet=192.168.30.0/24\nGateway=192.168.30.1\nLabel=app=wordpress", "__podman_quadlet_template_src": "" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25 Saturday 24 August 2024 12:36:32 -0400 (0:00:00.045) 0:02:40.649 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_pull_image": true, "__podman_state": "absent", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] ***** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35 Saturday 24 August 2024 12:36:32 -0400 (0:00:00.039) 0:02:40.688 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48 Saturday 24 August 2024 12:36:32 -0400 (0:00:00.034) 0:02:40.723 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_name": "quadlet-demo", "__podman_quadlet_type": "network", "__podman_rootless": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57 Saturday 24 August 2024 12:36:32 -0400 (0:00:00.046) 0:02:40.769 ******* included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Saturday 24 August 2024 12:36:32 -0400 (0:00:00.062) 0:02:40.832 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Saturday 24 August 2024 12:36:32 -0400 (0:00:00.041) 0:02:40.874 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Saturday 24 August 2024 12:36:32 -0400 (0:00:00.040) 0:02:40.914 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get group information] **************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:28 Saturday 24 August 2024 12:36:32 -0400 (0:00:00.043) 0:02:40.958 ******* ok: [managed_node1] => { "ansible_facts": { "getent_group": { "root": [ "x", "0", "" ] } }, "changed": false } TASK [fedora.linux_system_roles.podman : Set group name] *********************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:35 Saturday 24 August 2024 12:36:32 -0400 (0:00:00.370) 0:02:41.328 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_group_name": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:39 Saturday 24 August 2024 12:36:32 -0400 (0:00:00.044) 0:02:41.373 ******* ok: [managed_node1] => { "changed": false, "stat": { "atime": 1724516929.7168462, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "bb5b46ffbafcaa8c4021f3c8b3cb8594f48ef34b", "ctime": 1724516896.725572, "dev": 51713, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 6986657, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-sharedlib", "mode": "0755", "mtime": 1700557386.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 12640, "uid": 0, "version": "1255679238", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************ task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:50 Saturday 24 August 2024 12:36:33 -0400 (0:00:00.358) 0:02:41.731 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check group with getsubids] *********** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:55 Saturday 24 August 2024 12:36:33 -0400 (0:00:00.074) 0:02:41.806 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:60 Saturday 24 August 2024 12:36:33 -0400 (0:00:00.033) 0:02:41.839 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subuid file] ********************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:74 Saturday 24 August 2024 12:36:33 -0400 (0:00:00.034) 0:02:41.873 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subgid file] ********************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:79 Saturday 24 August 2024 12:36:33 -0400 (0:00:00.033) 0:02:41.907 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:84 Saturday 24 August 2024 12:36:33 -0400 (0:00:00.033) 0:02:41.940 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:94 Saturday 24 August 2024 12:36:33 -0400 (0:00:00.033) 0:02:41.974 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] ***** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:101 Saturday 24 August 2024 12:36:33 -0400 (0:00:00.032) 0:02:42.007 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62 Saturday 24 August 2024 12:36:33 -0400 (0:00:00.031) 0:02:42.039 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_activate_systemd_unit": true, "__podman_images_found": [], "__podman_kube_yamls_raw": "", "__podman_service_name": "quadlet-demo-network.service", "__podman_systemd_scope": "system", "__podman_user_home_dir": "/root", "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73 Saturday 24 August 2024 12:36:33 -0400 (0:00:00.057) 0:02:42.096 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_path": "/etc/containers/systemd" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get kube yaml contents] *************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77 Saturday 24 August 2024 12:36:33 -0400 (0:00:00.034) 0:02:42.130 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87 Saturday 24 August 2024 12:36:33 -0400 (0:00:00.033) 0:02:42.163 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_images": [], "__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo.network", "__podman_volumes": [] }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 6] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:103 Saturday 24 August 2024 12:36:33 -0400 (0:00:00.070) 0:02:42.234 ******* ok: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Cleanup quadlets] ********************* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:110 Saturday 24 August 2024 12:36:33 -0400 (0:00:00.038) 0:02:42.273 ******* included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Stat XDG_RUNTIME_DIR] ***************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:4 Saturday 24 August 2024 12:36:33 -0400 (0:00:00.080) 0:02:42.353 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Stop and disable service] ************* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:12 Saturday 24 August 2024 12:36:33 -0400 (0:00:00.032) 0:02:42.386 ******* changed: [managed_node1] => { "changed": true, "enabled": false, "failed_when_result": false, "name": "quadlet-demo-network.service", "state": "stopped", "status": { "ActiveEnterTimestamp": "Sat 2024-08-24 12:34:30 EDT", "ActiveEnterTimestampMonotonic": "673375366", "ActiveExitTimestampMonotonic": "0", "ActiveState": "active", "After": "sysinit.target systemd-journald.socket system.slice basic.target -.mount", "AllowIsolate": "no", "AllowedCPUs": "", "AllowedMemoryNodes": "", "AmbientCapabilities": "", "AssertResult": "yes", "AssertTimestamp": "Sat 2024-08-24 12:34:30 EDT", "AssertTimestampMonotonic": "673320204", "Before": "shutdown.target", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "CPUAccounting": "no", "CPUAffinity": "", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "[not set]", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "no", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf", "CollectMode": "inactive", "ConditionResult": "yes", "ConditionTimestamp": "Sat 2024-08-24 12:34:30 EDT", "ConditionTimestampMonotonic": "673320203", "ConfigurationDirectoryMode": "0755", "Conflicts": "shutdown.target", "ControlGroup": "/system.slice/quadlet-demo-network.service", "ControlPID": "0", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "Delegate": "no", "Description": "quadlet-demo-network.service", "DevicePolicy": "auto", "DynamicUser": "no", "EffectiveCPUs": "", "EffectiveMemoryNodes": "", "ExecMainCode": "1", "ExecMainExitTimestamp": "Sat 2024-08-24 12:34:30 EDT", "ExecMainExitTimestampMonotonic": "673374927", "ExecMainPID": "81779", "ExecMainStartTimestamp": "Sat 2024-08-24 12:34:30 EDT", "ExecMainStartTimestampMonotonic": "673321150", "ExecMainStatus": "0", "ExecStart": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman network create --ignore --subnet=192.168.30.0/24 --gateway=192.168.30.1 --label app=wordpress systemd-quadlet-demo ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FragmentPath": "/run/systemd/generator/quadlet-demo-network.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOSchedulingClass": "0", "IOSchedulingPriority": "0", "IOWeight": "[not set]", "IPAccounting": "no", "IPEgressBytes": "18446744073709551615", "IPEgressPackets": "18446744073709551615", "IPIngressBytes": "18446744073709551615", "IPIngressPackets": "18446744073709551615", "Id": "quadlet-demo-network.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestamp": "Sat 2024-08-24 12:34:30 EDT", "InactiveExitTimestampMonotonic": "673321190", "InvocationID": "124aa2ffb79e4600a823e33d270d8a42", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "control-group", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "0", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "65536", "LimitMEMLOCKSoft": "65536", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "262144", "LimitNOFILESoft": "1024", "LimitNPROC": "14003", "LimitNPROCSoft": "14003", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "14003", "LimitSIGPENDINGSoft": "14003", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "0", "MemoryAccounting": "yes", "MemoryCurrent": "0", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemorySwapMax": "infinity", "MountAPIVFS": "no", "MountFlags": "", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAMask": "", "NUMAPolicy": "n/a", "Names": "quadlet-demo-network.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "PermissionsStartOnly": "no", "Perpetual": "no", "PrivateDevices": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "RemainAfterExit": "yes", "RemoveIPC": "no", "Requires": "sysinit.target system.slice -.mount", "RequiresMountsFor": "/run/containers", "Restart": "no", "RestartUSec": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "Slice": "system.slice", "StandardError": "inherit", "StandardInput": "null", "StandardInputData": "", "StandardOutput": "journal", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StateChangeTimestamp": "Sat 2024-08-24 12:34:30 EDT", "StateChangeTimestampMonotonic": "673375366", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "exited", "SuccessAction": "none", "SyslogFacility": "3", "SyslogIdentifier": "quadlet-demo-network", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "0", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "0", "TasksMax": "22405", "TimeoutStartUSec": "infinity", "TimeoutStopUSec": "1min 30s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "oneshot", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "disabled", "UnitFileState": "generated", "UtmpMode": "init", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "0" } } TASK [fedora.linux_system_roles.podman : See if quadlet file exists] *********** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:33 Saturday 24 August 2024 12:36:34 -0400 (0:00:00.629) 0:02:43.015 ******* ok: [managed_node1] => { "changed": false, "stat": { "atime": 1724517269.990629, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 8, "charset": "us-ascii", "checksum": "e57c08d49aff4bae8daab138d913aeddaa8682a0", "ctime": 1724517269.3686237, "dev": 51713, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 186646789, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "text/plain", "mode": "0644", "mtime": 1724517269.029621, "nlink": 1, "path": "/etc/containers/systemd/quadlet-demo.network", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 74, "uid": 0, "version": "2153314366", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [fedora.linux_system_roles.podman : Parse quadlet file] ******************* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:38 Saturday 24 August 2024 12:36:34 -0400 (0:00:00.369) 0:02:43.385 ******* included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Slurp quadlet file] ******************* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml:6 Saturday 24 August 2024 12:36:34 -0400 (0:00:00.062) 0:02:43.447 ******* ok: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Parse quadlet file] ******************* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml:12 Saturday 24 August 2024 12:36:35 -0400 (0:00:00.353) 0:02:43.801 ******* ok: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Parse quadlet yaml file] ************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml:44 Saturday 24 August 2024 12:36:35 -0400 (0:00:00.047) 0:02:43.849 ******* skipping: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Reset raw variable] ******************* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml:52 Saturday 24 August 2024 12:36:35 -0400 (0:00:00.032) 0:02:43.881 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_raw": null }, "changed": false } TASK [fedora.linux_system_roles.podman : Remove quadlet file] ****************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:42 Saturday 24 August 2024 12:36:35 -0400 (0:00:00.031) 0:02:43.913 ******* changed: [managed_node1] => { "changed": true, "path": "/etc/containers/systemd/quadlet-demo.network", "state": "absent" } TASK [fedora.linux_system_roles.podman : Refresh systemd] ********************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:48 Saturday 24 August 2024 12:36:35 -0400 (0:00:00.357) 0:02:44.270 ******* ok: [managed_node1] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.podman : Remove managed resource] ************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:58 Saturday 24 August 2024 12:36:36 -0400 (0:00:00.588) 0:02:44.858 ******* changed: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": true } TASK [fedora.linux_system_roles.podman : Remove volumes] *********************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:95 Saturday 24 August 2024 12:36:36 -0400 (0:00:00.458) 0:02:45.316 ******* skipping: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Clear parsed podman variable] ********* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:112 Saturday 24 August 2024 12:36:36 -0400 (0:00:00.050) 0:02:45.367 ******* ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_parsed": null }, "changed": false } TASK [fedora.linux_system_roles.podman : Prune images no longer in use] ******** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:116 Saturday 24 August 2024 12:36:36 -0400 (0:00:00.032) 0:02:45.400 ******* changed: [managed_node1] => { "changed": true, "cmd": [ "podman", "image", "prune", "--all", "-f" ], "delta": "0:00:00.037303", "end": "2024-08-24 12:36:37.138650", "rc": 0, "start": "2024-08-24 12:36:37.101347" } TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:127 Saturday 24 August 2024 12:36:37 -0400 (0:00:00.421) 0:02:45.822 ******* included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Saturday 24 August 2024 12:36:37 -0400 (0:00:00.060) 0:02:45.883 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Saturday 24 August 2024 12:36:37 -0400 (0:00:00.034) 0:02:45.917 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Saturday 24 August 2024 12:36:37 -0400 (0:00:00.033) 0:02:45.951 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : For testing and debugging - images] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:137 Saturday 24 August 2024 12:36:37 -0400 (0:00:00.033) 0:02:45.985 ******* ok: [managed_node1] => { "changed": false, "cmd": [ "podman", "images", "-n" ], "delta": "0:00:00.035636", "end": "2024-08-24 12:36:37.703710", "rc": 0, "start": "2024-08-24 12:36:37.668074" } TASK [fedora.linux_system_roles.podman : For testing and debugging - volumes] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:146 Saturday 24 August 2024 12:36:37 -0400 (0:00:00.448) 0:02:46.433 ******* ok: [managed_node1] => { "changed": false, "cmd": [ "podman", "volume", "ls", "-n" ], "delta": "0:00:00.034941", "end": "2024-08-24 12:36:38.153593", "rc": 0, "start": "2024-08-24 12:36:38.118652" } STDOUT: local wp-pv-claim local envoy-proxy-config local envoy-certificates TASK [fedora.linux_system_roles.podman : For testing and debugging - containers] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:155 Saturday 24 August 2024 12:36:38 -0400 (0:00:00.403) 0:02:46.836 ******* ok: [managed_node1] => { "changed": false, "cmd": [ "podman", "ps", "--noheading" ], "delta": "0:00:00.032042", "end": "2024-08-24 12:36:38.553555", "rc": 0, "start": "2024-08-24 12:36:38.521513" } TASK [fedora.linux_system_roles.podman : For testing and debugging - networks] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:164 Saturday 24 August 2024 12:36:38 -0400 (0:00:00.399) 0:02:47.236 ******* ok: [managed_node1] => { "changed": false, "cmd": [ "podman", "network", "ls", "-n", "-q" ], "delta": "0:00:00.051693", "end": "2024-08-24 12:36:38.969610", "rc": 0, "start": "2024-08-24 12:36:38.917917" } STDOUT: podman podman-default-kube-network TASK [fedora.linux_system_roles.podman : For testing and debugging - secrets] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:173 Saturday 24 August 2024 12:36:39 -0400 (0:00:00.417) 0:02:47.653 ******* ok: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : For testing and debugging - services] *** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:183 Saturday 24 August 2024 12:36:39 -0400 (0:00:00.396) 0:02:48.049 ******* ok: [managed_node1] => { "ansible_facts": { "services": { "NetworkManager-dispatcher.service": { "name": "NetworkManager-dispatcher.service", "source": "systemd", "state": "running", "status": "enabled" }, "NetworkManager-wait-online.service": { "name": "NetworkManager-wait-online.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "NetworkManager.service": { "name": "NetworkManager.service", "source": "systemd", "state": "running", "status": "enabled" }, "auditd.service": { "name": "auditd.service", "source": "systemd", "state": "running", "status": "enabled" }, "auth-rpcgss-module.service": { "name": "auth-rpcgss-module.service", "source": "systemd", "state": "stopped", "status": "static" }, "autovt@.service": { "name": "autovt@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "certmonger.service": { "name": "certmonger.service", "source": "systemd", "state": "running", "status": "enabled" }, "chrony-dnssrv@.service": { "name": "chrony-dnssrv@.service", "source": "systemd", "state": "unknown", "status": "static" }, "chrony-wait.service": { "name": "chrony-wait.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd.service": { "name": "chronyd.service", "source": "systemd", "state": "running", "status": "enabled" }, "cloud-config.service": { "name": "cloud-config.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-final.service": { "name": "cloud-final.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init-hotplugd.service": { "name": "cloud-init-hotplugd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cloud-init-local.service": { "name": "cloud-init-local.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init.service": { "name": "cloud-init.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cni-dhcp.service": { "name": "cni-dhcp.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "console-getty.service": { "name": "console-getty.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "container-getty@.service": { "name": "container-getty@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cpupower.service": { "name": "cpupower.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "crond.service": { "name": "crond.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-org.fedoraproject.FirewallD1.service": { "name": "dbus-org.fedoraproject.FirewallD1.service", "source": "systemd", "state": "active", "status": "enabled" }, "dbus-org.freedesktop.hostname1.service": { "name": "dbus-org.freedesktop.hostname1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.locale1.service": { "name": "dbus-org.freedesktop.locale1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.login1.service": { "name": "dbus-org.freedesktop.login1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.nm-dispatcher.service": { "name": "dbus-org.freedesktop.nm-dispatcher.service", "source": "systemd", "state": "active", "status": "enabled" }, "dbus-org.freedesktop.portable1.service": { "name": "dbus-org.freedesktop.portable1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.timedate1.service": { "name": "dbus-org.freedesktop.timedate1.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "dbus.service": { "name": "dbus.service", "source": "systemd", "state": "running", "status": "static" }, "debug-shell.service": { "name": "debug-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dnf-makecache.service": { "name": "dnf-makecache.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-system-upgrade-cleanup.service": { "name": "dnf-system-upgrade-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "dnf-system-upgrade.service": { "name": "dnf-system-upgrade.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dnsmasq.service": { "name": "dnsmasq.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dracut-cmdline.service": { "name": "dracut-cmdline.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-initqueue.service": { "name": "dracut-initqueue.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-mount.service": { "name": "dracut-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-mount.service": { "name": "dracut-pre-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-pivot.service": { "name": "dracut-pre-pivot.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-trigger.service": { "name": "dracut-pre-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-udev.service": { "name": "dracut-pre-udev.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown-onfailure.service": { "name": "dracut-shutdown-onfailure.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown.service": { "name": "dracut-shutdown.service", "source": "systemd", "state": "stopped", "status": "static" }, "ebtables.service": { "name": "ebtables.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "emergency.service": { "name": "emergency.service", "source": "systemd", "state": "stopped", "status": "static" }, "firewalld.service": { "name": "firewalld.service", "source": "systemd", "state": "running", "status": "enabled" }, "fstrim.service": { "name": "fstrim.service", "source": "systemd", "state": "inactive", "status": "static" }, "getty@.service": { "name": "getty@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "getty@tty1.service": { "name": "getty@tty1.service", "source": "systemd", "state": "running", "status": "unknown" }, "grub-boot-indeterminate.service": { "name": "grub-boot-indeterminate.service", "source": "systemd", "state": "inactive", "status": "static" }, "gssproxy.service": { "name": "gssproxy.service", "source": "systemd", "state": "running", "status": "disabled" }, "halt-local.service": { "name": "halt-local.service", "source": "systemd", "state": "inactive", "status": "static" }, "import-state.service": { "name": "import-state.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "initrd-cleanup.service": { "name": "initrd-cleanup.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-parse-etc.service": { "name": "initrd-parse-etc.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-switch-root.service": { "name": "initrd-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-udevadm-cleanup-db.service": { "name": "initrd-udevadm-cleanup-db.service", "source": "systemd", "state": "stopped", "status": "static" }, "iprdump.service": { "name": "iprdump.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "iprinit.service": { "name": "iprinit.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "iprupdate.service": { "name": "iprupdate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "irqbalance.service": { "name": "irqbalance.service", "source": "systemd", "state": "running", "status": "enabled" }, "kdump.service": { "name": "kdump.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "kmod-static-nodes.service": { "name": "kmod-static-nodes.service", "source": "systemd", "state": "stopped", "status": "static" }, "kvm_stat.service": { "name": "kvm_stat.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "ldconfig.service": { "name": "ldconfig.service", "source": "systemd", "state": "stopped", "status": "static" }, "loadmodules.service": { "name": "loadmodules.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "man-db-cache-update.service": { "name": "man-db-cache-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "man-db-restart-cache-update.service": { "name": "man-db-restart-cache-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "messagebus.service": { "name": "messagebus.service", "source": "systemd", "state": "active", "status": "static" }, "microcode.service": { "name": "microcode.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "nfs-blkmap.service": { "name": "nfs-blkmap.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-convert.service": { "name": "nfs-convert.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-idmapd.service": { "name": "nfs-idmapd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-mountd.service": { "name": "nfs-mountd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-server.service": { "name": "nfs-server.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-utils.service": { "name": "nfs-utils.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfsdcld.service": { "name": "nfsdcld.service", "source": "systemd", "state": "stopped", "status": "static" }, "nftables.service": { "name": "nftables.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nis-domainname.service": { "name": "nis-domainname.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "oddjobd.service": { "name": "oddjobd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "plymouth-halt.service": { "name": "plymouth-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-kexec.service": { "name": "plymouth-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-poweroff.service": { "name": "plymouth-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-quit-wait.service": { "name": "plymouth-quit-wait.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-quit.service": { "name": "plymouth-quit.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-read-write.service": { "name": "plymouth-read-write.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-reboot.service": { "name": "plymouth-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-start.service": { "name": "plymouth-start.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-switch-root-initramfs.service": { "name": "plymouth-switch-root-initramfs.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-switch-root.service": { "name": "plymouth-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "podman-auto-update.service": { "name": "podman-auto-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "podman-clean-transient.service": { "name": "podman-clean-transient.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "podman-kube@.service": { "name": "podman-kube@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "podman-restart.service": { "name": "podman-restart.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "podman.service": { "name": "podman.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "polkit.service": { "name": "polkit.service", "source": "systemd", "state": "running", "status": "static" }, "qemu-guest-agent.service": { "name": "qemu-guest-agent.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "quotaon.service": { "name": "quotaon.service", "source": "systemd", "state": "inactive", "status": "static" }, "rc-local.service": { "name": "rc-local.service", "source": "systemd", "state": "stopped", "status": "static" }, "rdisc.service": { "name": "rdisc.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rescue.service": { "name": "rescue.service", "source": "systemd", "state": "stopped", "status": "static" }, "restraintd.service": { "name": "restraintd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rngd.service": { "name": "rngd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpc-gssd.service": { "name": "rpc-gssd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd-notify.service": { "name": "rpc-statd-notify.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd.service": { "name": "rpc-statd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpcbind.service": { "name": "rpcbind.service", "source": "systemd", "state": "running", "status": "enabled" }, "rsyslog.service": { "name": "rsyslog.service", "source": "systemd", "state": "running", "status": "enabled" }, "selinux-autorelabel-mark.service": { "name": "selinux-autorelabel-mark.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "selinux-autorelabel.service": { "name": "selinux-autorelabel.service", "source": "systemd", "state": "inactive", "status": "static" }, "serial-getty@.service": { "name": "serial-getty@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@.service": { "name": "sshd-keygen@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@ecdsa.service": { "name": "sshd-keygen@ecdsa.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd-keygen@ed25519.service": { "name": "sshd-keygen@ed25519.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd-keygen@rsa.service": { "name": "sshd-keygen@rsa.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd.service": { "name": "sshd.service", "source": "systemd", "state": "running", "status": "enabled" }, "sshd@.service": { "name": "sshd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "sssd-autofs.service": { "name": "sssd-autofs.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-kcm.service": { "name": "sssd-kcm.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "sssd-nss.service": { "name": "sssd-nss.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pac.service": { "name": "sssd-pac.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pam.service": { "name": "sssd-pam.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-ssh.service": { "name": "sssd-ssh.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-sudo.service": { "name": "sssd-sudo.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd.service": { "name": "sssd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "syslog.service": { "name": "syslog.service", "source": "systemd", "state": "active", "status": "enabled" }, "system-update-cleanup.service": { "name": "system-update-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-ask-password-console.service": { "name": "systemd-ask-password-console.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-plymouth.service": { "name": "systemd-ask-password-plymouth.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-wall.service": { "name": "systemd-ask-password-wall.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-backlight@.service": { "name": "systemd-backlight@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-binfmt.service": { "name": "systemd-binfmt.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-coredump@.service": { "name": "systemd-coredump@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-exit.service": { "name": "systemd-exit.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-firstboot.service": { "name": "systemd-firstboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck-root.service": { "name": "systemd-fsck-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck@.service": { "name": "systemd-fsck@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-halt.service": { "name": "systemd-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hibernate-resume@.service": { "name": "systemd-hibernate-resume@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-hibernate.service": { "name": "systemd-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hostnamed.service": { "name": "systemd-hostnamed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hwdb-update.service": { "name": "systemd-hwdb-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-hybrid-sleep.service": { "name": "systemd-hybrid-sleep.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-initctl.service": { "name": "systemd-initctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-catalog-update.service": { "name": "systemd-journal-catalog-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-flush.service": { "name": "systemd-journal-flush.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journald.service": { "name": "systemd-journald.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-kexec.service": { "name": "systemd-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-localed.service": { "name": "systemd-localed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-logind.service": { "name": "systemd-logind.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-machine-id-commit.service": { "name": "systemd-machine-id-commit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-modules-load.service": { "name": "systemd-modules-load.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-portabled.service": { "name": "systemd-portabled.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-poweroff.service": { "name": "systemd-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pstore.service": { "name": "systemd-pstore.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-quotacheck.service": { "name": "systemd-quotacheck.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-random-seed.service": { "name": "systemd-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-reboot.service": { "name": "systemd-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-remount-fs.service": { "name": "systemd-remount-fs.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-resolved.service": { "name": "systemd-resolved.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-rfkill.service": { "name": "systemd-rfkill.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend-then-hibernate.service": { "name": "systemd-suspend-then-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend.service": { "name": "systemd-suspend.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-sysctl.service": { "name": "systemd-sysctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-sysusers.service": { "name": "systemd-sysusers.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-timedated.service": { "name": "systemd-timedated.service", "source": "systemd", "state": "inactive", "status": "masked" }, "systemd-tmpfiles-clean.service": { "name": "systemd-tmpfiles-clean.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev.service": { "name": "systemd-tmpfiles-setup-dev.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup.service": { "name": "systemd-tmpfiles-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-settle.service": { "name": "systemd-udev-settle.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-udev-trigger.service": { "name": "systemd-udev-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udevd.service": { "name": "systemd-udevd.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-update-done.service": { "name": "systemd-update-done.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp-runlevel.service": { "name": "systemd-update-utmp-runlevel.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp.service": { "name": "systemd-update-utmp.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-user-sessions.service": { "name": "systemd-user-sessions.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-vconsole-setup.service": { "name": "systemd-vconsole-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-volatile-root.service": { "name": "systemd-volatile-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "tcsd.service": { "name": "tcsd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "teamd@.service": { "name": "teamd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "timedatex.service": { "name": "timedatex.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "tuned.service": { "name": "tuned.service", "source": "systemd", "state": "running", "status": "enabled" }, "unbound-anchor.service": { "name": "unbound-anchor.service", "source": "systemd", "state": "stopped", "status": "static" }, "user-runtime-dir@.service": { "name": "user-runtime-dir@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user-runtime-dir@0.service": { "name": "user-runtime-dir@0.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "user@.service": { "name": "user@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user@0.service": { "name": "user@0.service", "source": "systemd", "state": "running", "status": "unknown" } } }, "changed": false } TASK [fedora.linux_system_roles.podman : Create and update quadlets] *********** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:114 Saturday 24 August 2024 12:36:41 -0400 (0:00:01.615) 0:02:49.665 ******* skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Cancel linger] ************************ task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:167 Saturday 24 August 2024 12:36:41 -0400 (0:00:00.032) 0:02:49.698 ******* TASK [fedora.linux_system_roles.podman : Handle credential files - absent] ***** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:173 Saturday 24 August 2024 12:36:41 -0400 (0:00:00.065) 0:02:49.763 ******* skipping: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Handle certs.d files - absent] ******** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:182 Saturday 24 August 2024 12:36:41 -0400 (0:00:00.032) 0:02:49.796 ******* skipping: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [Ensure no resources] ***************************************************** task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_demo.yml:188 Saturday 24 August 2024 12:36:41 -0400 (0:00:00.030) 0:02:49.827 ******* fatal: [managed_node1]: FAILED! => { "assertion": "__podman_test_debug_volumes.stdout == \"\"", "changed": false, "evaluated_to": false } MSG: Assertion failed TASK [Debug] ******************************************************************* task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_demo.yml:199 Saturday 24 August 2024 12:36:41 -0400 (0:00:00.043) 0:02:49.871 ******* ok: [managed_node1] => { "changed": false, "cmd": "exec 1>&2\nset -x\nset -o pipefail\nsystemctl list-units --plain -l --all | grep quadlet || :\nsystemctl list-unit-files --all | grep quadlet || :\nsystemctl list-units --plain --failed -l --all | grep quadlet || :\n", "delta": "0:00:00.322673", "end": "2024-08-24 12:36:41.863123", "rc": 0, "start": "2024-08-24 12:36:41.540450" } STDERR: + set -o pipefail + systemctl list-units --plain -l --all + grep quadlet + : + systemctl list-unit-files --all + grep quadlet + : + systemctl list-units --plain --failed -l --all + grep quadlet + : TASK [Get journald] ************************************************************ task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_demo.yml:209 Saturday 24 August 2024 12:36:41 -0400 (0:00:00.674) 0:02:50.545 ******* fatal: [managed_node1]: FAILED! => { "changed": false, "cmd": [ "journalctl", "-ex" ], "delta": "0:00:00.106022", "end": "2024-08-24 12:36:42.321270", "failed_when_result": true, "rc": 0, "start": "2024-08-24 12:36:42.215248" } STDOUT: -- Logs begin at Sat 2024-08-24 12:23:16 EDT, end at Sat 2024-08-24 12:36:42 EDT. -- Aug 24 12:31:03 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state. Aug 24 12:31:03 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 32ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources. Aug 24 12:31:03 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813. -- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up. -- -- The start-up result is done. Aug 24 12:31:03 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state. Aug 24 12:31:03 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 33ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources. Aug 24 12:31:03 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813. -- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up. -- -- The start-up result is done. Aug 24 12:31:03 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[37282]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None Aug 24 12:31:03 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state. Aug 24 12:31:03 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 38ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources. Aug 24 12:31:03 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813. -- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up. -- -- The start-up result is done. Aug 24 12:31:04 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state. Aug 24 12:31:04 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 32ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources. Aug 24 12:31:04 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813. -- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up. -- -- The start-up result is done. Aug 24 12:31:04 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state. Aug 24 12:31:04 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 33ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources. Aug 24 12:31:04 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813. -- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up. -- -- The start-up result is done. Aug 24 12:31:04 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[37501]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Aug 24 12:31:04 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state. Aug 24 12:31:04 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 40ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources. Aug 24 12:31:04 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813. -- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up. -- -- The start-up result is done. Aug 24 12:31:04 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state. Aug 24 12:31:04 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 37ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources. Aug 24 12:31:04 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813. -- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up. -- -- The start-up result is done. Aug 24 12:31:04 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state. Aug 24 12:31:04 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 31ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources. Aug 24 12:31:05 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813. -- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up. -- -- The start-up result is done. Aug 24 12:31:05 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state. Aug 24 12:31:05 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 31ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources. Aug 24 12:31:05 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813. -- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up. -- -- The start-up result is done. Aug 24 12:31:05 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state. Aug 24 12:31:05 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 30ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources. Aug 24 12:31:05 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813. -- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up. -- -- The start-up result is done. Aug 24 12:31:05 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state. Aug 24 12:31:05 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 30ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources. Aug 24 12:31:05 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813. -- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up. -- -- The start-up result is done. Aug 24 12:31:05 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state. Aug 24 12:31:05 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 31ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources. Aug 24 12:31:05 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813. -- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up. -- -- The start-up result is done. Aug 24 12:31:05 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state. Aug 24 12:31:05 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 30ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources. Aug 24 12:31:05 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813. -- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up. -- -- The start-up result is done. Aug 24 12:31:06 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state. Aug 24 12:31:06 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 32ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources. Aug 24 12:31:06 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813. -- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up. -- -- The start-up result is done. Aug 24 12:31:06 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state. Aug 24 12:31:06 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 33ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources. Aug 24 12:31:06 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813. -- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up. -- -- The start-up result is done. Aug 24 12:31:06 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state. Aug 24 12:31:06 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 39ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources. Aug 24 12:31:06 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813. -- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up. -- -- The start-up result is done. Aug 24 12:31:06 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state. Aug 24 12:31:06 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 37ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources. Aug 24 12:31:06 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813. -- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up. -- -- The start-up result is done. Aug 24 12:31:07 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state. Aug 24 12:31:07 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 38ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources. Aug 24 12:31:07 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813. -- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up. -- -- The start-up result is done. Aug 24 12:31:07 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state. Aug 24 12:31:07 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 37ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources. Aug 24 12:31:07 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813. -- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up. -- -- The start-up result is done. Aug 24 12:31:07 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state. Aug 24 12:31:07 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 38ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources. Aug 24 12:31:07 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813. -- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up. -- -- The start-up result is done. Aug 24 12:31:07 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state. Aug 24 12:31:07 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 32ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources. Aug 24 12:31:07 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813. -- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up. -- -- The start-up result is done. Aug 24 12:31:08 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state. Aug 24 12:31:08 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 30ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources. Aug 24 12:31:08 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813. -- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up. -- -- The start-up result is done. Aug 24 12:31:08 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state. Aug 24 12:31:08 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 33ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources. Aug 24 12:31:08 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813. -- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up. -- -- The start-up result is done. Aug 24 12:31:08 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state. Aug 24 12:31:08 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 37ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources. Aug 24 12:31:08 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[38402]: ansible-command Invoked with _raw_params=podman --version warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:31:08 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813. -- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up. -- -- The start-up result is done. Aug 24 12:31:08 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state. Aug 24 12:31:08 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 31ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources. Aug 24 12:31:08 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813. -- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up. -- -- The start-up result is done. Aug 24 12:31:08 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state. Aug 24 12:31:08 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 30ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources. Aug 24 12:31:09 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813. -- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up. -- -- The start-up result is done. Aug 24 12:31:09 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state. Aug 24 12:31:09 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 33ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources. Aug 24 12:31:09 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813. -- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up. -- -- The start-up result is done. Aug 24 12:31:09 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state. Aug 24 12:31:09 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 32ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources. Aug 24 12:31:09 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813. -- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up. -- -- The start-up result is done. Aug 24 12:31:09 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state. Aug 24 12:31:09 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 36ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources. Aug 24 12:31:09 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[38707]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None Aug 24 12:31:09 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813. -- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up. -- -- The start-up result is done. Aug 24 12:31:09 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state. Aug 24 12:31:09 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 32ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources. Aug 24 12:31:09 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813. -- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up. -- -- The start-up result is done. Aug 24 12:31:09 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state. Aug 24 12:31:09 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 34ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources. Aug 24 12:31:10 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813. -- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up. -- -- The start-up result is done. Aug 24 12:31:10 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state. Aug 24 12:31:10 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 40ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources. Aug 24 12:31:10 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[38937]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Aug 24 12:31:10 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813. -- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up. -- -- The start-up result is done. Aug 24 12:31:10 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state. Aug 24 12:31:10 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 32ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources. Aug 24 12:31:10 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813. -- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up. -- -- The start-up result is done. Aug 24 12:31:10 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state. Aug 24 12:31:10 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 32ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources. Aug 24 12:31:10 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813. -- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up. -- -- The start-up result is done. Aug 24 12:31:10 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state. Aug 24 12:31:10 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 31ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources. Aug 24 12:31:10 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813. -- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up. -- -- The start-up result is done. Aug 24 12:31:11 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state. Aug 24 12:31:11 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 31ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources. Aug 24 12:31:11 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813. -- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up. -- -- The start-up result is done. Aug 24 12:31:11 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state. Aug 24 12:31:11 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 32ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources. Aug 24 12:31:11 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813. -- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up. -- -- The start-up result is done. Aug 24 12:31:11 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state. Aug 24 12:31:11 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 31ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources. Aug 24 12:31:11 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813. -- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up. -- -- The start-up result is done. Aug 24 12:31:11 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state. Aug 24 12:31:11 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 30ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources. Aug 24 12:31:11 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813. -- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up. -- -- The start-up result is done. Aug 24 12:31:11 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state. Aug 24 12:31:11 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 32ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources. Aug 24 12:31:11 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813. -- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up. -- -- The start-up result is done. Aug 24 12:31:11 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state. Aug 24 12:31:11 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 30ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources. Aug 24 12:31:12 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813. -- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up. -- -- The start-up result is done. Aug 24 12:31:12 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state. Aug 24 12:31:12 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 31ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources. Aug 24 12:31:12 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813. -- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up. -- -- The start-up result is done. Aug 24 12:31:12 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state. Aug 24 12:31:12 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 33ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources. Aug 24 12:31:12 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813. -- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up. -- -- The start-up result is done. Aug 24 12:31:12 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state. Aug 24 12:31:12 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 38ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources. Aug 24 12:31:12 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[39479]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None Aug 24 12:31:12 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813. -- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up. -- -- The start-up result is done. Aug 24 12:31:12 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state. Aug 24 12:31:12 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 32ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources. Aug 24 12:31:12 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813. -- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up. -- -- The start-up result is done. Aug 24 12:31:13 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state. Aug 24 12:31:13 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 34ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources. Aug 24 12:31:13 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[39681]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Aug 24 12:31:13 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813. -- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up. -- -- The start-up result is done. Aug 24 12:31:13 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state. Aug 24 12:31:13 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 31ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources. Aug 24 12:31:13 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813. -- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up. -- -- The start-up result is done. Aug 24 12:31:13 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state. Aug 24 12:31:13 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 31ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources. Aug 24 12:31:13 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813. -- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up. -- -- The start-up result is done. Aug 24 12:31:13 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state. Aug 24 12:31:13 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 30ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources. Aug 24 12:31:13 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813. -- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up. -- -- The start-up result is done. Aug 24 12:31:13 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state. Aug 24 12:31:13 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 30ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources. Aug 24 12:31:13 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813. -- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up. -- -- The start-up result is done. Aug 24 12:31:13 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state. Aug 24 12:31:13 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 29ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources. Aug 24 12:31:14 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813. -- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up. -- -- The start-up result is done. Aug 24 12:31:14 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state. Aug 24 12:31:14 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 30ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources. Aug 24 12:31:14 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813. -- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up. -- -- The start-up result is done. Aug 24 12:31:14 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state. Aug 24 12:31:14 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 30ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources. Aug 24 12:31:14 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813. -- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up. -- -- The start-up result is done. Aug 24 12:31:14 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state. Aug 24 12:31:14 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 33ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources. Aug 24 12:31:14 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813. -- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up. -- -- The start-up result is done. Aug 24 12:31:14 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[40107]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None Aug 24 12:31:14 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state. Aug 24 12:31:14 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 38ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources. Aug 24 12:31:14 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813. -- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up. -- -- The start-up result is done. Aug 24 12:31:15 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state. Aug 24 12:31:15 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 32ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources. Aug 24 12:31:15 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813. -- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up. -- -- The start-up result is done. Aug 24 12:31:15 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state. Aug 24 12:31:15 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 38ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources. Aug 24 12:31:15 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[40323]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Aug 24 12:31:15 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813. -- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up. -- -- The start-up result is done. Aug 24 12:31:15 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state. Aug 24 12:31:15 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 31ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources. Aug 24 12:31:15 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813. -- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up. -- -- The start-up result is done. Aug 24 12:31:15 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state. Aug 24 12:31:15 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 32ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources. Aug 24 12:31:15 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813. -- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up. -- -- The start-up result is done. Aug 24 12:31:15 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state. Aug 24 12:31:15 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 30ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources. Aug 24 12:31:15 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813. -- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up. -- -- The start-up result is done. Aug 24 12:31:16 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state. Aug 24 12:31:16 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 30ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources. Aug 24 12:31:16 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813. -- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up. -- -- The start-up result is done. Aug 24 12:31:16 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state. Aug 24 12:31:16 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 30ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources. Aug 24 12:31:16 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813. -- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up. -- -- The start-up result is done. Aug 24 12:31:16 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state. Aug 24 12:31:16 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 30ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources. Aug 24 12:31:16 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813. -- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up. -- -- The start-up result is done. Aug 24 12:31:16 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state. Aug 24 12:31:16 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 32ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources. Aug 24 12:31:16 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813. -- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up. -- -- The start-up result is done. Aug 24 12:31:16 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state. Aug 24 12:31:16 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 30ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources. Aug 24 12:31:16 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813. -- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up. -- -- The start-up result is done. Aug 24 12:31:16 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state. Aug 24 12:31:16 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 30ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources. Aug 24 12:31:17 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813. -- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up. -- -- The start-up result is done. Aug 24 12:31:17 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state. Aug 24 12:31:17 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 31ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources. Aug 24 12:31:17 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813. -- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up. -- -- The start-up result is done. Aug 24 12:31:17 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state. Aug 24 12:31:17 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 35ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources. Aug 24 12:31:17 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813. -- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up. -- -- The start-up result is done. Aug 24 12:31:17 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state. Aug 24 12:31:17 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 35ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources. Aug 24 12:31:17 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[40875]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None Aug 24 12:31:17 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813. -- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up. -- -- The start-up result is done. Aug 24 12:31:17 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state. Aug 24 12:31:17 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 33ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources. Aug 24 12:31:17 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813. -- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up. -- -- The start-up result is done. Aug 24 12:31:18 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state. Aug 24 12:31:18 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 33ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources. Aug 24 12:31:18 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813. -- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up. -- -- The start-up result is done. Aug 24 12:31:18 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state. Aug 24 12:31:18 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 37ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources. Aug 24 12:31:18 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[41104]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Aug 24 12:31:18 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813. -- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up. -- -- The start-up result is done. Aug 24 12:31:18 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state. Aug 24 12:31:18 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 30ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources. Aug 24 12:31:18 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813. -- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up. -- -- The start-up result is done. Aug 24 12:31:18 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state. Aug 24 12:31:18 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 32ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources. Aug 24 12:31:18 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813. -- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up. -- -- The start-up result is done. Aug 24 12:31:18 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state. Aug 24 12:31:18 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 30ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources. Aug 24 12:31:18 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813. -- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up. -- -- The start-up result is done. Aug 24 12:31:19 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state. Aug 24 12:31:19 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 33ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources. Aug 24 12:31:19 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813. -- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up. -- -- The start-up result is done. Aug 24 12:31:19 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state. Aug 24 12:31:19 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 38ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources. Aug 24 12:31:19 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[41406]: ansible-command Invoked with _raw_params= warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:31:19 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813. -- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up. -- -- The start-up result is done. Aug 24 12:31:19 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state. Aug 24 12:31:19 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 33ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources. Aug 24 12:31:19 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813. -- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up. -- -- The start-up result is done. Aug 24 12:31:19 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state. Aug 24 12:31:19 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 34ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources. Aug 24 12:31:19 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813. -- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up. -- -- The start-up result is done. Aug 24 12:31:19 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state. Aug 24 12:31:19 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 36ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources. Aug 24 12:31:20 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813. -- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up. -- -- The start-up result is done. Aug 24 12:31:20 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[41637]: ansible-systemd Invoked with name= scope=system state=stopped enabled=False daemon_reload=False daemon_reexec=False no_block=False force=None masked=None user=None Aug 24 12:31:20 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Reloading. Aug 24 12:31:20 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state. Aug 24 12:31:20 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 38ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources. Aug 24 12:31:20 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Stopping A template for running K8s workloads via podman-kube-play... -- Subject: Unit podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service has begun shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service has begun shutting down. Aug 24 12:31:20 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813. -- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up. -- -- The start-up result is done. Aug 24 12:31:20 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-49bd35886401f2696f39559535070a7acae98c93915196aab21a7de3174e59ba.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-49bd35886401f2696f39559535070a7acae98c93915196aab21a7de3174e59ba.scope has successfully entered the 'dead' state. Aug 24 12:31:20 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-49bd35886401f2696f39559535070a7acae98c93915196aab21a7de3174e59ba.scope: Consumed 32ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-49bd35886401f2696f39559535070a7acae98c93915196aab21a7de3174e59ba.scope completed and consumed the indicated resources. Aug 24 12:31:20 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state. Aug 24 12:31:20 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Stopped libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813. -- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished shutting down. Aug 24 12:31:20 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 600us CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources. Aug 24 12:31:20 ip-10-31-40-15.us-east-1.aws.redhat.com conmon[41736]: conmon 765d8c2cddaf785a845e : runtime stderr: runc create failed: unable to start container process: can't get final child's PID from pipe: EOF Aug 24 12:31:20 ip-10-31-40-15.us-east-1.aws.redhat.com conmon[41736]: conmon 765d8c2cddaf785a845e : Failed to create container: exit status 1 Aug 24 12:31:20 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay-df494c4e221d99e0f620fb00cd28f2cc006ffcc851b7d7eade776335b55ac463-merged.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay-df494c4e221d99e0f620fb00cd28f2cc006ffcc851b7d7eade776335b55ac463-merged.mount has successfully entered the 'dead' state. Aug 24 12:31:20 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: cni-podman1: port 1(veth727493b2) entered disabled state Aug 24 12:31:20 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: device veth727493b2 left promiscuous mode Aug 24 12:31:20 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: cni-podman1: port 1(veth727493b2) entered disabled state Aug 24 12:31:20 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: run-netns-netns\x2de2bae42d\x2da696\x2d4b35\x2d2126\x2d284fd3d6a73b.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit run-netns-netns\x2de2bae42d\x2da696\x2d4b35\x2d2126\x2d284fd3d6a73b.mount has successfully entered the 'dead' state. Aug 24 12:31:20 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-49bd35886401f2696f39559535070a7acae98c93915196aab21a7de3174e59ba-userdata-shm.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay\x2dcontainers-49bd35886401f2696f39559535070a7acae98c93915196aab21a7de3174e59ba-userdata-shm.mount has successfully entered the 'dead' state. Aug 24 12:31:20 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay-536a0786a016d674449abd21820ed1ec2659103dcfcc60e560eb1b13ba0fb457-merged.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay-536a0786a016d674449abd21820ed1ec2659103dcfcc60e560eb1b13ba0fb457-merged.mount has successfully entered the 'dead' state. Aug 24 12:31:20 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Removed slice cgroup machine-libpod_pod_c5e3848e0e861be93b14b306c309ae7576e20076beb5aaaf5660f3b6bec6eb4e.slice. -- Subject: Unit machine-libpod_pod_c5e3848e0e861be93b14b306c309ae7576e20076beb5aaaf5660f3b6bec6eb4e.slice has finished shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit machine-libpod_pod_c5e3848e0e861be93b14b306c309ae7576e20076beb5aaaf5660f3b6bec6eb4e.slice has finished shutting down. Aug 24 12:31:20 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: machine-libpod_pod_c5e3848e0e861be93b14b306c309ae7576e20076beb5aaaf5660f3b6bec6eb4e.slice: Consumed 11.825s CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit machine-libpod_pod_c5e3848e0e861be93b14b306c309ae7576e20076beb5aaaf5660f3b6bec6eb4e.slice completed and consumed the indicated resources. Aug 24 12:31:20 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-80915bd8fbddf68e48445610f66f32c8466405cc58a3790cf9782f24652d9f2c.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-80915bd8fbddf68e48445610f66f32c8466405cc58a3790cf9782f24652d9f2c.scope has successfully entered the 'dead' state. Aug 24 12:31:20 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-80915bd8fbddf68e48445610f66f32c8466405cc58a3790cf9782f24652d9f2c.scope: Consumed 33ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-80915bd8fbddf68e48445610f66f32c8466405cc58a3790cf9782f24652d9f2c.scope completed and consumed the indicated resources. Aug 24 12:31:20 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-80915bd8fbddf68e48445610f66f32c8466405cc58a3790cf9782f24652d9f2c-userdata-shm.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay\x2dcontainers-80915bd8fbddf68e48445610f66f32c8466405cc58a3790cf9782f24652d9f2c-userdata-shm.mount has successfully entered the 'dead' state. Aug 24 12:31:20 ip-10-31-40-15.us-east-1.aws.redhat.com podman[41721]: Pods stopped: Aug 24 12:31:20 ip-10-31-40-15.us-east-1.aws.redhat.com podman[41721]: c5e3848e0e861be93b14b306c309ae7576e20076beb5aaaf5660f3b6bec6eb4e Aug 24 12:31:20 ip-10-31-40-15.us-east-1.aws.redhat.com podman[41721]: Pods removed: Aug 24 12:31:20 ip-10-31-40-15.us-east-1.aws.redhat.com podman[41721]: c5e3848e0e861be93b14b306c309ae7576e20076beb5aaaf5660f3b6bec6eb4e Aug 24 12:31:20 ip-10-31-40-15.us-east-1.aws.redhat.com podman[41721]: Secrets removed: Aug 24 12:31:20 ip-10-31-40-15.us-east-1.aws.redhat.com podman[41721]: Volumes removed: Aug 24 12:31:20 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service has successfully entered the 'dead' state. Aug 24 12:31:20 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Stopped A template for running K8s workloads via podman-kube-play. -- Subject: Unit podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service has finished shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service has finished shutting down. Aug 24 12:31:21 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[41993]: ansible-stat Invoked with path=/etc/containers/ansible-kubernetes.d/auth_test_1_kube.yml follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Aug 24 12:31:21 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay-a55b6a86d999a458b4a32b86117737eed9e58e0082a4227e2d3fc866782b9269-merged.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay-a55b6a86d999a458b4a32b86117737eed9e58e0082a4227e2d3fc866782b9269-merged.mount has successfully entered the 'dead' state. Aug 24 12:31:21 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[42118]: ansible-containers.podman.podman_play Invoked with state=absent kube_file=/etc/containers/ansible-kubernetes.d/auth_test_1_kube.yml executable=podman annotation=None authfile=None build=None cert_dir=None configmap=None context_dir=None seccomp_profile_root=None username=None password=NOT_LOGGING_PARAMETER log_driver=None log_opt=None network=None tls_verify=None debug=None quiet=None recreate=None userns=None log_level=None quadlet_dir=None quadlet_filename=None quadlet_options=None Aug 24 12:31:21 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[42118]: ansible-containers.podman.podman_play version: 4.9.4-dev, kube file /etc/containers/ansible-kubernetes.d/auth_test_1_kube.yml Aug 24 12:31:22 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[42254]: ansible-file Invoked with path=/etc/containers/ansible-kubernetes.d/auth_test_1_kube.yml state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Aug 24 12:31:22 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[42377]: ansible-command Invoked with _raw_params=podman image prune -f warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:31:24 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[42507]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None Aug 24 12:31:24 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[42631]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Aug 24 12:31:26 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[42756]: ansible-systemd Invoked with name=auth_test_1_quadlet.service scope=system state=stopped enabled=False force=True daemon_reload=False daemon_reexec=False no_block=False masked=None user=None Aug 24 12:31:26 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Reloading. Aug 24 12:31:26 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[42911]: ansible-stat Invoked with path=/etc/containers/systemd/auth_test_1_quadlet.container follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Aug 24 12:31:27 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[43159]: ansible-file Invoked with path=/etc/containers/systemd/auth_test_1_quadlet.container state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Aug 24 12:31:28 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[43282]: ansible-systemd Invoked with daemon_reload=True scope=system daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None user=None Aug 24 12:31:28 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Reloading. Aug 24 12:31:31 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[43565]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None Aug 24 12:31:31 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[43689]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Aug 24 12:31:34 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[44060]: ansible-file Invoked with path=/root/.config/containers state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Aug 24 12:31:35 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[44183]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None Aug 24 12:31:36 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[44307]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Aug 24 12:31:39 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[45047]: ansible-user Invoked with name=auth_test_user1 uid=2001 state=present non_unique=False force=False remove=False create_home=True system=False move_home=False append=False ssh_key_bits=0 ssh_key_type=rsa ssh_key_comment=ansible-generated on ip-10-31-40-15.us-east-1.aws.redhat.com update_password=always group=None groups=None comment=None home=None shell=None password=NOT_LOGGING_PARAMETER login_class=None hidden=None seuser=None skeleton=None generate_ssh_key=None ssh_key_file=None ssh_key_passphrase=NOT_LOGGING_PARAMETER expires=None password_lock=None local=None profile=None authorization=None role=None Aug 24 12:31:39 ip-10-31-40-15.us-east-1.aws.redhat.com useradd[45051]: new group: name=auth_test_user1, GID=2001 Aug 24 12:31:39 ip-10-31-40-15.us-east-1.aws.redhat.com useradd[45051]: new user: name=auth_test_user1, UID=2001, GID=2001, home=/home/auth_test_user1, shell=/bin/bash Aug 24 12:31:43 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[45304]: ansible-command Invoked with _raw_params=podman --version warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:31:43 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[45433]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None Aug 24 12:31:44 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[45557]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Aug 24 12:31:47 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[45682]: ansible-getent Invoked with database=passwd key=auth_test_user1 fail_key=False service=None split=None Aug 24 12:31:47 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[45806]: ansible-getent Invoked with database=group key=2001 fail_key=False service=None split=None Aug 24 12:31:48 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[45930]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Aug 24 12:31:48 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[46055]: ansible-command Invoked with _raw_params=getsubids auth_test_user1 warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:31:49 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[46179]: ansible-command Invoked with _raw_params=getsubids -g auth_test_user1 warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:31:50 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[46303]: ansible-file Invoked with path=/home/auth_test_user1/.config/containers/certs.d/localhost:5000 state=directory owner=auth_test_user1 group=2001 mode=0700 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Aug 24 12:31:52 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[46650]: ansible-getent Invoked with database=group key=2001 fail_key=False service=None split=None Aug 24 12:31:53 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[46774]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Aug 24 12:31:53 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[46899]: ansible-command Invoked with _raw_params=getsubids auth_test_user1 warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:31:53 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[47023]: ansible-command Invoked with _raw_params=getsubids -g auth_test_user1 warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:31:54 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[47147]: ansible-file Invoked with path=/home/auth_test_user1/.config/containers state=directory owner=auth_test_user1 group=2001 mode=0700 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Aug 24 12:31:56 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[47494]: ansible-getent Invoked with database=group key=2001 fail_key=False service=None split=None Aug 24 12:31:57 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[47618]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Aug 24 12:31:57 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[47743]: ansible-command Invoked with _raw_params=getsubids auth_test_user1 warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:31:58 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[47867]: ansible-command Invoked with _raw_params=getsubids -g auth_test_user1 warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:31:59 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[47991]: ansible-command Invoked with _raw_params= warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:32:00 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[48115]: ansible-command Invoked with creates=/var/lib/systemd/linger/auth_test_user1 _raw_params=loginctl enable-linger auth_test_user1 warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None removes=None stdin=None Aug 24 12:32:00 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Created slice User Slice of UID 2001. -- Subject: Unit user-2001.slice has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit user-2001.slice has finished starting up. -- -- The start-up result is done. Aug 24 12:32:00 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Starting User runtime directory /run/user/2001... -- Subject: Unit user-runtime-dir@2001.service has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit user-runtime-dir@2001.service has begun starting up. Aug 24 12:32:00 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started User runtime directory /run/user/2001. -- Subject: Unit user-runtime-dir@2001.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit user-runtime-dir@2001.service has finished starting up. -- -- The start-up result is done. Aug 24 12:32:00 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Starting User Manager for UID 2001... -- Subject: Unit user@2001.service has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit user@2001.service has begun starting up. Aug 24 12:32:00 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: pam_unix(systemd-user:session): session opened for user auth_test_user1 by (uid=0) Aug 24 12:32:00 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Reached target Paths. -- Subject: Unit UNIT has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished starting up. -- -- The start-up result is done. Aug 24 12:32:00 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Starting D-Bus User Message Bus Socket. -- Subject: Unit UNIT has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has begun starting up. Aug 24 12:32:00 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Started Mark boot as successful after the user session has run 2 minutes. -- Subject: Unit UNIT has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished starting up. -- -- The start-up result is done. Aug 24 12:32:00 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Reached target Timers. -- Subject: Unit UNIT has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished starting up. -- -- The start-up result is done. Aug 24 12:32:00 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Listening on D-Bus User Message Bus Socket. -- Subject: Unit UNIT has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished starting up. -- -- The start-up result is done. Aug 24 12:32:00 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Reached target Sockets. -- Subject: Unit UNIT has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished starting up. -- -- The start-up result is done. Aug 24 12:32:00 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Reached target Basic System. -- Subject: Unit UNIT has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished starting up. -- -- The start-up result is done. Aug 24 12:32:00 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Reached target Default. -- Subject: Unit UNIT has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished starting up. -- -- The start-up result is done. Aug 24 12:32:00 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Startup finished in 37ms. -- Subject: User manager start-up is now complete -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The user manager instance for user 2001 has been started. All services queued -- for starting have been started. Note that other services might still be starting -- up or be started at any later time. -- -- Startup of the manager took 37357 microseconds. Aug 24 12:32:00 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started User Manager for UID 2001. -- Subject: Unit user@2001.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit user@2001.service has finished starting up. -- -- The start-up result is done. Aug 24 12:32:01 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[48255]: root : TTY=pts/0 ; PWD=/root ; USER=auth_test_user1 ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-ucyqvqaeqcbwzlreelmtcakpvqhqhicr ; XDG_RUNTIME_DIR=/run/user/2001 /usr/libexec/platform-python /var/tmp/ansible-tmp-1724517121.0929248-20827-180187680463393/AnsiballZ_podman_image.py' Aug 24 12:32:01 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[48255]: pam_unix(sudo:session): session opened for user auth_test_user1 by root(uid=0) Aug 24 12:32:01 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Started D-Bus User Message Bus. -- Subject: Unit UNIT has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished starting up. -- -- The start-up result is done. Aug 24 12:32:01 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Created slice user.slice. -- Subject: Unit UNIT has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished starting up. -- -- The start-up result is done. Aug 24 12:32:01 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Started podman-48267.scope. -- Subject: Unit UNIT has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished starting up. -- -- The start-up result is done. Aug 24 12:32:01 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Started podman-pause-0e20a755.scope. -- Subject: Unit UNIT has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished starting up. -- -- The start-up result is done. Aug 24 12:32:01 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Started podman-48283.scope. -- Subject: Unit UNIT has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished starting up. -- -- The start-up result is done. Aug 24 12:32:01 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Started podman-48290.scope. -- Subject: Unit UNIT has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished starting up. -- -- The start-up result is done. Aug 24 12:32:02 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Started podman-48306.scope. -- Subject: Unit UNIT has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished starting up. -- -- The start-up result is done. Aug 24 12:32:02 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[48255]: pam_unix(sudo:session): session closed for user auth_test_user1 Aug 24 12:32:02 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[48436]: ansible-stat Invoked with path=/home/auth_test_user1/.config/containers/ansible-kubernetes.d/auth_test_1_kube.yml follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Aug 24 12:32:03 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[48559]: ansible-file Invoked with path=/home/auth_test_user1/.config/containers/ansible-kubernetes.d state=directory owner=auth_test_user1 group=2001 mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Aug 24 12:32:03 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[48682]: ansible-stat Invoked with path=/home/auth_test_user1/.config/containers/ansible-kubernetes.d/auth_test_1_kube.yml follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Aug 24 12:32:03 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[48781]: ansible-copy Invoked with dest=/home/auth_test_user1/.config/containers/ansible-kubernetes.d/auth_test_1_kube.yml owner=auth_test_user1 group=2001 mode=0644 src=/root/.ansible/tmp/ansible-tmp-1724517123.15822-20917-242171895731390/source _original_basename=tmpj93_bt8l follow=False checksum=fb0097683a2e5c8909a8037d64ddc1b350aed0be backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None regexp=None delimiter=None Aug 24 12:32:04 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[48906]: root : TTY=pts/0 ; PWD=/root ; USER=auth_test_user1 ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-tltabhiijzzijcbpxvlygvztxmltcxps ; XDG_RUNTIME_DIR=/run/user/2001 /usr/libexec/platform-python /var/tmp/ansible-tmp-1724517123.8727906-20945-22576776740070/AnsiballZ_podman_play.py' Aug 24 12:32:04 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[48906]: pam_unix(sudo:session): session opened for user auth_test_user1 by root(uid=0) Aug 24 12:32:04 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[48909]: ansible-containers.podman.podman_play Invoked with state=started kube_file=/home/auth_test_user1/.config/containers/ansible-kubernetes.d/auth_test_1_kube.yml executable=podman annotation=None authfile=None build=None cert_dir=None configmap=None context_dir=None seccomp_profile_root=None username=None password=NOT_LOGGING_PARAMETER log_driver=None log_opt=None network=None tls_verify=None debug=None quiet=None recreate=None userns=None log_level=None quadlet_dir=None quadlet_filename=None quadlet_options=None Aug 24 12:32:04 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Started podman-48917.scope. -- Subject: Unit UNIT has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished starting up. -- -- The start-up result is done. Aug 24 12:32:04 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: tun: Universal TUN/TAP device driver, 1.6 Aug 24 12:32:04 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Started rootless-netns-804780fa.scope. -- Subject: Unit UNIT has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished starting up. -- -- The start-up result is done. Aug 24 12:32:04 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: IPv6: ADDRCONF(NETDEV_UP): vethd69b4c65: link is not ready Aug 24 12:32:04 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: cni-podman2: port 1(vethd69b4c65) entered blocking state Aug 24 12:32:04 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: cni-podman2: port 1(vethd69b4c65) entered disabled state Aug 24 12:32:04 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: device vethd69b4c65 entered promiscuous mode Aug 24 12:32:04 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: IPv6: ADDRCONF(NETDEV_UP): eth0: link is not ready Aug 24 12:32:04 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: IPv6: ADDRCONF(NETDEV_CHANGE): eth0: link becomes ready Aug 24 12:32:04 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: IPv6: ADDRCONF(NETDEV_CHANGE): vethd69b4c65: link becomes ready Aug 24 12:32:04 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: cni-podman2: port 1(vethd69b4c65) entered blocking state Aug 24 12:32:04 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: cni-podman2: port 1(vethd69b4c65) entered forwarding state Aug 24 12:32:04 ip-10-31-40-15.us-east-1.aws.redhat.com dnsmasq[49050]: listening on cni-podman2(#3): 10.89.1.1 Aug 24 12:32:04 ip-10-31-40-15.us-east-1.aws.redhat.com dnsmasq[49052]: started, version 2.79 cachesize 150 Aug 24 12:32:04 ip-10-31-40-15.us-east-1.aws.redhat.com dnsmasq[49052]: compile time options: IPv6 GNU-getopt DBus no-i18n IDN2 DHCP DHCPv6 no-Lua TFTP no-conntrack ipset auth DNSSEC loop-detect inotify Aug 24 12:32:04 ip-10-31-40-15.us-east-1.aws.redhat.com dnsmasq[49052]: using local addresses only for domain dns.podman Aug 24 12:32:04 ip-10-31-40-15.us-east-1.aws.redhat.com dnsmasq[49052]: reading /etc/resolv.conf Aug 24 12:32:04 ip-10-31-40-15.us-east-1.aws.redhat.com dnsmasq[49052]: using local addresses only for domain dns.podman Aug 24 12:32:04 ip-10-31-40-15.us-east-1.aws.redhat.com dnsmasq[49052]: using nameserver 10.0.2.3#53 Aug 24 12:32:04 ip-10-31-40-15.us-east-1.aws.redhat.com dnsmasq[49052]: using nameserver 10.29.169.13#53 Aug 24 12:32:04 ip-10-31-40-15.us-east-1.aws.redhat.com dnsmasq[49052]: using nameserver 10.29.170.12#53 Aug 24 12:32:04 ip-10-31-40-15.us-east-1.aws.redhat.com dnsmasq[49052]: using nameserver 10.2.32.1#53 Aug 24 12:32:04 ip-10-31-40-15.us-east-1.aws.redhat.com dnsmasq[49052]: read /run/user/2001/containers/cni/dnsname/podman-default-kube-network/addnhosts - 1 addresses Aug 24 12:32:05 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[48906]: pam_unix(sudo:session): session closed for user auth_test_user1 Aug 24 12:32:05 ip-10-31-40-15.us-east-1.aws.redhat.com conmon[49128]: conmon d83f34166aafd6cd4a14 : Failed to get working directory Aug 24 12:32:05 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: cni-podman2: port 1(vethd69b4c65) entered disabled state Aug 24 12:32:05 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: device vethd69b4c65 left promiscuous mode Aug 24 12:32:05 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: cni-podman2: port 1(vethd69b4c65) entered disabled state Aug 24 12:32:05 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[49321]: root : TTY=pts/0 ; PWD=/root ; USER=auth_test_user1 ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-ezgtectxydrzklmrvefcwvynfynwneuk ; XDG_RUNTIME_DIR=/run/user/2001 /usr/libexec/platform-python /var/tmp/ansible-tmp-1724517125.1997957-20990-248056207770319/AnsiballZ_systemd.py' Aug 24 12:32:05 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[49321]: pam_unix(sudo:session): session opened for user auth_test_user1 by root(uid=0) Aug 24 12:32:05 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[49324]: ansible-systemd Invoked with daemon_reload=True scope=user daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None user=None Aug 24 12:32:05 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Reloading. Aug 24 12:32:05 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[49321]: pam_unix(sudo:session): session closed for user auth_test_user1 Aug 24 12:32:05 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[49458]: root : TTY=pts/0 ; PWD=/root ; USER=auth_test_user1 ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-vqjixfcjzxxcureisbskictotkbcqgsl ; XDG_RUNTIME_DIR=/run/user/2001 /usr/libexec/platform-python /var/tmp/ansible-tmp-1724517125.7649283-21020-115307811609571/AnsiballZ_systemd.py' Aug 24 12:32:05 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[49458]: pam_unix(sudo:session): session opened for user auth_test_user1 by root(uid=0) Aug 24 12:32:06 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[49461]: ansible-systemd Invoked with name= scope=user enabled=True daemon_reload=False daemon_reexec=False no_block=False state=None force=None masked=None user=None Aug 24 12:32:06 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Reloading. Aug 24 12:32:06 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[49458]: pam_unix(sudo:session): session closed for user auth_test_user1 Aug 24 12:32:06 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[49597]: root : TTY=pts/0 ; PWD=/root ; USER=auth_test_user1 ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-gyttobypdwnvbqfzidqzdzrfvaoifrzt ; XDG_RUNTIME_DIR=/run/user/2001 /usr/libexec/platform-python /var/tmp/ansible-tmp-1724517126.3745768-21029-141501193737783/AnsiballZ_systemd.py' Aug 24 12:32:06 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[49597]: pam_unix(sudo:session): session opened for user auth_test_user1 by root(uid=0) Aug 24 12:32:06 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[49600]: ansible-systemd Invoked with name= scope=user state=started daemon_reload=False daemon_reexec=False no_block=False enabled=None force=None masked=None user=None Aug 24 12:32:06 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Created slice podman\x2dkube.slice. -- Subject: Unit UNIT has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished starting up. -- -- The start-up result is done. Aug 24 12:32:06 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Starting A template for running K8s workloads via podman-kube-play... -- Subject: Unit UNIT has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has begun starting up. Aug 24 12:32:06 ip-10-31-40-15.us-east-1.aws.redhat.com podman[49606]: Pods stopped: Aug 24 12:32:06 ip-10-31-40-15.us-east-1.aws.redhat.com podman[49606]: a8fd13030011bcabb65b997412ad0a15d6982f0a908bdee2d8266896c3183b71 Aug 24 12:32:06 ip-10-31-40-15.us-east-1.aws.redhat.com podman[49606]: Pods removed: Aug 24 12:32:06 ip-10-31-40-15.us-east-1.aws.redhat.com podman[49606]: a8fd13030011bcabb65b997412ad0a15d6982f0a908bdee2d8266896c3183b71 Aug 24 12:32:06 ip-10-31-40-15.us-east-1.aws.redhat.com podman[49606]: Secrets removed: Aug 24 12:32:06 ip-10-31-40-15.us-east-1.aws.redhat.com podman[49606]: Volumes removed: Aug 24 12:32:07 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Started rootless-netns-f870481d.scope. -- Subject: Unit UNIT has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished starting up. -- -- The start-up result is done. Aug 24 12:32:07 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: IPv6: ADDRCONF(NETDEV_UP): veth30c7d91e: link is not ready Aug 24 12:32:07 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: cni-podman2: port 1(veth30c7d91e) entered blocking state Aug 24 12:32:07 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: cni-podman2: port 1(veth30c7d91e) entered disabled state Aug 24 12:32:07 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: device veth30c7d91e entered promiscuous mode Aug 24 12:32:07 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: IPv6: ADDRCONF(NETDEV_UP): eth0: link is not ready Aug 24 12:32:07 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: IPv6: ADDRCONF(NETDEV_CHANGE): eth0: link becomes ready Aug 24 12:32:07 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: IPv6: ADDRCONF(NETDEV_CHANGE): veth30c7d91e: link becomes ready Aug 24 12:32:07 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: cni-podman2: port 1(veth30c7d91e) entered blocking state Aug 24 12:32:07 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: cni-podman2: port 1(veth30c7d91e) entered forwarding state Aug 24 12:32:07 ip-10-31-40-15.us-east-1.aws.redhat.com dnsmasq[49727]: listening on cni-podman2(#3): 10.89.1.1 Aug 24 12:32:07 ip-10-31-40-15.us-east-1.aws.redhat.com dnsmasq[49729]: started, version 2.79 cachesize 150 Aug 24 12:32:07 ip-10-31-40-15.us-east-1.aws.redhat.com dnsmasq[49729]: compile time options: IPv6 GNU-getopt DBus no-i18n IDN2 DHCP DHCPv6 no-Lua TFTP no-conntrack ipset auth DNSSEC loop-detect inotify Aug 24 12:32:07 ip-10-31-40-15.us-east-1.aws.redhat.com dnsmasq[49729]: using local addresses only for domain dns.podman Aug 24 12:32:07 ip-10-31-40-15.us-east-1.aws.redhat.com dnsmasq[49729]: reading /etc/resolv.conf Aug 24 12:32:07 ip-10-31-40-15.us-east-1.aws.redhat.com dnsmasq[49729]: using local addresses only for domain dns.podman Aug 24 12:32:07 ip-10-31-40-15.us-east-1.aws.redhat.com dnsmasq[49729]: using nameserver 10.0.2.3#53 Aug 24 12:32:07 ip-10-31-40-15.us-east-1.aws.redhat.com dnsmasq[49729]: using nameserver 10.29.169.13#53 Aug 24 12:32:07 ip-10-31-40-15.us-east-1.aws.redhat.com dnsmasq[49729]: using nameserver 10.29.170.12#53 Aug 24 12:32:07 ip-10-31-40-15.us-east-1.aws.redhat.com dnsmasq[49729]: using nameserver 10.2.32.1#53 Aug 24 12:32:07 ip-10-31-40-15.us-east-1.aws.redhat.com dnsmasq[49729]: read /run/user/2001/containers/cni/dnsname/podman-default-kube-network/addnhosts - 1 addresses Aug 24 12:32:07 ip-10-31-40-15.us-east-1.aws.redhat.com podman[49606]: Pod: Aug 24 12:32:07 ip-10-31-40-15.us-east-1.aws.redhat.com podman[49606]: 297f77339fed772a57c23ceaa620fab6c5fb2d8c698c32ef1aa2c3b2467c30e1 Aug 24 12:32:07 ip-10-31-40-15.us-east-1.aws.redhat.com podman[49606]: Container: Aug 24 12:32:07 ip-10-31-40-15.us-east-1.aws.redhat.com podman[49606]: 3a4bdf2de00a30d3fdd9c342e9c24cef103280a69c6a44caaf41f265a3c3073b Aug 24 12:32:07 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Started A template for running K8s workloads via podman-kube-play. -- Subject: Unit UNIT has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished starting up. -- -- The start-up result is done. Aug 24 12:32:07 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[49597]: pam_unix(sudo:session): session closed for user auth_test_user1 Aug 24 12:32:08 ip-10-31-40-15.us-east-1.aws.redhat.com dnsmasq[49729]: listening on cni-podman2(#3): fe80::48be:c4ff:feb2:ce88%cni-podman2 Aug 24 12:32:08 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[50119]: ansible-getent Invoked with database=group key=2001 fail_key=False service=None split=None Aug 24 12:32:09 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[50330]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Aug 24 12:32:09 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[50539]: ansible-command Invoked with _raw_params=getsubids auth_test_user1 warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:32:10 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[50754]: ansible-command Invoked with _raw_params=getsubids -g auth_test_user1 warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:32:11 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[51138]: ansible-command Invoked with creates=/var/lib/systemd/linger/auth_test_user1 _raw_params=loginctl enable-linger auth_test_user1 warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None removes=None stdin=None Aug 24 12:32:12 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[51469]: root : TTY=pts/0 ; PWD=/root ; USER=auth_test_user1 ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-olbrlxtogvmvtfwqowhzmwejfuiazfub ; XDG_RUNTIME_DIR=/run/user/2001 /usr/libexec/platform-python /var/tmp/ansible-tmp-1724517132.3173523-21358-60429077144067/AnsiballZ_podman_image.py' Aug 24 12:32:12 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[51469]: pam_unix(sudo:session): session opened for user auth_test_user1 by root(uid=0) Aug 24 12:32:13 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Started podman-51497.scope. -- Subject: Unit UNIT has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished starting up. -- -- The start-up result is done. Aug 24 12:32:13 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Started podman-51506.scope. -- Subject: Unit UNIT has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished starting up. -- -- The start-up result is done. Aug 24 12:32:13 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Started podman-51525.scope. -- Subject: Unit UNIT has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished starting up. -- -- The start-up result is done. Aug 24 12:32:13 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Started podman-51570.scope. -- Subject: Unit UNIT has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished starting up. -- -- The start-up result is done. Aug 24 12:32:13 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[51469]: pam_unix(sudo:session): session closed for user auth_test_user1 Aug 24 12:32:13 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[51763]: ansible-file Invoked with path=/home/auth_test_user1/.config/containers/systemd state=directory owner=auth_test_user1 group=2001 mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Aug 24 12:32:14 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[52004]: ansible-stat Invoked with path=/home/auth_test_user1/.config/containers/systemd/auth_test_1_quadlet.container follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Aug 24 12:32:15 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[52189]: ansible-copy Invoked with src=/root/.ansible/tmp/ansible-tmp-1724517134.1960437-21444-42694398646366/source dest=/home/auth_test_user1/.config/containers/systemd/auth_test_1_quadlet.container owner=auth_test_user1 group=2001 mode=0644 follow=False _original_basename=systemd.j2 checksum=1127036af0ba22b83d33a5a35ab54db2c5855175 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None regexp=None delimiter=None Aug 24 12:32:16 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[52595]: ansible-getent Invoked with database=group key=2001 fail_key=False service=None split=None Aug 24 12:32:17 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[52790]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Aug 24 12:32:17 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[53020]: ansible-command Invoked with _raw_params=getsubids auth_test_user1 warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:32:18 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[53234]: ansible-command Invoked with _raw_params=getsubids -g auth_test_user1 warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:32:21 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[53863]: ansible-getent Invoked with database=group key=2001 fail_key=False service=None split=None Aug 24 12:32:21 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[54092]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Aug 24 12:32:22 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[54357]: ansible-command Invoked with _raw_params=getsubids auth_test_user1 warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:32:23 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[54587]: ansible-command Invoked with _raw_params=getsubids -g auth_test_user1 warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:32:27 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[55567]: ansible-command Invoked with _raw_params=podman --version warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:32:28 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[55823]: ansible-getent Invoked with database=passwd key=root fail_key=False service=None split=None Aug 24 12:32:28 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[56046]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None Aug 24 12:32:29 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[56276]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Aug 24 12:32:32 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[56917]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None Aug 24 12:32:32 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[57122]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Aug 24 12:32:34 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[57546]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None Aug 24 12:32:34 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[57757]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Aug 24 12:32:36 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[58223]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None Aug 24 12:32:37 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[58431]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Aug 24 12:32:38 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[58768]: ansible-command Invoked with _raw_params= warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:32:39 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[59053]: ansible-systemd Invoked with name= scope=system state=stopped enabled=False daemon_reload=False daemon_reexec=False no_block=False force=None masked=None user=None Aug 24 12:32:40 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[59285]: ansible-stat Invoked with path=/etc/containers/ansible-kubernetes.d/auth_test_1_kube.yml follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Aug 24 12:32:40 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[59492]: ansible-file Invoked with path=/etc/containers/ansible-kubernetes.d/auth_test_1_kube.yml state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Aug 24 12:32:42 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[59929]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None Aug 24 12:32:42 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[60152]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Aug 24 12:32:44 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[60573]: ansible-systemd Invoked with name=auth_test_1_quadlet.service scope=system state=stopped enabled=False force=True daemon_reload=False daemon_reexec=False no_block=False masked=None user=None Aug 24 12:32:45 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[60796]: ansible-stat Invoked with path=/etc/containers/systemd/auth_test_1_quadlet.container follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Aug 24 12:32:45 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[61014]: ansible-file Invoked with path=/etc/containers/systemd/auth_test_1_quadlet.container state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Aug 24 12:32:48 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[61558]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None Aug 24 12:32:48 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[61787]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Aug 24 12:32:50 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[62514]: ansible-file Invoked with path=/root/.config/containers state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Aug 24 12:32:52 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[62856]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None Aug 24 12:32:52 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[63074]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Aug 24 12:33:00 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[65152]: ansible-command Invoked with _raw_params=podman --version warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:33:01 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[65407]: ansible-getent Invoked with database=passwd key=auth_test_user1 fail_key=False service=None split=None Aug 24 12:33:02 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[65666]: ansible-getent Invoked with database=group key=2001 fail_key=False service=None split=None Aug 24 12:33:02 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[65900]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Aug 24 12:33:03 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[66109]: ansible-command Invoked with _raw_params=getsubids auth_test_user1 warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:33:03 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[66319]: ansible-command Invoked with _raw_params=getsubids -g auth_test_user1 warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:33:06 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[67038]: ansible-getent Invoked with database=group key=2001 fail_key=False service=None split=None Aug 24 12:33:07 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[67267]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Aug 24 12:33:08 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[67481]: ansible-command Invoked with _raw_params=getsubids auth_test_user1 warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:33:08 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[67691]: ansible-command Invoked with _raw_params=getsubids -g auth_test_user1 warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:33:10 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[68173]: ansible-getent Invoked with database=group key=2001 fail_key=False service=None split=None Aug 24 12:33:11 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[68399]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Aug 24 12:33:11 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[68613]: ansible-command Invoked with _raw_params=getsubids auth_test_user1 warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:33:12 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[68824]: ansible-command Invoked with _raw_params=getsubids -g auth_test_user1 warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:33:14 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[69299]: ansible-getent Invoked with database=group key=2001 fail_key=False service=None split=None Aug 24 12:33:14 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[69507]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Aug 24 12:33:15 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[69743]: ansible-command Invoked with _raw_params=getsubids auth_test_user1 warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:33:16 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[69954]: ansible-command Invoked with _raw_params=getsubids -g auth_test_user1 warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:33:17 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[70238]: ansible-command Invoked with _raw_params= warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:33:17 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[70531]: ansible-stat Invoked with path=/run/user/2001 follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Aug 24 12:33:18 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[70763]: root : TTY=pts/0 ; PWD=/root ; USER=auth_test_user1 ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-hpurnidubzilbykhwwpidsvjgctdbqzs ; XDG_RUNTIME_DIR=/run/user/2001 /usr/libexec/platform-python /var/tmp/ansible-tmp-1724517198.2391145-25226-139773372393366/AnsiballZ_systemd.py' Aug 24 12:33:18 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[70763]: pam_unix(sudo:session): session opened for user auth_test_user1 by root(uid=0) Aug 24 12:33:18 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[70766]: ansible-systemd Invoked with name= scope=user state=stopped enabled=False daemon_reload=False daemon_reexec=False no_block=False force=None masked=None user=None Aug 24 12:33:18 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Reloading. Aug 24 12:33:18 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Stopping A template for running K8s workloads via podman-kube-play... -- Subject: Unit UNIT has begun shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has begun shutting down. Aug 24 12:33:19 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: cni-podman2: port 1(veth30c7d91e) entered disabled state Aug 24 12:33:19 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: device veth30c7d91e left promiscuous mode Aug 24 12:33:19 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: cni-podman2: port 1(veth30c7d91e) entered disabled state Aug 24 12:33:19 ip-10-31-40-15.us-east-1.aws.redhat.com podman[70832]: Pods stopped: Aug 24 12:33:19 ip-10-31-40-15.us-east-1.aws.redhat.com podman[70832]: 297f77339fed772a57c23ceaa620fab6c5fb2d8c698c32ef1aa2c3b2467c30e1 Aug 24 12:33:19 ip-10-31-40-15.us-east-1.aws.redhat.com podman[70832]: Pods removed: Aug 24 12:33:19 ip-10-31-40-15.us-east-1.aws.redhat.com podman[70832]: 297f77339fed772a57c23ceaa620fab6c5fb2d8c698c32ef1aa2c3b2467c30e1 Aug 24 12:33:19 ip-10-31-40-15.us-east-1.aws.redhat.com podman[70832]: Secrets removed: Aug 24 12:33:19 ip-10-31-40-15.us-east-1.aws.redhat.com podman[70832]: Volumes removed: Aug 24 12:33:19 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Stopped A template for running K8s workloads via podman-kube-play. -- Subject: Unit UNIT has finished shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished shutting down. Aug 24 12:33:19 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[70763]: pam_unix(sudo:session): session closed for user auth_test_user1 Aug 24 12:33:19 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[71095]: ansible-stat Invoked with path=/home/auth_test_user1/.config/containers/ansible-kubernetes.d/auth_test_1_kube.yml follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Aug 24 12:33:20 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[71220]: root : TTY=pts/0 ; PWD=/root ; USER=auth_test_user1 ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-nrmnvxwsvdeezidilmvvzjftxotbozoo ; XDG_RUNTIME_DIR=/run/user/2001 /usr/libexec/platform-python /var/tmp/ansible-tmp-1724517200.0476031-25340-149551025494898/AnsiballZ_podman_play.py' Aug 24 12:33:20 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[71220]: pam_unix(sudo:session): session opened for user auth_test_user1 by root(uid=0) Aug 24 12:33:20 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[71223]: ansible-containers.podman.podman_play Invoked with state=absent kube_file=/home/auth_test_user1/.config/containers/ansible-kubernetes.d/auth_test_1_kube.yml executable=podman annotation=None authfile=None build=None cert_dir=None configmap=None context_dir=None seccomp_profile_root=None username=None password=NOT_LOGGING_PARAMETER log_driver=None log_opt=None network=None tls_verify=None debug=None quiet=None recreate=None userns=None log_level=None quadlet_dir=None quadlet_filename=None quadlet_options=None Aug 24 12:33:20 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[71223]: ansible-containers.podman.podman_play version: 4.9.4-dev, kube file /home/auth_test_user1/.config/containers/ansible-kubernetes.d/auth_test_1_kube.yml Aug 24 12:33:20 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Started podman-71231.scope. -- Subject: Unit UNIT has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished starting up. -- -- The start-up result is done. Aug 24 12:33:20 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[71220]: pam_unix(sudo:session): session closed for user auth_test_user1 Aug 24 12:33:21 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[71360]: ansible-file Invoked with path=/home/auth_test_user1/.config/containers/ansible-kubernetes.d/auth_test_1_kube.yml state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Aug 24 12:33:21 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[71483]: root : TTY=pts/0 ; PWD=/root ; USER=auth_test_user1 ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-komntehjdpgyonyrnjsqzcpuefesrwto ; XDG_RUNTIME_DIR=/run/user/2001 /usr/libexec/platform-python /var/tmp/ansible-tmp-1724517201.3468843-25419-53322898559826/AnsiballZ_command.py' Aug 24 12:33:21 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[71483]: pam_unix(sudo:session): session opened for user auth_test_user1 by root(uid=0) Aug 24 12:33:21 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[71486]: ansible-command Invoked with _raw_params=podman image prune -f warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:33:21 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Started podman-71488.scope. -- Subject: Unit UNIT has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished starting up. -- -- The start-up result is done. Aug 24 12:33:21 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[71483]: pam_unix(sudo:session): session closed for user auth_test_user1 Aug 24 12:33:23 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[71617]: ansible-getent Invoked with database=group key=2001 fail_key=False service=None split=None Aug 24 12:33:24 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[71741]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Aug 24 12:33:24 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[71866]: ansible-command Invoked with _raw_params=getsubids auth_test_user1 warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:33:24 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[71990]: ansible-command Invoked with _raw_params=getsubids -g auth_test_user1 warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:33:26 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[72114]: ansible-stat Invoked with path=/run/user/2001 follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Aug 24 12:33:26 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[72239]: root : TTY=pts/0 ; PWD=/root ; USER=auth_test_user1 ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-vlnruqoerlngvryvqackcpgmuhtwqhbn ; XDG_RUNTIME_DIR=/run/user/2001 /usr/libexec/platform-python /var/tmp/ansible-tmp-1724517206.5066495-25749-97258820368346/AnsiballZ_systemd.py' Aug 24 12:33:26 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[72239]: pam_unix(sudo:session): session opened for user auth_test_user1 by root(uid=0) Aug 24 12:33:26 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[72242]: ansible-systemd Invoked with name=auth_test_1_quadlet.service scope=user state=stopped enabled=False force=True daemon_reload=False daemon_reexec=False no_block=False masked=None user=None Aug 24 12:33:27 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Reloading. Aug 24 12:33:27 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[72239]: pam_unix(sudo:session): session closed for user auth_test_user1 Aug 24 12:33:27 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[72378]: ansible-stat Invoked with path=/home/auth_test_user1/.config/containers/systemd/auth_test_1_quadlet.container follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Aug 24 12:33:28 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[72626]: ansible-file Invoked with path=/home/auth_test_user1/.config/containers/systemd/auth_test_1_quadlet.container state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Aug 24 12:33:28 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[72749]: root : TTY=pts/0 ; PWD=/root ; USER=auth_test_user1 ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-acokvsmjelcatbnfbypmoricsgnapgxd ; XDG_RUNTIME_DIR=/run/user/2001 /usr/libexec/platform-python /var/tmp/ansible-tmp-1724517208.7611573-25879-195251231871213/AnsiballZ_systemd.py' Aug 24 12:33:28 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[72749]: pam_unix(sudo:session): session opened for user auth_test_user1 by root(uid=0) Aug 24 12:33:29 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[72752]: ansible-systemd Invoked with daemon_reload=True scope=user daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None user=None Aug 24 12:33:29 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Reloading. Aug 24 12:33:29 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[72749]: pam_unix(sudo:session): session closed for user auth_test_user1 Aug 24 12:33:29 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[72886]: root : TTY=pts/0 ; PWD=/root ; USER=auth_test_user1 ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-ckhmtkhqnlzafleknmfufbzovnrishtl ; XDG_RUNTIME_DIR=/run/user/2001 /usr/libexec/platform-python /var/tmp/ansible-tmp-1724517209.3543239-25915-271137461694598/AnsiballZ_command.py' Aug 24 12:33:29 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[72886]: pam_unix(sudo:session): session opened for user auth_test_user1 by root(uid=0) Aug 24 12:33:29 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Started podman-72891.scope. -- Subject: Unit UNIT has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished starting up. -- -- The start-up result is done. Aug 24 12:33:29 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[72886]: pam_unix(sudo:session): session closed for user auth_test_user1 Aug 24 12:33:31 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[73020]: ansible-stat Invoked with path=/run/user/2001 follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Aug 24 12:33:31 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[73145]: root : TTY=pts/0 ; PWD=/root ; USER=auth_test_user1 ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-ocdeozodtlruaphkgintpmlufvdoozeh ; XDG_RUNTIME_DIR=/run/user/2001 /usr/libexec/platform-python /var/tmp/ansible-tmp-1724517211.6403909-26040-73967260498135/AnsiballZ_podman_container_info.py' Aug 24 12:33:31 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[73145]: pam_unix(sudo:session): session opened for user auth_test_user1 by root(uid=0) Aug 24 12:33:32 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[73148]: ansible-containers.podman.podman_container_info Invoked with executable=podman name=None Aug 24 12:33:32 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Started podman-73150.scope. -- Subject: Unit UNIT has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished starting up. -- -- The start-up result is done. Aug 24 12:33:32 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[73145]: pam_unix(sudo:session): session closed for user auth_test_user1 Aug 24 12:33:32 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[73279]: root : TTY=pts/0 ; PWD=/root ; USER=auth_test_user1 ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-lswpsxdmfeeknwcgchdovupwlnrcgqve ; XDG_RUNTIME_DIR=/run/user/2001 /usr/libexec/platform-python /var/tmp/ansible-tmp-1724517212.2750049-26080-100716737600628/AnsiballZ_command.py' Aug 24 12:33:32 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[73279]: pam_unix(sudo:session): session opened for user auth_test_user1 by root(uid=0) Aug 24 12:33:32 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[73282]: ansible-command Invoked with _raw_params=podman network ls -q warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:33:32 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Started podman-73284.scope. -- Subject: Unit UNIT has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished starting up. -- -- The start-up result is done. Aug 24 12:33:32 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[73279]: pam_unix(sudo:session): session closed for user auth_test_user1 Aug 24 12:33:32 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[73438]: root : TTY=pts/0 ; PWD=/root ; USER=auth_test_user1 ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-khzsauhujhmmwmbstbwuicigahvopfvy ; XDG_RUNTIME_DIR=/run/user/2001 /usr/libexec/platform-python /var/tmp/ansible-tmp-1724517212.7601972-26109-106997805110866/AnsiballZ_command.py' Aug 24 12:33:32 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[73438]: pam_unix(sudo:session): session opened for user auth_test_user1 by root(uid=0) Aug 24 12:33:33 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[73441]: ansible-command Invoked with _raw_params=podman secret ls -n -q warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:33:33 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Started podman-73443.scope. -- Subject: Unit UNIT has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished starting up. -- -- The start-up result is done. Aug 24 12:33:33 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[73438]: pam_unix(sudo:session): session closed for user auth_test_user1 Aug 24 12:33:33 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[73572]: ansible-command Invoked with removes=/var/lib/systemd/linger/auth_test_user1 _raw_params=loginctl disable-linger auth_test_user1 warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None stdin=None Aug 24 12:33:33 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Stopping User Manager for UID 2001... -- Subject: Unit user@2001.service has begun shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit user@2001.service has begun shutting down. Aug 24 12:33:33 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Removed slice podman\x2dkube.slice. -- Subject: Unit UNIT has finished shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished shutting down. Aug 24 12:33:33 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Stopped target Default. -- Subject: Unit UNIT has finished shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished shutting down. Aug 24 12:33:33 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Stopping podman-pause-0e20a755.scope. -- Subject: Unit UNIT has begun shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has begun shutting down. Aug 24 12:33:33 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Stopping D-Bus User Message Bus... -- Subject: Unit UNIT has begun shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has begun shutting down. Aug 24 12:33:33 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Stopped D-Bus User Message Bus. -- Subject: Unit UNIT has finished shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished shutting down. Aug 24 12:33:33 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Stopped target Basic System. -- Subject: Unit UNIT has finished shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished shutting down. Aug 24 12:33:33 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Stopped target Paths. -- Subject: Unit UNIT has finished shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished shutting down. Aug 24 12:33:33 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Stopped target Sockets. -- Subject: Unit UNIT has finished shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished shutting down. Aug 24 12:33:33 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Stopped target Timers. -- Subject: Unit UNIT has finished shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished shutting down. Aug 24 12:33:33 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Stopped Mark boot as successful after the user session has run 2 minutes. -- Subject: Unit UNIT has finished shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished shutting down. Aug 24 12:33:33 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Closed D-Bus User Message Bus Socket. -- Subject: Unit UNIT has finished shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished shutting down. Aug 24 12:33:33 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Stopped podman-pause-0e20a755.scope. -- Subject: Unit UNIT has finished shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished shutting down. Aug 24 12:33:33 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Removed slice user.slice. -- Subject: Unit UNIT has finished shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished shutting down. Aug 24 12:33:33 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Reached target Shutdown. -- Subject: Unit UNIT has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished starting up. -- -- The start-up result is done. Aug 24 12:33:33 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Started Exit the Session. -- Subject: Unit UNIT has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished starting up. -- -- The start-up result is done. Aug 24 12:33:33 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Reached target Exit the Session. -- Subject: Unit UNIT has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished starting up. -- -- The start-up result is done. Aug 24 12:33:33 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48123]: pam_unix(systemd-user:session): session closed for user auth_test_user1 Aug 24 12:33:33 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: user@2001.service: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit user@2001.service has successfully entered the 'dead' state. Aug 24 12:33:33 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Stopped User Manager for UID 2001. -- Subject: Unit user@2001.service has finished shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit user@2001.service has finished shutting down. Aug 24 12:33:33 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Stopping User runtime directory /run/user/2001... -- Subject: Unit user-runtime-dir@2001.service has begun shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit user-runtime-dir@2001.service has begun shutting down. Aug 24 12:33:33 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: run-user-2001.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit run-user-2001.mount has successfully entered the 'dead' state. Aug 24 12:33:33 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: user-runtime-dir@2001.service: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit user-runtime-dir@2001.service has successfully entered the 'dead' state. Aug 24 12:33:33 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Stopped User runtime directory /run/user/2001. -- Subject: Unit user-runtime-dir@2001.service has finished shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit user-runtime-dir@2001.service has finished shutting down. Aug 24 12:33:33 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Removed slice User Slice of UID 2001. -- Subject: Unit user-2001.slice has finished shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit user-2001.slice has finished shutting down. Aug 24 12:33:34 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[73704]: ansible-command Invoked with _raw_params=loginctl show-user -P State auth_test_user1 warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:33:34 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[73828]: ansible-systemd Invoked with name=systemd-logind state=stopped daemon_reload=False daemon_reexec=False no_block=False enabled=None force=None masked=None user=None scope=None Aug 24 12:33:34 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Stopping Login Service... -- Subject: Unit systemd-logind.service has begun shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit systemd-logind.service has begun shutting down. Aug 24 12:33:34 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: systemd-logind.service: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit systemd-logind.service has successfully entered the 'dead' state. Aug 24 12:33:34 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Stopped Login Service. -- Subject: Unit systemd-logind.service has finished shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit systemd-logind.service has finished shutting down. Aug 24 12:33:35 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[73957]: ansible-command Invoked with _raw_params=loginctl show-user -P State auth_test_user1 warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:33:36 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[74081]: ansible-getent Invoked with database=group key=2001 fail_key=False service=None split=None Aug 24 12:33:36 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[74205]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Aug 24 12:33:37 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[74330]: ansible-command Invoked with _raw_params=getsubids auth_test_user1 warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:33:37 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[74454]: ansible-command Invoked with _raw_params=getsubids -g auth_test_user1 warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:33:40 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[74824]: ansible-getent Invoked with database=group key=2001 fail_key=False service=None split=None Aug 24 12:33:41 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[74948]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Aug 24 12:33:41 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[75073]: ansible-command Invoked with _raw_params=getsubids auth_test_user1 warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:33:42 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[75197]: ansible-command Invoked with _raw_params=getsubids -g auth_test_user1 warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:33:46 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[75936]: ansible-user Invoked with name=auth_test_user1 state=absent non_unique=False force=False remove=False create_home=True system=False move_home=False append=False ssh_key_bits=0 ssh_key_type=rsa ssh_key_comment=ansible-generated on ip-10-31-40-15.us-east-1.aws.redhat.com update_password=always uid=None group=None groups=None comment=None home=None shell=None password=NOT_LOGGING_PARAMETER login_class=None hidden=None seuser=None skeleton=None generate_ssh_key=None ssh_key_file=None ssh_key_passphrase=NOT_LOGGING_PARAMETER expires=None password_lock=None local=None profile=None authorization=None role=None Aug 24 12:33:46 ip-10-31-40-15.us-east-1.aws.redhat.com userdel[75940]: delete user 'auth_test_user1' Aug 24 12:33:46 ip-10-31-40-15.us-east-1.aws.redhat.com userdel[75940]: removed group 'auth_test_user1' owned by 'auth_test_user1' Aug 24 12:33:46 ip-10-31-40-15.us-east-1.aws.redhat.com userdel[75940]: removed shadow group 'auth_test_user1' owned by 'auth_test_user1' Aug 24 12:33:46 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[76068]: ansible-file Invoked with path=/home/auth_test_user1 state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Aug 24 12:33:47 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[76191]: ansible-command Invoked with _raw_params=podman inspect podman_registry --format '{{range .}}{{range .Mounts}}{{if eq .Type "volume"}}{{.Name}}{{end}}{{end}}{{end}}' warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:33:48 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[76322]: ansible-command Invoked with _raw_params=podman rm -f podman_registry warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:33:48 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-bf8d6a8a130cd57044b05b59250727fd32420dde3068441219e120a629a7ae4d.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-bf8d6a8a130cd57044b05b59250727fd32420dde3068441219e120a629a7ae4d.scope has successfully entered the 'dead' state. Aug 24 12:33:48 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-bf8d6a8a130cd57044b05b59250727fd32420dde3068441219e120a629a7ae4d.scope: Consumed 319ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-bf8d6a8a130cd57044b05b59250727fd32420dde3068441219e120a629a7ae4d.scope completed and consumed the indicated resources. Aug 24 12:33:48 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: cni-podman0: port 1(veth85345cee) entered disabled state Aug 24 12:33:48 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: device veth85345cee left promiscuous mode Aug 24 12:33:48 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: cni-podman0: port 1(veth85345cee) entered disabled state Aug 24 12:33:48 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: run-netns-netns\x2dcf1675ac\x2d9007\x2dd2d5\x2d8d1a\x2da8931a022d46.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit run-netns-netns\x2dcf1675ac\x2d9007\x2dd2d5\x2d8d1a\x2da8931a022d46.mount has successfully entered the 'dead' state. Aug 24 12:33:48 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-bf8d6a8a130cd57044b05b59250727fd32420dde3068441219e120a629a7ae4d-userdata-shm.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay\x2dcontainers-bf8d6a8a130cd57044b05b59250727fd32420dde3068441219e120a629a7ae4d-userdata-shm.mount has successfully entered the 'dead' state. Aug 24 12:33:48 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay-77219d5a585fd51d0e3bdf98787258f43de7656281a74f5b8cd3e9918c22274b-merged.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay-77219d5a585fd51d0e3bdf98787258f43de7656281a74f5b8cd3e9918c22274b-merged.mount has successfully entered the 'dead' state. Aug 24 12:33:48 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Aug 24 12:33:48 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-conmon-bf8d6a8a130cd57044b05b59250727fd32420dde3068441219e120a629a7ae4d.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-conmon-bf8d6a8a130cd57044b05b59250727fd32420dde3068441219e120a629a7ae4d.scope has successfully entered the 'dead' state. Aug 24 12:33:49 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[76553]: ansible-command Invoked with _raw_params=podman volume rm a24eff084e57a603a42be183d8da720f785cafeabd9b1200393e3034f9f518d6 warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:33:49 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[76683]: ansible-file Invoked with path=/tmp/lsr_iftawaex_podman state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Aug 24 12:33:52 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[76845]: ansible-setup Invoked with gather_subset=['all'] gather_timeout=10 filter=* fact_path=/etc/ansible/facts.d Aug 24 12:33:53 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[77001]: ansible-stat Invoked with path=/run/ostree-booted follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Aug 24 12:33:54 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[77124]: ansible-dnf Invoked with name=['python3-pyasn1', 'python3-cryptography', 'python3-dbus'] state=present allow_downgrade=False autoremove=False bugfix=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True lock_timeout=30 conf_file=None disable_excludes=None download_dir=None list=None releasever=None Aug 24 12:33:57 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[77247]: ansible-dnf Invoked with name=['certmonger'] state=present allow_downgrade=False autoremove=False bugfix=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True lock_timeout=30 conf_file=None disable_excludes=None download_dir=None list=None releasever=None Aug 24 12:33:59 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[77370]: ansible-file Invoked with name=/etc/certmonger//pre-scripts owner=root group=root mode=0700 state=directory path=/etc/certmonger//pre-scripts recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Aug 24 12:34:00 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[77493]: ansible-file Invoked with name=/etc/certmonger//post-scripts owner=root group=root mode=0700 state=directory path=/etc/certmonger//post-scripts recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Aug 24 12:34:01 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[77616]: ansible-systemd Invoked with name=certmonger state=started enabled=True daemon_reload=False daemon_reexec=False no_block=False force=None masked=None user=None scope=None Aug 24 12:34:02 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[77743]: ansible-fedora.linux_system_roles.certificate_request Invoked with name=quadlet_demo dns=['localhost'] directory=/etc/pki/tls wait=True ca=self-sign __header=# # Ansible managed # # system_role:certificate provider_config_directory=/etc/certmonger provider=certmonger key_usage=['digitalSignature', 'keyEncipherment'] extended_key_usage=['id-kp-serverAuth', 'id-kp-clientAuth'] auto_renew=True ip=None email=None common_name=None country=None state=None locality=None organization=None organizational_unit=None contact_email=None key_size=None owner=None group=None mode=None principal=None run_before=None run_after=None Aug 24 12:34:02 ip-10-31-40-15.us-east-1.aws.redhat.com certmonger[12107]: 2024-08-24 12:34:02 [12107] Wrote to /var/lib/certmonger/requests/20240824163402 Aug 24 12:34:02 ip-10-31-40-15.us-east-1.aws.redhat.com certmonger[12107]: 2024-08-24 12:34:02 [12107] Wrote to /var/lib/certmonger/requests/20240824163402 Aug 24 12:34:02 ip-10-31-40-15.us-east-1.aws.redhat.com certmonger[12107]: 2024-08-24 12:34:02 [12107] Wrote to /var/lib/certmonger/requests/20240824163402 Aug 24 12:34:02 ip-10-31-40-15.us-east-1.aws.redhat.com certmonger[12107]: 2024-08-24 12:34:02 [12107] Wrote to /var/lib/certmonger/requests/20240824163402 Aug 24 12:34:02 ip-10-31-40-15.us-east-1.aws.redhat.com certmonger[12107]: 2024-08-24 12:34:02 [12107] Wrote to /var/lib/certmonger/requests/20240824163402 Aug 24 12:34:02 ip-10-31-40-15.us-east-1.aws.redhat.com certmonger[12107]: 2024-08-24 12:34:02 [12107] Wrote to /var/lib/certmonger/requests/20240824163402 Aug 24 12:34:02 ip-10-31-40-15.us-east-1.aws.redhat.com certmonger[12107]: 2024-08-24 12:34:02 [12107] Wrote to /var/lib/certmonger/requests/20240824163402 Aug 24 12:34:02 ip-10-31-40-15.us-east-1.aws.redhat.com certmonger[12107]: 2024-08-24 12:34:02 [12107] Wrote to /var/lib/certmonger/requests/20240824163402 Aug 24 12:34:02 ip-10-31-40-15.us-east-1.aws.redhat.com certmonger[12107]: 2024-08-24 12:34:02 [12107] Wrote to /var/lib/certmonger/requests/20240824163402 Aug 24 12:34:02 ip-10-31-40-15.us-east-1.aws.redhat.com certmonger[12107]: 2024-08-24 12:34:02 [12107] Wrote to /var/lib/certmonger/requests/20240824163402 Aug 24 12:34:02 ip-10-31-40-15.us-east-1.aws.redhat.com certmonger[12107]: 2024-08-24 12:34:02 [12107] Wrote to /var/lib/certmonger/requests/20240824163402 Aug 24 12:34:02 ip-10-31-40-15.us-east-1.aws.redhat.com certmonger[12107]: 2024-08-24 12:34:02 [12107] Wrote to /var/lib/certmonger/requests/20240824163402 Aug 24 12:34:02 ip-10-31-40-15.us-east-1.aws.redhat.com certmonger[12107]: 2024-08-24 12:34:02 [12107] Wrote to /var/lib/certmonger/requests/20240824163402 Aug 24 12:34:02 ip-10-31-40-15.us-east-1.aws.redhat.com certmonger[12107]: 2024-08-24 12:34:02 [12107] Wrote to /var/lib/certmonger/requests/20240824163402 Aug 24 12:34:02 ip-10-31-40-15.us-east-1.aws.redhat.com certmonger[12107]: 2024-08-24 12:34:02 [12107] Wrote to /var/lib/certmonger/requests/20240824163402 Aug 24 12:34:02 ip-10-31-40-15.us-east-1.aws.redhat.com certmonger[12107]: 2024-08-24 12:34:02 [12107] Wrote to /var/lib/certmonger/requests/20240824163402 Aug 24 12:34:02 ip-10-31-40-15.us-east-1.aws.redhat.com certmonger[12107]: 2024-08-24 12:34:02 [12107] Wrote to /var/lib/certmonger/requests/20240824163402 Aug 24 12:34:02 ip-10-31-40-15.us-east-1.aws.redhat.com certmonger[12107]: 2024-08-24 12:34:02 [12107] Wrote to /var/lib/certmonger/requests/20240824163402 Aug 24 12:34:02 ip-10-31-40-15.us-east-1.aws.redhat.com certmonger[12107]: 2024-08-24 12:34:02 [12107] Wrote to /var/lib/certmonger/requests/20240824163402 Aug 24 12:34:02 ip-10-31-40-15.us-east-1.aws.redhat.com certmonger[12107]: 2024-08-24 12:34:02 [12107] Wrote to /var/lib/certmonger/requests/20240824163402 Aug 24 12:34:02 ip-10-31-40-15.us-east-1.aws.redhat.com certmonger[12107]: 2024-08-24 12:34:02 [12107] Wrote to /var/lib/certmonger/requests/20240824163402 Aug 24 12:34:02 ip-10-31-40-15.us-east-1.aws.redhat.com certmonger[12107]: 2024-08-24 12:34:02 [12107] Wrote to /var/lib/certmonger/requests/20240824163402 Aug 24 12:34:02 ip-10-31-40-15.us-east-1.aws.redhat.com certmonger[12107]: 2024-08-24 12:34:02 [12107] Wrote to /var/lib/certmonger/requests/20240824163402 Aug 24 12:34:02 ip-10-31-40-15.us-east-1.aws.redhat.com certmonger[12107]: 2024-08-24 12:34:02 [12107] Wrote to /var/lib/certmonger/requests/20240824163402 Aug 24 12:34:02 ip-10-31-40-15.us-east-1.aws.redhat.com certmonger[12107]: 2024-08-24 12:34:02 [12107] Wrote to /var/lib/certmonger/requests/20240824163402 Aug 24 12:34:02 ip-10-31-40-15.us-east-1.aws.redhat.com certmonger[12107]: 2024-08-24 12:34:02 [12107] Wrote to /var/lib/certmonger/requests/20240824163402 Aug 24 12:34:02 ip-10-31-40-15.us-east-1.aws.redhat.com certmonger[77759]: Certificate in file "/etc/pki/tls/certs/quadlet_demo.crt" issued by CA and saved. Aug 24 12:34:02 ip-10-31-40-15.us-east-1.aws.redhat.com certmonger[12107]: 2024-08-24 12:34:02 [12107] Wrote to /var/lib/certmonger/requests/20240824163402 Aug 24 12:34:03 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[77881]: ansible-slurp Invoked with path=/etc/pki/tls/certs/quadlet_demo.crt src=/etc/pki/tls/certs/quadlet_demo.crt Aug 24 12:34:03 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[78004]: ansible-slurp Invoked with path=/etc/pki/tls/private/quadlet_demo.key src=/etc/pki/tls/private/quadlet_demo.key Aug 24 12:34:04 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[78127]: ansible-slurp Invoked with path=/etc/pki/tls/certs/quadlet_demo.crt src=/etc/pki/tls/certs/quadlet_demo.crt Aug 24 12:34:04 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[78250]: ansible-command Invoked with _raw_params=getcert stop-tracking -f /etc/pki/tls/certs/quadlet_demo.crt warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:34:04 ip-10-31-40-15.us-east-1.aws.redhat.com certmonger[12107]: 2024-08-24 12:34:04 [12107] Wrote to /var/lib/certmonger/requests/20240824163402 Aug 24 12:34:05 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[78374]: ansible-file Invoked with path=/etc/pki/tls/certs/quadlet_demo.crt state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Aug 24 12:34:05 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[78497]: ansible-file Invoked with path=/etc/pki/tls/private/quadlet_demo.key state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Aug 24 12:34:06 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[78620]: ansible-file Invoked with path=/etc/pki/tls/certs/quadlet_demo.crt state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Aug 24 12:34:07 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[78743]: ansible-stat Invoked with path=/run/ostree-booted follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Aug 24 12:34:07 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[78866]: ansible-stat Invoked with path=/sbin/transactional-update follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Aug 24 12:34:10 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[79114]: ansible-command Invoked with _raw_params=podman --version warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:34:11 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[79243]: ansible-getent Invoked with database=passwd key=root fail_key=False service=None split=None Aug 24 12:34:12 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[79367]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None Aug 24 12:34:13 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[79491]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Aug 24 12:34:14 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[79616]: ansible-stat Invoked with path=/run/ostree-booted follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Aug 24 12:34:15 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[79739]: ansible-stat Invoked with path=/sbin/transactional-update follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Aug 24 12:34:16 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[79862]: ansible-dnf Invoked with name=['firewalld'] state=present allow_downgrade=False autoremove=False bugfix=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True lock_timeout=30 conf_file=None disable_excludes=None download_dir=None list=None releasever=None Aug 24 12:34:19 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[79985]: ansible-systemd Invoked with name=firewalld masked=False daemon_reload=False daemon_reexec=False no_block=False state=None enabled=None force=None user=None scope=None Aug 24 12:34:19 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[80112]: ansible-systemd Invoked with name=firewalld state=started enabled=True daemon_reload=False daemon_reexec=False no_block=False force=None masked=None user=None scope=None Aug 24 12:34:19 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Reloading. Aug 24 12:34:19 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Starting firewalld - dynamic firewall daemon... -- Subject: Unit firewalld.service has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit firewalld.service has begun starting up. Aug 24 12:34:20 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started firewalld - dynamic firewall daemon. -- Subject: Unit firewalld.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit firewalld.service has finished starting up. -- -- The start-up result is done. Aug 24 12:34:20 ip-10-31-40-15.us-east-1.aws.redhat.com firewalld[80148]: WARNING: AllowZoneDrifting is enabled. This is considered an insecure configuration option. It will be removed in a future release. Please consider disabling it now. Aug 24 12:34:21 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[80323]: ansible-fedora.linux_system_roles.firewall_lib Invoked with port=['8000/tcp'] permanent=True runtime=True state=enabled __report_changed=True service=[] source_port=[] forward_port=[] rich_rule=[] source=[] interface=[] interface_pci_id=[] icmp_block=[] timeout=0 ipset_entries=[] protocol=[] helper_module=[] destination=[] firewalld_conf=None masquerade=None icmp_block_inversion=None target=None zone=None set_default_zone=None ipset=None ipset_type=None description=None short=None Aug 24 12:34:21 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[80446]: ansible-fedora.linux_system_roles.firewall_lib Invoked with port=['9000/tcp'] permanent=True runtime=True state=enabled __report_changed=True service=[] source_port=[] forward_port=[] rich_rule=[] source=[] interface=[] interface_pci_id=[] icmp_block=[] timeout=0 ipset_entries=[] protocol=[] helper_module=[] destination=[] firewalld_conf=None masquerade=None icmp_block_inversion=None target=None zone=None set_default_zone=None ipset=None ipset_type=None description=None short=None Aug 24 12:34:21 ip-10-31-40-15.us-east-1.aws.redhat.com rsyslogd[1022]: imjournal: journal files changed, reloading... [v8.2102.0-15.el8 try https://www.rsyslog.com/e/0 ] Aug 24 12:34:23 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Aug 24 12:34:24 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Aug 24 12:34:25 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Aug 24 12:34:26 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[81023]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None Aug 24 12:34:26 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[81147]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Aug 24 12:34:28 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[81272]: ansible-file Invoked with path=/etc/containers/systemd state=directory owner=root group=0 mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Aug 24 12:34:28 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[81395]: ansible-stat Invoked with path=/etc/containers/systemd/quadlet-demo.network follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Aug 24 12:34:29 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[81494]: ansible-copy Invoked with src=/root/.ansible/tmp/ansible-tmp-1724517268.7169204-29276-193597116746162/source dest=/etc/containers/systemd/quadlet-demo.network owner=root group=0 mode=0644 _original_basename=quadlet-demo.network follow=False checksum=e57c08d49aff4bae8daab138d913aeddaa8682a0 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None regexp=None delimiter=None Aug 24 12:34:29 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[81619]: ansible-systemd Invoked with daemon_reload=True scope=system daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None user=None Aug 24 12:34:29 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Reloading. Aug 24 12:34:30 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[81772]: ansible-systemd Invoked with name=quadlet-demo-network.service scope=system state=started daemon_reload=False daemon_reexec=False no_block=False enabled=None force=None masked=None user=None Aug 24 12:34:30 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Starting quadlet-demo-network.service... -- Subject: Unit quadlet-demo-network.service has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit quadlet-demo-network.service has begun starting up. Aug 24 12:34:30 ip-10-31-40-15.us-east-1.aws.redhat.com quadlet-demo-network[81779]: systemd-quadlet-demo Aug 24 12:34:30 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Aug 24 12:34:30 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started quadlet-demo-network.service. -- Subject: Unit quadlet-demo-network.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit quadlet-demo-network.service has finished starting up. -- -- The start-up result is done. Aug 24 12:34:31 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[81933]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None Aug 24 12:34:32 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[82057]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Aug 24 12:34:33 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[82182]: ansible-file Invoked with path=/etc/containers/systemd state=directory owner=root group=0 mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Aug 24 12:34:34 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[82305]: ansible-stat Invoked with path=/etc/containers/systemd/quadlet-demo-mysql.volume follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Aug 24 12:34:34 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[82404]: ansible-copy Invoked with src=/root/.ansible/tmp/ansible-tmp-1724517273.9811473-29551-138782588482123/source dest=/etc/containers/systemd/quadlet-demo-mysql.volume owner=root group=0 mode=0644 _original_basename=quadlet-demo-mysql.volume follow=False checksum=585f8cbdf0ec73000f9227dcffbef71e9552ea4a backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None regexp=None delimiter=None Aug 24 12:34:35 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[82529]: ansible-systemd Invoked with daemon_reload=True scope=system daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None user=None Aug 24 12:34:35 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Reloading. Aug 24 12:34:35 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[82682]: ansible-systemd Invoked with name=quadlet-demo-mysql-volume.service scope=system state=started daemon_reload=False daemon_reexec=False no_block=False enabled=None force=None masked=None user=None Aug 24 12:34:35 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Starting quadlet-demo-mysql-volume.service... -- Subject: Unit quadlet-demo-mysql-volume.service has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit quadlet-demo-mysql-volume.service has begun starting up. Aug 24 12:34:35 ip-10-31-40-15.us-east-1.aws.redhat.com quadlet-demo-mysql-volume[82689]: systemd-quadlet-demo-mysql Aug 24 12:34:35 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Aug 24 12:34:35 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started quadlet-demo-mysql-volume.service. -- Subject: Unit quadlet-demo-mysql-volume.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit quadlet-demo-mysql-volume.service has finished starting up. -- -- The start-up result is done. Aug 24 12:34:36 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[82819]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None Aug 24 12:34:37 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[82943]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Aug 24 12:34:38 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[83068]: ansible-file Invoked with path=/tmp/quadlet_demo state=directory owner=root group=root mode=0777 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Aug 24 12:34:39 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Aug 24 12:34:46 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Aug 24 12:34:46 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Aug 24 12:34:46 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[83411]: ansible-file Invoked with path=/etc/containers/systemd state=directory owner=root group=0 mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Aug 24 12:34:46 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[83534]: ansible-stat Invoked with path=/etc/containers/systemd/quadlet-demo-mysql.container follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Aug 24 12:34:47 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[83633]: ansible-copy Invoked with dest=/etc/containers/systemd/quadlet-demo-mysql.container owner=root group=0 mode=0644 src=/root/.ansible/tmp/ansible-tmp-1724517286.7033756-30107-189472742671718/source _original_basename=tmpca1ro4lg follow=False checksum=ca62b2ad3cc9afb5b5371ebbf797b9bc4fd7edd4 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None regexp=None delimiter=None Aug 24 12:34:47 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[83758]: ansible-systemd Invoked with daemon_reload=True scope=system daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None user=None Aug 24 12:34:47 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Reloading. Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[83911]: ansible-systemd Invoked with name=quadlet-demo-mysql.service scope=system state=started daemon_reload=False daemon_reexec=False no_block=False enabled=None force=None masked=None user=None Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Starting quadlet-demo-mysql.service... -- Subject: Unit quadlet-demo-mysql.service has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit quadlet-demo-mysql.service has begun starting up. Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: IPv6: ADDRCONF(NETDEV_UP): veth10ec5328: link is not ready Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: cni-podman2: port 1(veth10ec5328) entered blocking state Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: cni-podman2: port 1(veth10ec5328) entered disabled state Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: device veth10ec5328 entered promiscuous mode Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: cni-podman2: port 1(veth10ec5328) entered blocking state Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: cni-podman2: port 1(veth10ec5328) entered forwarding state Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com systemd-udevd[83985]: Using default interface naming scheme 'rhel-8.0'. Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com systemd-udevd[83985]: link_config: autonegotiation is unset or enabled, the speed and duplex are not writable. Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com systemd-udevd[83989]: link_config: autonegotiation is unset or enabled, the speed and duplex are not writable. Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com systemd-udevd[83985]: Could not generate persistent MAC address for cni-podman2: No such file or directory Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com systemd-udevd[83989]: Could not generate persistent MAC address for veth10ec5328: No such file or directory Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com NetworkManager[660]: [1724517288.6462] manager: (cni-podman2): new Bridge device (/org/freedesktop/NetworkManager/Devices/8) Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com NetworkManager[660]: [1724517288.6473] manager: (veth10ec5328): new Veth device (/org/freedesktop/NetworkManager/Devices/9) Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: IPv6: ADDRCONF(NETDEV_CHANGE): veth10ec5328: link becomes ready Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com NetworkManager[660]: [1724517288.6545] device (veth10ec5328): carrier: link connected Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com NetworkManager[660]: [1724517288.6546] device (cni-podman2): carrier: link connected Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com NetworkManager[660]: [1724517288.6798] device (cni-podman2): state change: unmanaged -> unavailable (reason 'connection-assumed', sys-iface-state: 'external') Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com NetworkManager[660]: [1724517288.6805] device (cni-podman2): state change: unavailable -> disconnected (reason 'connection-assumed', sys-iface-state: 'external') Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com NetworkManager[660]: [1724517288.6813] device (cni-podman2): Activation: starting connection 'cni-podman2' (27cdadd2-7388-418d-a98e-53ae883e6b12) Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com NetworkManager[660]: [1724517288.6814] device (cni-podman2): state change: disconnected -> prepare (reason 'none', sys-iface-state: 'external') Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com NetworkManager[660]: [1724517288.6817] device (cni-podman2): state change: prepare -> config (reason 'none', sys-iface-state: 'external') Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com NetworkManager[660]: [1724517288.6818] device (cni-podman2): state change: config -> ip-config (reason 'none', sys-iface-state: 'external') Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com NetworkManager[660]: [1724517288.6820] device (cni-podman2): state change: ip-config -> ip-check (reason 'none', sys-iface-state: 'external') Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com dbus-daemon[590]: [system] Activating via systemd: service name='org.freedesktop.nm_dispatcher' unit='dbus-org.freedesktop.nm-dispatcher.service' requested by ':1.5' (uid=0 pid=660 comm="/usr/sbin/NetworkManager --no-daemon " label="system_u:system_r:NetworkManager_t:s0") Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Starting Network Manager Script Dispatcher Service... -- Subject: Unit NetworkManager-dispatcher.service has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit NetworkManager-dispatcher.service has begun starting up. Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com dbus-daemon[590]: [system] Successfully activated service 'org.freedesktop.nm_dispatcher' Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started Network Manager Script Dispatcher Service. -- Subject: Unit NetworkManager-dispatcher.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit NetworkManager-dispatcher.service has finished starting up. -- -- The start-up result is done. Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com NetworkManager[660]: [1724517288.7196] device (cni-podman2): state change: ip-check -> secondaries (reason 'none', sys-iface-state: 'external') Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com NetworkManager[660]: [1724517288.7198] device (cni-podman2): state change: secondaries -> activated (reason 'none', sys-iface-state: 'external') Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com NetworkManager[660]: [1724517288.7202] device (cni-podman2): Activation: successful, device activated. Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com dnsmasq[84059]: listening on cni-podman2(#8): 192.168.30.1 Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com dnsmasq[84071]: started, version 2.79 cachesize 150 Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com dnsmasq[84071]: compile time options: IPv6 GNU-getopt DBus no-i18n IDN2 DHCP DHCPv6 no-Lua TFTP no-conntrack ipset auth DNSSEC loop-detect inotify Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com dnsmasq[84071]: using local addresses only for domain dns.podman Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com dnsmasq[84071]: reading /etc/resolv.conf Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com dnsmasq[84071]: using local addresses only for domain dns.podman Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com dnsmasq[84071]: using nameserver 10.29.169.13#53 Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com dnsmasq[84071]: using nameserver 10.29.170.12#53 Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com dnsmasq[84071]: using nameserver 10.2.32.1#53 Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com dnsmasq[84071]: read /run/containers/cni/dnsname/systemd-quadlet-demo/addnhosts - 1 addresses Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started /usr/bin/podman healthcheck run 68250d5fc0218bc6db3ab2ece605ebe15c37a833d3345ad2e2df2c1916dd6cc3. -- Subject: Unit 68250d5fc0218bc6db3ab2ece605ebe15c37a833d3345ad2e2df2c1916dd6cc3.timer has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit 68250d5fc0218bc6db3ab2ece605ebe15c37a833d3345ad2e2df2c1916dd6cc3.timer has finished starting up. -- -- The start-up result is done. Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started quadlet-demo-mysql.service. -- Subject: Unit quadlet-demo-mysql.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit quadlet-demo-mysql.service has finished starting up. -- -- The start-up result is done. Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started /usr/bin/podman healthcheck run 68250d5fc0218bc6db3ab2ece605ebe15c37a833d3345ad2e2df2c1916dd6cc3. -- Subject: Unit 68250d5fc0218bc6db3ab2ece605ebe15c37a833d3345ad2e2df2c1916dd6cc3.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit 68250d5fc0218bc6db3ab2ece605ebe15c37a833d3345ad2e2df2c1916dd6cc3.service has finished starting up. -- -- The start-up result is done. Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com quadlet-demo-mysql[83918]: 68250d5fc0218bc6db3ab2ece605ebe15c37a833d3345ad2e2df2c1916dd6cc3 Aug 24 12:34:49 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: 68250d5fc0218bc6db3ab2ece605ebe15c37a833d3345ad2e2df2c1916dd6cc3.service: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit 68250d5fc0218bc6db3ab2ece605ebe15c37a833d3345ad2e2df2c1916dd6cc3.service has successfully entered the 'dead' state. Aug 24 12:34:49 ip-10-31-40-15.us-east-1.aws.redhat.com dnsmasq[84071]: listening on cni-podman2(#8): fe80::50:afff:fe96:3495%cni-podman2 Aug 24 12:34:49 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[84309]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None Aug 24 12:34:50 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[84452]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Aug 24 12:34:52 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[84577]: ansible-file Invoked with path=/etc/containers/systemd state=directory owner=root group=0 mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Aug 24 12:34:52 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[84724]: ansible-stat Invoked with path=/etc/containers/systemd/envoy-proxy-configmap.yml follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Aug 24 12:34:52 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[84823]: ansible-copy Invoked with src=/root/.ansible/tmp/ansible-tmp-1724517292.2012258-30396-55356854144861/source dest=/etc/containers/systemd/envoy-proxy-configmap.yml owner=root group=0 mode=0644 _original_basename=envoy-proxy-configmap.yml follow=False checksum=d681c7d56f912150d041873e880818b22a90c188 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None regexp=None delimiter=None Aug 24 12:34:53 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[84948]: ansible-systemd Invoked with daemon_reload=True scope=system daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None user=None Aug 24 12:34:53 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Reloading. Aug 24 12:34:54 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[85129]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None Aug 24 12:34:55 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[85262]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Aug 24 12:34:56 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[85394]: ansible-file Invoked with path=/etc/containers/systemd state=directory owner=root group=0 mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Aug 24 12:34:57 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[85517]: ansible-stat Invoked with path=/etc/containers/systemd/quadlet-demo.yml follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Aug 24 12:34:57 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[85616]: ansible-copy Invoked with dest=/etc/containers/systemd/quadlet-demo.yml owner=root group=0 mode=0644 src=/root/.ansible/tmp/ansible-tmp-1724517297.0551069-30667-219028600000855/source _original_basename=tmpwlrrnyo9 follow=False checksum=998dccde0483b1654327a46ddd89cbaa47650370 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None regexp=None delimiter=None Aug 24 12:34:58 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[85741]: ansible-systemd Invoked with daemon_reload=True scope=system daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None user=None Aug 24 12:34:58 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Reloading. Aug 24 12:34:58 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: NetworkManager-dispatcher.service: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit NetworkManager-dispatcher.service has successfully entered the 'dead' state. Aug 24 12:34:59 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[85918]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None Aug 24 12:34:59 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[86042]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Aug 24 12:35:00 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[86167]: ansible-slurp Invoked with path=/etc/containers/systemd/quadlet-demo.yml src=/etc/containers/systemd/quadlet-demo.yml Aug 24 12:35:01 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[86290]: ansible-file Invoked with path=/tmp/httpd3-create state=directory owner=root group=root recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Aug 24 12:35:02 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[86413]: ansible-file Invoked with path=/tmp/httpd3 state=directory owner=root group=root recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Aug 24 12:35:19 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started /usr/bin/podman healthcheck run 68250d5fc0218bc6db3ab2ece605ebe15c37a833d3345ad2e2df2c1916dd6cc3. -- Subject: Unit 68250d5fc0218bc6db3ab2ece605ebe15c37a833d3345ad2e2df2c1916dd6cc3.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit 68250d5fc0218bc6db3ab2ece605ebe15c37a833d3345ad2e2df2c1916dd6cc3.service has finished starting up. -- -- The start-up result is done. Aug 24 12:35:20 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: run-runc-68250d5fc0218bc6db3ab2ece605ebe15c37a833d3345ad2e2df2c1916dd6cc3-runc.4xgW0s.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit run-runc-68250d5fc0218bc6db3ab2ece605ebe15c37a833d3345ad2e2df2c1916dd6cc3-runc.4xgW0s.mount has successfully entered the 'dead' state. Aug 24 12:35:20 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: 68250d5fc0218bc6db3ab2ece605ebe15c37a833d3345ad2e2df2c1916dd6cc3.service: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit 68250d5fc0218bc6db3ab2ece605ebe15c37a833d3345ad2e2df2c1916dd6cc3.service has successfully entered the 'dead' state. Aug 24 12:35:20 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[87030]: ansible-file Invoked with path=/etc/containers/systemd state=directory owner=root group=0 mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Aug 24 12:35:21 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[87153]: ansible-stat Invoked with path=/etc/containers/systemd/quadlet-demo.kube follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Aug 24 12:35:21 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[87252]: ansible-copy Invoked with src=/root/.ansible/tmp/ansible-tmp-1724517321.03834-31464-224102551766944/source dest=/etc/containers/systemd/quadlet-demo.kube owner=root group=0 mode=0644 _original_basename=quadlet-demo.kube follow=False checksum=7a5c73a5d935a42431c87bcdbeb8a04ed0909dc7 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None regexp=None delimiter=None Aug 24 12:35:22 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[87377]: ansible-systemd Invoked with daemon_reload=True scope=system daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None user=None Aug 24 12:35:22 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Reloading. Aug 24 12:35:22 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[87530]: ansible-systemd Invoked with name=quadlet-demo.service scope=system state=started daemon_reload=False daemon_reexec=False no_block=False enabled=None force=None masked=None user=None Aug 24 12:35:22 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Starting quadlet-demo.service... -- Subject: Unit quadlet-demo.service has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit quadlet-demo.service has begun starting up. Aug 24 12:35:22 ip-10-31-40-15.us-east-1.aws.redhat.com quadlet-demo[87537]: Pods stopped: Aug 24 12:35:22 ip-10-31-40-15.us-east-1.aws.redhat.com quadlet-demo[87537]: Pods removed: Aug 24 12:35:22 ip-10-31-40-15.us-east-1.aws.redhat.com quadlet-demo[87537]: Secrets removed: Aug 24 12:35:22 ip-10-31-40-15.us-east-1.aws.redhat.com quadlet-demo[87537]: Volumes removed: Aug 24 12:35:22 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Created slice cgroup machine-libpod_pod_bbca104112c49d0abd527ad6773ff340419b9009bed96d17f6799e23cb2e88d4.slice. -- Subject: Unit machine-libpod_pod_bbca104112c49d0abd527ad6773ff340419b9009bed96d17f6799e23cb2e88d4.slice has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit machine-libpod_pod_bbca104112c49d0abd527ad6773ff340419b9009bed96d17f6799e23cb2e88d4.slice has finished starting up. -- -- The start-up result is done. Aug 24 12:35:23 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container a3a21fc2dc266efda8f166acd56c9ec75e90514a1f55d114243efd101b6d363e. -- Subject: Unit libpod-a3a21fc2dc266efda8f166acd56c9ec75e90514a1f55d114243efd101b6d363e.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-a3a21fc2dc266efda8f166acd56c9ec75e90514a1f55d114243efd101b6d363e.scope has finished starting up. -- -- The start-up result is done. Aug 24 12:35:23 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: IPv6: ADDRCONF(NETDEV_UP): veth1c0859f2: link is not ready Aug 24 12:35:23 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: cni-podman2: port 2(veth1c0859f2) entered blocking state Aug 24 12:35:23 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: cni-podman2: port 2(veth1c0859f2) entered disabled state Aug 24 12:35:23 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: device veth1c0859f2 entered promiscuous mode Aug 24 12:35:23 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: cni-podman2: port 2(veth1c0859f2) entered blocking state Aug 24 12:35:23 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: cni-podman2: port 2(veth1c0859f2) entered forwarding state Aug 24 12:35:23 ip-10-31-40-15.us-east-1.aws.redhat.com NetworkManager[660]: [1724517323.1746] manager: (veth1c0859f2): new Veth device (/org/freedesktop/NetworkManager/Devices/10) Aug 24 12:35:23 ip-10-31-40-15.us-east-1.aws.redhat.com systemd-udevd[87625]: link_config: autonegotiation is unset or enabled, the speed and duplex are not writable. Aug 24 12:35:23 ip-10-31-40-15.us-east-1.aws.redhat.com systemd-udevd[87625]: Could not generate persistent MAC address for veth1c0859f2: No such file or directory Aug 24 12:35:23 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: IPv6: ADDRCONF(NETDEV_CHANGE): veth1c0859f2: link becomes ready Aug 24 12:35:23 ip-10-31-40-15.us-east-1.aws.redhat.com NetworkManager[660]: [1724517323.1833] device (veth1c0859f2): carrier: link connected Aug 24 12:35:23 ip-10-31-40-15.us-east-1.aws.redhat.com dnsmasq[84071]: read /run/containers/cni/dnsname/systemd-quadlet-demo/addnhosts - 2 addresses Aug 24 12:35:23 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container c14ac2390f14575fc25c4d23b475705801a8d72451bc5ef8674c274401e2fc71. -- Subject: Unit libpod-c14ac2390f14575fc25c4d23b475705801a8d72451bc5ef8674c274401e2fc71.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-c14ac2390f14575fc25c4d23b475705801a8d72451bc5ef8674c274401e2fc71.scope has finished starting up. -- -- The start-up result is done. Aug 24 12:35:23 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 3ecc5b442f005a249ca7a72e1a6c6e613010d0c5c989a576b5aea86daadc9e6c. -- Subject: Unit libpod-3ecc5b442f005a249ca7a72e1a6c6e613010d0c5c989a576b5aea86daadc9e6c.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-3ecc5b442f005a249ca7a72e1a6c6e613010d0c5c989a576b5aea86daadc9e6c.scope has finished starting up. -- -- The start-up result is done. Aug 24 12:35:23 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 14fe7e5a1bc45af6ba6a3d0270a844a86eaa844a3fdfb935158304797b010f6e. -- Subject: Unit libpod-14fe7e5a1bc45af6ba6a3d0270a844a86eaa844a3fdfb935158304797b010f6e.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-14fe7e5a1bc45af6ba6a3d0270a844a86eaa844a3fdfb935158304797b010f6e.scope has finished starting up. -- -- The start-up result is done. Aug 24 12:35:23 ip-10-31-40-15.us-east-1.aws.redhat.com quadlet-demo[87537]: Volumes: Aug 24 12:35:23 ip-10-31-40-15.us-east-1.aws.redhat.com quadlet-demo[87537]: wp-pv-claim Aug 24 12:35:23 ip-10-31-40-15.us-east-1.aws.redhat.com quadlet-demo[87537]: Pod: Aug 24 12:35:23 ip-10-31-40-15.us-east-1.aws.redhat.com quadlet-demo[87537]: bbca104112c49d0abd527ad6773ff340419b9009bed96d17f6799e23cb2e88d4 Aug 24 12:35:23 ip-10-31-40-15.us-east-1.aws.redhat.com quadlet-demo[87537]: Containers: Aug 24 12:35:23 ip-10-31-40-15.us-east-1.aws.redhat.com quadlet-demo[87537]: 3ecc5b442f005a249ca7a72e1a6c6e613010d0c5c989a576b5aea86daadc9e6c Aug 24 12:35:23 ip-10-31-40-15.us-east-1.aws.redhat.com quadlet-demo[87537]: 14fe7e5a1bc45af6ba6a3d0270a844a86eaa844a3fdfb935158304797b010f6e Aug 24 12:35:23 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started quadlet-demo.service. -- Subject: Unit quadlet-demo.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit quadlet-demo.service has finished starting up. -- -- The start-up result is done. Aug 24 12:35:24 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[87981]: ansible-command Invoked with _raw_params=ls -alrtF /etc/containers/systemd warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:35:24 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[88178]: ansible-command Invoked with _raw_params=podman ps -a warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:35:25 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[88350]: ansible-command Invoked with _raw_params=podman volume ls warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:35:25 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[88480]: ansible-command Invoked with _raw_params=podman pod ps --ctr-ids --ctr-names --ctr-status warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:35:26 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[88611]: ansible-command Invoked with _raw_params=set -euo pipefail; systemctl list-units | grep quadlet _uses_shell=True warn=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:35:26 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[88737]: ansible-get_url Invoked with url=https://localhost:8000 dest=/run/out mode=0600 validate_certs=False force=False http_agent=ansible-httpget use_proxy=True force_basic_auth=False sha256sum= checksum= timeout=10 follow=False unsafe_writes=False url_username=None url_password=NOT_LOGGING_PARAMETER client_cert=None client_key=None backup=None headers=None tmp_dest=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None src=None content=NOT_LOGGING_PARAMETER remote_src=None regexp=None delimiter=None directory_mode=None Aug 24 12:35:27 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[88862]: ansible-command Invoked with _raw_params=cat /run/out warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:35:27 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[88986]: ansible-command Invoked with _raw_params=podman ps -a warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:35:28 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[89118]: ansible-command Invoked with _raw_params=podman pod ps --ctr-ids --ctr-names --ctr-status warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:35:28 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[89249]: ansible-command Invoked with _raw_params=set -euo pipefail; systemctl list-units --all | grep quadlet _uses_shell=True warn=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:35:29 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[89375]: ansible-command Invoked with _raw_params=ls -alrtF /etc/systemd/system warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:35:31 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[89624]: ansible-command Invoked with _raw_params=podman --version warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:35:32 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[89753]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None Aug 24 12:35:32 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[89877]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Aug 24 12:35:35 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[90002]: ansible-dnf Invoked with name=['firewalld'] state=present allow_downgrade=False autoremove=False bugfix=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True lock_timeout=30 conf_file=None disable_excludes=None download_dir=None list=None releasever=None Aug 24 12:35:38 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[90125]: ansible-systemd Invoked with name=firewalld masked=False daemon_reload=False daemon_reexec=False no_block=False state=None enabled=None force=None user=None scope=None Aug 24 12:35:38 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[90252]: ansible-systemd Invoked with name=firewalld state=started enabled=True daemon_reload=False daemon_reexec=False no_block=False force=None masked=None user=None scope=None Aug 24 12:35:39 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[90379]: ansible-fedora.linux_system_roles.firewall_lib Invoked with port=['8000/tcp'] permanent=True runtime=True state=enabled __report_changed=True service=[] source_port=[] forward_port=[] rich_rule=[] source=[] interface=[] interface_pci_id=[] icmp_block=[] timeout=0 ipset_entries=[] protocol=[] helper_module=[] destination=[] firewalld_conf=None masquerade=None icmp_block_inversion=None target=None zone=None set_default_zone=None ipset=None ipset_type=None description=None short=None Aug 24 12:35:39 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[90502]: ansible-fedora.linux_system_roles.firewall_lib Invoked with port=['9000/tcp'] permanent=True runtime=True state=enabled __report_changed=True service=[] source_port=[] forward_port=[] rich_rule=[] source=[] interface=[] interface_pci_id=[] icmp_block=[] timeout=0 ipset_entries=[] protocol=[] helper_module=[] destination=[] firewalld_conf=None masquerade=None icmp_block_inversion=None target=None zone=None set_default_zone=None ipset=None ipset_type=None description=None short=None Aug 24 12:35:43 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[91015]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None Aug 24 12:35:44 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[91139]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Aug 24 12:35:45 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[91264]: ansible-systemd Invoked with name=quadlet-demo.service scope=system state=stopped enabled=False force=True daemon_reload=False daemon_reexec=False no_block=False masked=None user=None Aug 24 12:35:45 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Reloading. Aug 24 12:35:45 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Stopping quadlet-demo.service... -- Subject: Unit quadlet-demo.service has begun shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit quadlet-demo.service has begun shutting down. Aug 24 12:35:45 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-a3a21fc2dc266efda8f166acd56c9ec75e90514a1f55d114243efd101b6d363e.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-a3a21fc2dc266efda8f166acd56c9ec75e90514a1f55d114243efd101b6d363e.scope has successfully entered the 'dead' state. Aug 24 12:35:45 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-a3a21fc2dc266efda8f166acd56c9ec75e90514a1f55d114243efd101b6d363e.scope: Consumed 35ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-a3a21fc2dc266efda8f166acd56c9ec75e90514a1f55d114243efd101b6d363e.scope completed and consumed the indicated resources. Aug 24 12:35:45 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-a3a21fc2dc266efda8f166acd56c9ec75e90514a1f55d114243efd101b6d363e-userdata-shm.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay\x2dcontainers-a3a21fc2dc266efda8f166acd56c9ec75e90514a1f55d114243efd101b6d363e-userdata-shm.mount has successfully entered the 'dead' state. Aug 24 12:35:45 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay-a98d97c31be0232a5dea2f79c3b27620993f5e7c2943a034f4688bd8808862df-merged.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay-a98d97c31be0232a5dea2f79c3b27620993f5e7c2943a034f4688bd8808862df-merged.mount has successfully entered the 'dead' state. Aug 24 12:35:47 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-3ecc5b442f005a249ca7a72e1a6c6e613010d0c5c989a576b5aea86daadc9e6c.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-3ecc5b442f005a249ca7a72e1a6c6e613010d0c5c989a576b5aea86daadc9e6c.scope has successfully entered the 'dead' state. Aug 24 12:35:47 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-3ecc5b442f005a249ca7a72e1a6c6e613010d0c5c989a576b5aea86daadc9e6c.scope: Consumed 1.056s CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-3ecc5b442f005a249ca7a72e1a6c6e613010d0c5c989a576b5aea86daadc9e6c.scope completed and consumed the indicated resources. Aug 24 12:35:47 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-14fe7e5a1bc45af6ba6a3d0270a844a86eaa844a3fdfb935158304797b010f6e.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-14fe7e5a1bc45af6ba6a3d0270a844a86eaa844a3fdfb935158304797b010f6e.scope has successfully entered the 'dead' state. Aug 24 12:35:47 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-14fe7e5a1bc45af6ba6a3d0270a844a86eaa844a3fdfb935158304797b010f6e.scope: Consumed 121ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-14fe7e5a1bc45af6ba6a3d0270a844a86eaa844a3fdfb935158304797b010f6e.scope completed and consumed the indicated resources. Aug 24 12:35:47 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay-a76eb00d8a3761c8d189a1ddebc0f043c3aa2c138bd8fa985fd3c1fc009a9aab-merged.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay-a76eb00d8a3761c8d189a1ddebc0f043c3aa2c138bd8fa985fd3c1fc009a9aab-merged.mount has successfully entered the 'dead' state. Aug 24 12:35:47 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay-fcf523d74cdb47e960956221b411933c2ce38983b3a6dcb9cceccb4b004baa54-merged.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay-fcf523d74cdb47e960956221b411933c2ce38983b3a6dcb9cceccb4b004baa54-merged.mount has successfully entered the 'dead' state. Aug 24 12:35:47 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-c14ac2390f14575fc25c4d23b475705801a8d72451bc5ef8674c274401e2fc71.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-c14ac2390f14575fc25c4d23b475705801a8d72451bc5ef8674c274401e2fc71.scope has successfully entered the 'dead' state. Aug 24 12:35:47 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-c14ac2390f14575fc25c4d23b475705801a8d72451bc5ef8674c274401e2fc71.scope: Consumed 31ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-c14ac2390f14575fc25c4d23b475705801a8d72451bc5ef8674c274401e2fc71.scope completed and consumed the indicated resources. Aug 24 12:35:47 ip-10-31-40-15.us-east-1.aws.redhat.com dnsmasq[84071]: read /run/containers/cni/dnsname/systemd-quadlet-demo/addnhosts - 1 addresses Aug 24 12:35:47 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: cni-podman2: port 2(veth1c0859f2) entered disabled state Aug 24 12:35:47 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: device veth1c0859f2 left promiscuous mode Aug 24 12:35:47 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: cni-podman2: port 2(veth1c0859f2) entered disabled state Aug 24 12:35:47 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: run-netns-netns\x2df8ada618\x2d4309\x2d2e92\x2d2bb3\x2d8c587e2dd2e5.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit run-netns-netns\x2df8ada618\x2d4309\x2d2e92\x2d2bb3\x2d8c587e2dd2e5.mount has successfully entered the 'dead' state. Aug 24 12:35:47 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-c14ac2390f14575fc25c4d23b475705801a8d72451bc5ef8674c274401e2fc71-userdata-shm.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay\x2dcontainers-c14ac2390f14575fc25c4d23b475705801a8d72451bc5ef8674c274401e2fc71-userdata-shm.mount has successfully entered the 'dead' state. Aug 24 12:35:47 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay-3dce984c4dfcb31a9a28077551ecae2ac7b04fc9ddd6862a8083fe91e9189f7e-merged.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay-3dce984c4dfcb31a9a28077551ecae2ac7b04fc9ddd6862a8083fe91e9189f7e-merged.mount has successfully entered the 'dead' state. Aug 24 12:35:47 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Removed slice cgroup machine-libpod_pod_bbca104112c49d0abd527ad6773ff340419b9009bed96d17f6799e23cb2e88d4.slice. -- Subject: Unit machine-libpod_pod_bbca104112c49d0abd527ad6773ff340419b9009bed96d17f6799e23cb2e88d4.slice has finished shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit machine-libpod_pod_bbca104112c49d0abd527ad6773ff340419b9009bed96d17f6799e23cb2e88d4.slice has finished shutting down. Aug 24 12:35:47 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: machine-libpod_pod_bbca104112c49d0abd527ad6773ff340419b9009bed96d17f6799e23cb2e88d4.slice: Consumed 1.210s CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit machine-libpod_pod_bbca104112c49d0abd527ad6773ff340419b9009bed96d17f6799e23cb2e88d4.slice completed and consumed the indicated resources. Aug 24 12:35:48 ip-10-31-40-15.us-east-1.aws.redhat.com quadlet-demo[91314]: Pods stopped: Aug 24 12:35:48 ip-10-31-40-15.us-east-1.aws.redhat.com quadlet-demo[91314]: bbca104112c49d0abd527ad6773ff340419b9009bed96d17f6799e23cb2e88d4 Aug 24 12:35:48 ip-10-31-40-15.us-east-1.aws.redhat.com quadlet-demo[91314]: Pods removed: Aug 24 12:35:48 ip-10-31-40-15.us-east-1.aws.redhat.com quadlet-demo[91314]: bbca104112c49d0abd527ad6773ff340419b9009bed96d17f6799e23cb2e88d4 Aug 24 12:35:48 ip-10-31-40-15.us-east-1.aws.redhat.com quadlet-demo[91314]: Secrets removed: Aug 24 12:35:48 ip-10-31-40-15.us-east-1.aws.redhat.com quadlet-demo[91314]: Volumes removed: Aug 24 12:35:48 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: quadlet-demo.service: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit quadlet-demo.service has successfully entered the 'dead' state. Aug 24 12:35:48 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Stopped quadlet-demo.service. -- Subject: Unit quadlet-demo.service has finished shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit quadlet-demo.service has finished shutting down. Aug 24 12:35:48 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[91620]: ansible-stat Invoked with path=/etc/containers/systemd/quadlet-demo.kube follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Aug 24 12:35:49 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[91868]: ansible-file Invoked with path=/etc/containers/systemd/quadlet-demo.kube state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Aug 24 12:35:50 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[91991]: ansible-systemd Invoked with daemon_reload=True scope=system daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None user=None Aug 24 12:35:50 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Reloading. Aug 24 12:35:50 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[92144]: ansible-command Invoked with _raw_params=podman image prune --all -f warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:35:50 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started /usr/bin/podman healthcheck run 68250d5fc0218bc6db3ab2ece605ebe15c37a833d3345ad2e2df2c1916dd6cc3. -- Subject: Unit 68250d5fc0218bc6db3ab2ece605ebe15c37a833d3345ad2e2df2c1916dd6cc3.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit 68250d5fc0218bc6db3ab2ece605ebe15c37a833d3345ad2e2df2c1916dd6cc3.service has finished starting up. -- -- The start-up result is done. Aug 24 12:35:51 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: 68250d5fc0218bc6db3ab2ece605ebe15c37a833d3345ad2e2df2c1916dd6cc3.service: Main process exited, code=exited, status=125/n/a Aug 24 12:35:51 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: 68250d5fc0218bc6db3ab2ece605ebe15c37a833d3345ad2e2df2c1916dd6cc3.service: Failed with result 'exit-code'. -- Subject: Unit failed -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit 68250d5fc0218bc6db3ab2ece605ebe15c37a833d3345ad2e2df2c1916dd6cc3.service has entered the 'failed' state with result 'exit-code'. Aug 24 12:35:51 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[92284]: ansible-command Invoked with _raw_params=podman images -n warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:35:52 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[92415]: ansible-command Invoked with _raw_params=podman volume ls -n warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:35:52 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[92545]: ansible-command Invoked with _raw_params=podman ps --noheading warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:35:53 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[92675]: ansible-command Invoked with _raw_params=podman network ls -n -q warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:35:54 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[92987]: ansible-service_facts Invoked Aug 24 12:35:56 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[93199]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None Aug 24 12:35:57 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[93323]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Aug 24 12:35:58 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[93448]: ansible-stat Invoked with path=/etc/containers/systemd/quadlet-demo.yml follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Aug 24 12:35:59 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[93696]: ansible-file Invoked with path=/etc/containers/systemd/quadlet-demo.yml state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Aug 24 12:35:59 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[93819]: ansible-systemd Invoked with daemon_reload=True scope=system daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None user=None Aug 24 12:35:59 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Reloading. Aug 24 12:36:00 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[93972]: ansible-command Invoked with _raw_params=podman image prune --all -f warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:36:00 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[94102]: ansible-command Invoked with _raw_params=podman images -n warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:36:01 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[94232]: ansible-command Invoked with _raw_params=podman volume ls -n warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:36:01 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[94362]: ansible-command Invoked with _raw_params=podman ps --noheading warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:36:01 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[94492]: ansible-command Invoked with _raw_params=podman network ls -n -q warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:36:02 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[94802]: ansible-service_facts Invoked Aug 24 12:36:04 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[95014]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None Aug 24 12:36:05 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[95138]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Aug 24 12:36:06 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[95263]: ansible-stat Invoked with path=/etc/containers/systemd/envoy-proxy-configmap.yml follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Aug 24 12:36:07 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[95511]: ansible-file Invoked with path=/etc/containers/systemd/envoy-proxy-configmap.yml state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Aug 24 12:36:07 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[95634]: ansible-systemd Invoked with daemon_reload=True scope=system daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None user=None Aug 24 12:36:07 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Reloading. Aug 24 12:36:08 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[95787]: ansible-command Invoked with _raw_params=podman image prune --all -f warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:36:08 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[95917]: ansible-command Invoked with _raw_params=podman images -n warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:36:09 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[96047]: ansible-command Invoked with _raw_params=podman volume ls -n warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:36:09 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[96177]: ansible-command Invoked with _raw_params=podman ps --noheading warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:36:10 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[96307]: ansible-command Invoked with _raw_params=podman network ls -n -q warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:36:10 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[96618]: ansible-service_facts Invoked Aug 24 12:36:12 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[96830]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None Aug 24 12:36:13 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[96954]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Aug 24 12:36:14 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[97079]: ansible-systemd Invoked with name=quadlet-demo-mysql.service scope=system state=stopped enabled=False force=True daemon_reload=False daemon_reexec=False no_block=False masked=None user=None Aug 24 12:36:14 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Reloading. Aug 24 12:36:14 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Stopping quadlet-demo-mysql.service... -- Subject: Unit quadlet-demo-mysql.service has begun shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit quadlet-demo-mysql.service has begun shutting down. Aug 24 12:36:15 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: 68250d5fc0218bc6db3ab2ece605ebe15c37a833d3345ad2e2df2c1916dd6cc3.timer: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit 68250d5fc0218bc6db3ab2ece605ebe15c37a833d3345ad2e2df2c1916dd6cc3.timer has successfully entered the 'dead' state. Aug 24 12:36:15 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Stopped /usr/bin/podman healthcheck run 68250d5fc0218bc6db3ab2ece605ebe15c37a833d3345ad2e2df2c1916dd6cc3. -- Subject: Unit 68250d5fc0218bc6db3ab2ece605ebe15c37a833d3345ad2e2df2c1916dd6cc3.timer has finished shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit 68250d5fc0218bc6db3ab2ece605ebe15c37a833d3345ad2e2df2c1916dd6cc3.timer has finished shutting down. Aug 24 12:36:15 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: cni-podman2: port 1(veth10ec5328) entered disabled state Aug 24 12:36:15 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: device veth10ec5328 left promiscuous mode Aug 24 12:36:15 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: cni-podman2: port 1(veth10ec5328) entered disabled state Aug 24 12:36:15 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: run-netns-netns\x2d700346f0\x2dda9a\x2d65a2\x2d3a64\x2d3b6e5ef382b2.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit run-netns-netns\x2d700346f0\x2dda9a\x2d65a2\x2d3a64\x2d3b6e5ef382b2.mount has successfully entered the 'dead' state. Aug 24 12:36:15 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-68250d5fc0218bc6db3ab2ece605ebe15c37a833d3345ad2e2df2c1916dd6cc3-userdata-shm.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay\x2dcontainers-68250d5fc0218bc6db3ab2ece605ebe15c37a833d3345ad2e2df2c1916dd6cc3-userdata-shm.mount has successfully entered the 'dead' state. Aug 24 12:36:15 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay-11b1e288ec3f02575f7a1a20a99b775e3487c1852d9248f0536b48f52045d256-merged.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay-11b1e288ec3f02575f7a1a20a99b775e3487c1852d9248f0536b48f52045d256-merged.mount has successfully entered the 'dead' state. Aug 24 12:36:15 ip-10-31-40-15.us-east-1.aws.redhat.com quadlet-demo-mysql[97115]: 68250d5fc0218bc6db3ab2ece605ebe15c37a833d3345ad2e2df2c1916dd6cc3 Aug 24 12:36:15 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Aug 24 12:36:15 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: quadlet-demo-mysql.service: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit quadlet-demo-mysql.service has successfully entered the 'dead' state. Aug 24 12:36:15 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Stopped quadlet-demo-mysql.service. -- Subject: Unit quadlet-demo-mysql.service has finished shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit quadlet-demo-mysql.service has finished shutting down. Aug 24 12:36:16 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[97363]: ansible-stat Invoked with path=/etc/containers/systemd/quadlet-demo-mysql.container follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Aug 24 12:36:17 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[97611]: ansible-file Invoked with path=/etc/containers/systemd/quadlet-demo-mysql.container state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Aug 24 12:36:17 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[97734]: ansible-systemd Invoked with daemon_reload=True scope=system daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None user=None Aug 24 12:36:17 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Reloading. Aug 24 12:36:18 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[98019]: ansible-command Invoked with _raw_params=podman image prune --all -f warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:36:18 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Aug 24 12:36:19 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[98150]: ansible-command Invoked with _raw_params=podman images -n warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:36:19 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Aug 24 12:36:19 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[98280]: ansible-command Invoked with _raw_params=podman volume ls -n warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:36:19 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Aug 24 12:36:20 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[98411]: ansible-command Invoked with _raw_params=podman ps --noheading warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:36:20 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[98541]: ansible-command Invoked with _raw_params=podman network ls -n -q warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:36:20 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Aug 24 12:36:21 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Aug 24 12:36:21 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[98853]: ansible-service_facts Invoked Aug 24 12:36:23 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[99064]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None Aug 24 12:36:23 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[99188]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Aug 24 12:36:25 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[99313]: ansible-systemd Invoked with name=quadlet-demo-mysql-volume.service scope=system state=stopped enabled=False force=True daemon_reload=False daemon_reexec=False no_block=False masked=None user=None Aug 24 12:36:25 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Reloading. Aug 24 12:36:25 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: quadlet-demo-mysql-volume.service: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit quadlet-demo-mysql-volume.service has successfully entered the 'dead' state. Aug 24 12:36:25 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Stopped quadlet-demo-mysql-volume.service. -- Subject: Unit quadlet-demo-mysql-volume.service has finished shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit quadlet-demo-mysql-volume.service has finished shutting down. Aug 24 12:36:25 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[99470]: ansible-stat Invoked with path=/etc/containers/systemd/quadlet-demo-mysql.volume follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Aug 24 12:36:26 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[99718]: ansible-file Invoked with path=/etc/containers/systemd/quadlet-demo-mysql.volume state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Aug 24 12:36:27 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[99841]: ansible-systemd Invoked with daemon_reload=True scope=system daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None user=None Aug 24 12:36:27 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Reloading. Aug 24 12:36:27 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Aug 24 12:36:28 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[100125]: ansible-command Invoked with _raw_params=podman image prune --all -f warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:36:28 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[100256]: ansible-command Invoked with _raw_params=podman images -n warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:36:28 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[100386]: ansible-command Invoked with _raw_params=podman volume ls -n warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:36:29 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[100517]: ansible-command Invoked with _raw_params=podman ps --noheading warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:36:29 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[100647]: ansible-command Invoked with _raw_params=podman network ls -n -q warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:36:29 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Aug 24 12:36:30 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[100957]: ansible-service_facts Invoked Aug 24 12:36:32 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[101168]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None Aug 24 12:36:33 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[101292]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Aug 24 12:36:34 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[101417]: ansible-systemd Invoked with name=quadlet-demo-network.service scope=system state=stopped enabled=False force=True daemon_reload=False daemon_reexec=False no_block=False masked=None user=None Aug 24 12:36:34 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Reloading. Aug 24 12:36:34 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: quadlet-demo-network.service: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit quadlet-demo-network.service has successfully entered the 'dead' state. Aug 24 12:36:34 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Stopped quadlet-demo-network.service. -- Subject: Unit quadlet-demo-network.service has finished shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit quadlet-demo-network.service has finished shutting down. Aug 24 12:36:34 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[101574]: ansible-stat Invoked with path=/etc/containers/systemd/quadlet-demo.network follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Aug 24 12:36:35 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[101822]: ansible-file Invoked with path=/etc/containers/systemd/quadlet-demo.network state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Aug 24 12:36:36 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[101945]: ansible-systemd Invoked with daemon_reload=True scope=system daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None user=None Aug 24 12:36:36 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Reloading. Aug 24 12:36:36 ip-10-31-40-15.us-east-1.aws.redhat.com NetworkManager[660]: [1724517396.6127] device (cni-podman2): state change: activated -> unmanaged (reason 'unmanaged', sys-iface-state: 'removed') Aug 24 12:36:36 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Aug 24 12:36:36 ip-10-31-40-15.us-east-1.aws.redhat.com dbus-daemon[590]: [system] Activating via systemd: service name='org.freedesktop.nm_dispatcher' unit='dbus-org.freedesktop.nm-dispatcher.service' requested by ':1.5' (uid=0 pid=660 comm="/usr/sbin/NetworkManager --no-daemon " label="system_u:system_r:NetworkManager_t:s0") Aug 24 12:36:36 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Starting Network Manager Script Dispatcher Service... -- Subject: Unit NetworkManager-dispatcher.service has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit NetworkManager-dispatcher.service has begun starting up. Aug 24 12:36:36 ip-10-31-40-15.us-east-1.aws.redhat.com dbus-daemon[590]: [system] Successfully activated service 'org.freedesktop.nm_dispatcher' Aug 24 12:36:36 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started Network Manager Script Dispatcher Service. -- Subject: Unit NetworkManager-dispatcher.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit NetworkManager-dispatcher.service has finished starting up. -- -- The start-up result is done. Aug 24 12:36:37 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[102290]: ansible-command Invoked with _raw_params=podman image prune --all -f warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:36:37 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[102420]: ansible-command Invoked with _raw_params=podman images -n warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:36:38 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[102550]: ansible-command Invoked with _raw_params=podman volume ls -n warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:36:38 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[102680]: ansible-command Invoked with _raw_params=podman ps --noheading warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:36:38 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[102810]: ansible-command Invoked with _raw_params=podman network ls -n -q warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:36:38 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Aug 24 12:36:39 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[103097]: ansible-service_facts Invoked Aug 24 12:36:41 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[103307]: ansible-command Invoked with _raw_params=exec 1>&2 set -x set -o pipefail systemctl list-units --plain -l --all | grep quadlet || : systemctl list-unit-files --all | grep quadlet || : systemctl list-units --plain --failed -l --all | grep quadlet || : _uses_shell=True warn=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Aug 24 12:36:42 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[103437]: ansible-command Invoked with _raw_params=journalctl -ex warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None PLAY RECAP ********************************************************************* managed_node1 : ok=414 changed=43 unreachable=0 failed=1 skipped=377 rescued=1 ignored=0 Saturday 24 August 2024 12:36:42 -0400 (0:00:00.474) 0:02:51.019 ******* =============================================================================== fedora.linux_system_roles.podman : Ensure container images are present -- 18.30s /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:18 fedora.linux_system_roles.podman : Ensure container images are present --- 7.23s /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:18 fedora.linux_system_roles.podman : Stop and disable service ------------- 3.21s /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:12 fedora.linux_system_roles.certificate : Ensure certificate role dependencies are installed --- 3.10s /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:5 fedora.linux_system_roles.certificate : Ensure provider packages are installed --- 2.81s /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:23 fedora.linux_system_roles.firewall : Install firewalld ------------------ 2.79s /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:31 fedora.linux_system_roles.firewall : Install firewalld ------------------ 2.78s /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:31 fedora.linux_system_roles.podman : For testing and debugging - services --- 1.99s /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:183 fedora.linux_system_roles.podman : Stop and disable service ------------- 1.94s /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:12 fedora.linux_system_roles.podman : Gather the package facts ------------- 1.88s /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:6 fedora.linux_system_roles.certificate : Remove files -------------------- 1.71s /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:181 fedora.linux_system_roles.podman : For testing and debugging - services --- 1.62s /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:183 fedora.linux_system_roles.podman : For testing and debugging - services --- 1.61s /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:183 fedora.linux_system_roles.podman : For testing and debugging - services --- 1.60s /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:183 fedora.linux_system_roles.podman : For testing and debugging - services --- 1.59s /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:183 fedora.linux_system_roles.podman : For testing and debugging - services --- 1.58s /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:183 fedora.linux_system_roles.podman : Gather the package facts ------------- 1.51s /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:6 fedora.linux_system_roles.podman : Start service ------------------------ 1.48s /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:110 fedora.linux_system_roles.certificate : Slurp the contents of the files --- 1.39s /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:152 fedora.linux_system_roles.certificate : Ensure certificate requests ----- 1.26s /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:101