ansible-playbook 2.9.27
config file = None
configured module search path = ['/root/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules']
ansible python module location = /usr/local/lib/python3.9/site-packages/ansible
executable location = /usr/local/bin/ansible-playbook
python version = 3.9.19 (main, May 16 2024, 11:40:09) [GCC 8.5.0 20210514 (Red Hat 8.5.0-22)]
No config file found; using defaults
[WARNING]: running playbook inside collection fedora.linux_system_roles
Skipping callback 'actionable', as we already have a stdout callback.
Skipping callback 'counter_enabled', as we already have a stdout callback.
Skipping callback 'debug', as we already have a stdout callback.
Skipping callback 'dense', as we already have a stdout callback.
Skipping callback 'dense', as we already have a stdout callback.
Skipping callback 'full_skip', as we already have a stdout callback.
Skipping callback 'json', as we already have a stdout callback.
Skipping callback 'minimal', as we already have a stdout callback.
Skipping callback 'null', as we already have a stdout callback.
Skipping callback 'oneline', as we already have a stdout callback.
Skipping callback 'selective', as we already have a stdout callback.
Skipping callback 'skippy', as we already have a stdout callback.
Skipping callback 'stderr', as we already have a stdout callback.
Skipping callback 'unixy', as we already have a stdout callback.
Skipping callback 'yaml', as we already have a stdout callback.
PLAYBOOK: tests_quadlet_demo.yml ***********************************************
2 plays in /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_demo.yml
PLAY [all] *********************************************************************
META: ran handlers
TASK [Include vault variables] *************************************************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_demo.yml:5
Saturday 24 August 2024 12:33:51 -0400 (0:00:00.030) 0:00:00.030 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_test_password": {
"__ansible_vault": "$ANSIBLE_VAULT;1.1;AES256\n35383939616163653333633431363463313831383037386236646138333162396161356130303461\n3932623930643263313563336163316337643562333936360a363538636631313039343233383732\n38666530383538656639363465313230343533386130303833336434303438333161656262346562\n3362626538613031640a663330613638366132356534363534353239616666653466353961323533\n6565\n"
},
"mysql_container_root_password": {
"__ansible_vault": "$ANSIBLE_VAULT;1.1;AES256\n61333932373230333539663035366431326163363166363036323963623131363530326231303634\n6635326161643165363366323062333334363730376631660a393566366139353861656364656661\n38653463363837336639363032646433666361646535366137303464623261313663643336306465\n6264663730656337310a343962353137386238383064646533366433333437303566656433386233\n34343235326665646661623131643335313236313131353661386338343366316261643634653633\n3832313034366536616531323963333234326461353130303532\n"
}
},
"ansible_included_var_files": [
"/tmp/podman-mRR/tests/vars/vault-variables.yml"
],
"changed": false
}
META: ran handlers
META: ran handlers
PLAY [Deploy the quadlet demo app] *********************************************
TASK [Gathering Facts] *********************************************************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_demo.yml:9
Saturday 24 August 2024 12:33:51 -0400 (0:00:00.042) 0:00:00.073 *******
ok: [managed_node1]
META: ran handlers
TASK [Test is only supported on x86_64] ****************************************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_demo.yml:38
Saturday 24 August 2024 12:33:52 -0400 (0:00:01.126) 0:00:01.200 *******
skipping: [managed_node1] => {}
META:
TASK [Generate certificates] ***************************************************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_demo.yml:51
Saturday 24 August 2024 12:33:52 -0400 (0:00:00.091) 0:00:01.292 *******
TASK [fedora.linux_system_roles.certificate : Set version specific variables] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:2
Saturday 24 August 2024 12:33:52 -0400 (0:00:00.069) 0:00:01.362 *******
included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/set_vars.yml for managed_node1
TASK [fedora.linux_system_roles.certificate : Ensure ansible_facts used by role] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/set_vars.yml:2
Saturday 24 August 2024 12:33:52 -0400 (0:00:00.037) 0:00:01.399 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.certificate : Check if system is ostree] *******
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/set_vars.yml:10
Saturday 24 August 2024 12:33:52 -0400 (0:00:00.021) 0:00:01.420 *******
ok: [managed_node1] => {
"changed": false,
"stat": {
"exists": false
}
}
TASK [fedora.linux_system_roles.certificate : Set flag to indicate system is ostree] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/set_vars.yml:15
Saturday 24 August 2024 12:33:53 -0400 (0:00:00.510) 0:00:01.931 *******
ok: [managed_node1] => {
"ansible_facts": {
"__certificate_is_ostree": false
},
"changed": false
}
TASK [fedora.linux_system_roles.certificate : Set platform/version specific variables] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/set_vars.yml:19
Saturday 24 August 2024 12:33:53 -0400 (0:00:00.035) 0:00:01.966 *******
skipping: [managed_node1] => (item=RedHat.yml) => {
"ansible_loop_var": "item",
"changed": false,
"item": "RedHat.yml",
"skip_reason": "Conditional result was False"
}
skipping: [managed_node1] => (item=CentOS.yml) => {
"ansible_loop_var": "item",
"changed": false,
"item": "CentOS.yml",
"skip_reason": "Conditional result was False"
}
skipping: [managed_node1] => (item=CentOS_8.yml) => {
"ansible_loop_var": "item",
"changed": false,
"item": "CentOS_8.yml",
"skip_reason": "Conditional result was False"
}
skipping: [managed_node1] => (item=CentOS_8.yml) => {
"ansible_loop_var": "item",
"changed": false,
"item": "CentOS_8.yml",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.certificate : Ensure certificate role dependencies are installed] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:5
Saturday 24 August 2024 12:33:53 -0400 (0:00:00.055) 0:00:02.022 *******
ok: [managed_node1] => {
"changed": false,
"rc": 0,
"results": []
}
MSG:
Nothing to do
lsrpackages: python3-cryptography python3-dbus python3-pyasn1
TASK [fedora.linux_system_roles.certificate : Ensure provider packages are installed] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:23
Saturday 24 August 2024 12:33:56 -0400 (0:00:03.099) 0:00:05.121 *******
ok: [managed_node1] => (item=certmonger) => {
"__certificate_provider": "certmonger",
"ansible_loop_var": "__certificate_provider",
"changed": false,
"rc": 0,
"results": []
}
MSG:
Nothing to do
lsrpackages: certmonger
TASK [fedora.linux_system_roles.certificate : Ensure pre-scripts hooks directory exists] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:35
Saturday 24 August 2024 12:33:59 -0400 (0:00:02.809) 0:00:07.930 *******
ok: [managed_node1] => (item=certmonger) => {
"__certificate_provider": "certmonger",
"ansible_loop_var": "__certificate_provider",
"changed": false,
"gid": 0,
"group": "root",
"mode": "0700",
"owner": "root",
"path": "/etc/certmonger//pre-scripts",
"secontext": "unconfined_u:object_r:etc_t:s0",
"size": 6,
"state": "directory",
"uid": 0
}
TASK [fedora.linux_system_roles.certificate : Ensure post-scripts hooks directory exists] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:61
Saturday 24 August 2024 12:34:00 -0400 (0:00:00.731) 0:00:08.661 *******
ok: [managed_node1] => (item=certmonger) => {
"__certificate_provider": "certmonger",
"ansible_loop_var": "__certificate_provider",
"changed": false,
"gid": 0,
"group": "root",
"mode": "0700",
"owner": "root",
"path": "/etc/certmonger//post-scripts",
"secontext": "unconfined_u:object_r:etc_t:s0",
"size": 6,
"state": "directory",
"uid": 0
}
TASK [fedora.linux_system_roles.certificate : Ensure provider service is running] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:90
Saturday 24 August 2024 12:34:00 -0400 (0:00:00.468) 0:00:09.130 *******
ok: [managed_node1] => (item=certmonger) => {
"__certificate_provider": "certmonger",
"ansible_loop_var": "__certificate_provider",
"changed": false,
"enabled": true,
"name": "certmonger",
"state": "started",
"status": {
"ActiveEnterTimestamp": "Sat 2024-08-24 12:29:03 EDT",
"ActiveEnterTimestampMonotonic": "346295924",
"ActiveExitTimestampMonotonic": "0",
"ActiveState": "active",
"After": "sysinit.target dbus.service systemd-journald.socket network.target basic.target dbus.socket system.slice syslog.target",
"AllowIsolate": "no",
"AllowedCPUs": "",
"AllowedMemoryNodes": "",
"AmbientCapabilities": "",
"AssertResult": "yes",
"AssertTimestamp": "Sat 2024-08-24 12:29:03 EDT",
"AssertTimestampMonotonic": "346284245",
"Before": "shutdown.target multi-user.target",
"BlockIOAccounting": "no",
"BlockIOWeight": "[not set]",
"BusName": "org.fedorahosted.certmonger",
"CPUAccounting": "no",
"CPUAffinity": "",
"CPUAffinityFromNUMA": "no",
"CPUQuotaPerSecUSec": "infinity",
"CPUQuotaPeriodUSec": "infinity",
"CPUSchedulingPolicy": "0",
"CPUSchedulingPriority": "0",
"CPUSchedulingResetOnFork": "no",
"CPUShares": "[not set]",
"CPUUsageNSec": "[not set]",
"CPUWeight": "[not set]",
"CacheDirectoryMode": "0755",
"CanFreeze": "yes",
"CanIsolate": "no",
"CanReload": "no",
"CanStart": "yes",
"CanStop": "yes",
"CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf",
"CollectMode": "inactive",
"ConditionResult": "yes",
"ConditionTimestamp": "Sat 2024-08-24 12:29:03 EDT",
"ConditionTimestampMonotonic": "346284243",
"ConfigurationDirectoryMode": "0755",
"Conflicts": "shutdown.target",
"ControlGroup": "/system.slice/certmonger.service",
"ControlPID": "0",
"DefaultDependencies": "yes",
"DefaultMemoryLow": "0",
"DefaultMemoryMin": "0",
"Delegate": "no",
"Description": "Certificate monitoring and PKI enrollment",
"DevicePolicy": "auto",
"DynamicUser": "no",
"EffectiveCPUs": "",
"EffectiveMemoryNodes": "",
"EnvironmentFiles": "/etc/sysconfig/certmonger (ignore_errors=yes)",
"ExecMainCode": "0",
"ExecMainExitTimestampMonotonic": "0",
"ExecMainPID": "12107",
"ExecMainStartTimestamp": "Sat 2024-08-24 12:29:03 EDT",
"ExecMainStartTimestampMonotonic": "346285051",
"ExecMainStatus": "0",
"ExecStart": "{ path=/usr/sbin/certmonger ; argv[]=/usr/sbin/certmonger -S -p /run/certmonger.pid -n $OPTS ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"FailureAction": "none",
"FileDescriptorStoreMax": "0",
"FragmentPath": "/usr/lib/systemd/system/certmonger.service",
"FreezerState": "running",
"GID": "[not set]",
"GuessMainPID": "yes",
"IOAccounting": "no",
"IOSchedulingClass": "0",
"IOSchedulingPriority": "0",
"IOWeight": "[not set]",
"IPAccounting": "no",
"IPEgressBytes": "18446744073709551615",
"IPEgressPackets": "18446744073709551615",
"IPIngressBytes": "18446744073709551615",
"IPIngressPackets": "18446744073709551615",
"Id": "certmonger.service",
"IgnoreOnIsolate": "no",
"IgnoreSIGPIPE": "yes",
"InactiveEnterTimestampMonotonic": "0",
"InactiveExitTimestamp": "Sat 2024-08-24 12:29:03 EDT",
"InactiveExitTimestampMonotonic": "346285098",
"InvocationID": "f4bd1c0399d64c7cb53a9acf6dd73285",
"JobRunningTimeoutUSec": "infinity",
"JobTimeoutAction": "none",
"JobTimeoutUSec": "infinity",
"KeyringMode": "private",
"KillMode": "control-group",
"KillSignal": "15",
"LimitAS": "infinity",
"LimitASSoft": "infinity",
"LimitCORE": "infinity",
"LimitCORESoft": "0",
"LimitCPU": "infinity",
"LimitCPUSoft": "infinity",
"LimitDATA": "infinity",
"LimitDATASoft": "infinity",
"LimitFSIZE": "infinity",
"LimitFSIZESoft": "infinity",
"LimitLOCKS": "infinity",
"LimitLOCKSSoft": "infinity",
"LimitMEMLOCK": "65536",
"LimitMEMLOCKSoft": "65536",
"LimitMSGQUEUE": "819200",
"LimitMSGQUEUESoft": "819200",
"LimitNICE": "0",
"LimitNICESoft": "0",
"LimitNOFILE": "262144",
"LimitNOFILESoft": "1024",
"LimitNPROC": "14003",
"LimitNPROCSoft": "14003",
"LimitRSS": "infinity",
"LimitRSSSoft": "infinity",
"LimitRTPRIO": "0",
"LimitRTPRIOSoft": "0",
"LimitRTTIME": "infinity",
"LimitRTTIMESoft": "infinity",
"LimitSIGPENDING": "14003",
"LimitSIGPENDINGSoft": "14003",
"LimitSTACK": "infinity",
"LimitSTACKSoft": "8388608",
"LoadState": "loaded",
"LockPersonality": "no",
"LogLevelMax": "-1",
"LogRateLimitBurst": "0",
"LogRateLimitIntervalUSec": "0",
"LogsDirectoryMode": "0755",
"MainPID": "12107",
"MemoryAccounting": "yes",
"MemoryCurrent": "3100672",
"MemoryDenyWriteExecute": "no",
"MemoryHigh": "infinity",
"MemoryLimit": "infinity",
"MemoryLow": "0",
"MemoryMax": "infinity",
"MemoryMin": "0",
"MemorySwapMax": "infinity",
"MountAPIVFS": "no",
"MountFlags": "",
"NFileDescriptorStore": "0",
"NRestarts": "0",
"NUMAMask": "",
"NUMAPolicy": "n/a",
"Names": "certmonger.service",
"NeedDaemonReload": "no",
"Nice": "0",
"NoNewPrivileges": "no",
"NonBlocking": "no",
"NotifyAccess": "none",
"OOMScoreAdjust": "0",
"OnFailureJobMode": "replace",
"PIDFile": "/run/certmonger.pid",
"PartOf": "dbus.service",
"PermissionsStartOnly": "no",
"Perpetual": "no",
"PrivateDevices": "no",
"PrivateMounts": "no",
"PrivateNetwork": "no",
"PrivateTmp": "no",
"PrivateUsers": "no",
"ProtectControlGroups": "no",
"ProtectHome": "no",
"ProtectKernelModules": "no",
"ProtectKernelTunables": "no",
"ProtectSystem": "no",
"RefuseManualStart": "no",
"RefuseManualStop": "no",
"RemainAfterExit": "no",
"RemoveIPC": "no",
"Requires": "sysinit.target system.slice dbus.socket",
"Restart": "no",
"RestartUSec": "100ms",
"RestrictNamespaces": "no",
"RestrictRealtime": "no",
"RestrictSUIDSGID": "no",
"Result": "success",
"RootDirectoryStartOnly": "no",
"RuntimeDirectoryMode": "0755",
"RuntimeDirectoryPreserve": "no",
"RuntimeMaxUSec": "infinity",
"SameProcessGroup": "no",
"SecureBits": "0",
"SendSIGHUP": "no",
"SendSIGKILL": "yes",
"Slice": "system.slice",
"StandardError": "inherit",
"StandardInput": "null",
"StandardInputData": "",
"StandardOutput": "journal",
"StartLimitAction": "none",
"StartLimitBurst": "5",
"StartLimitIntervalUSec": "10s",
"StartupBlockIOWeight": "[not set]",
"StartupCPUShares": "[not set]",
"StartupCPUWeight": "[not set]",
"StartupIOWeight": "[not set]",
"StateChangeTimestamp": "Sat 2024-08-24 12:29:03 EDT",
"StateChangeTimestampMonotonic": "346295924",
"StateDirectoryMode": "0755",
"StatusErrno": "0",
"StopWhenUnneeded": "no",
"SubState": "running",
"SuccessAction": "none",
"SyslogFacility": "3",
"SyslogLevel": "6",
"SyslogLevelPrefix": "yes",
"SyslogPriority": "30",
"SystemCallErrorNumber": "0",
"TTYReset": "no",
"TTYVHangup": "no",
"TTYVTDisallocate": "no",
"TasksAccounting": "yes",
"TasksCurrent": "1",
"TasksMax": "22405",
"TimeoutStartUSec": "1min 30s",
"TimeoutStopUSec": "1min 30s",
"TimerSlackNSec": "50000",
"Transient": "no",
"Type": "dbus",
"UID": "[not set]",
"UMask": "0022",
"UnitFilePreset": "disabled",
"UnitFileState": "enabled",
"UtmpMode": "init",
"WantedBy": "multi-user.target",
"WatchdogTimestamp": "Sat 2024-08-24 12:29:03 EDT",
"WatchdogTimestampMonotonic": "346295922",
"WatchdogUSec": "0"
}
}
TASK [fedora.linux_system_roles.certificate : Ensure certificate requests] *****
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:101
Saturday 24 August 2024 12:34:01 -0400 (0:00:01.061) 0:00:10.192 *******
changed: [managed_node1] => (item={'name': 'quadlet_demo', 'dns': ['localhost'], 'ca': 'self-sign'}) => {
"ansible_loop_var": "item",
"changed": true,
"item": {
"ca": "self-sign",
"dns": [
"localhost"
],
"name": "quadlet_demo"
}
}
MSG:
Certificate requested (new).
TASK [fedora.linux_system_roles.certificate : Slurp the contents of the files] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:152
Saturday 24 August 2024 12:34:02 -0400 (0:00:01.256) 0:00:11.448 *******
ok: [managed_node1] => (item=['cert', {'name': 'quadlet_demo', 'dns': ['localhost'], 'ca': 'self-sign'}]) => {
"ansible_loop_var": "item",
"changed": false,
"content": "LS0tLS1CRUdJTiBDRVJUSUZJQ0FURS0tLS0tCk1JSURnekNDQW11Z0F3SUJBZ0lSQUxxZnN3V1ZSa25NdTg3clBDa0JaVnN3RFFZSktvWklodmNOQVFFTEJRQXcKVURFZ01CNEdBMVVFQXd3WFRHOWpZV3dnVTJsbmJtbHVaeUJCZFhSb2IzSnBkSGt4TERBcUJnTlZCQU1NSTJKaApPV1ppTXpBMUxUazFORFkwT1dOakxXSmlZMlZsWWpOakxUSTVNREUyTlRVNU1CNFhEVEkwTURneU5ERTJNelF3Ck1sb1hEVEkxTURneU5ERTJNamt3TTFvd0ZERVNNQkFHQTFVRUF4TUpiRzlqWVd4b2IzTjBNSUlCSWpBTkJna3EKaGtpRzl3MEJBUUVGQUFPQ0FROEFNSUlCQ2dLQ0FRRUFrekNka1R1VnAxSzRxSmxtWFhGTG1Bb3ZqbVVhT1FYRApVRk5UOVkyeHM2VHZleEFHaEh0S1Z4b3VxSW8zdUhOT2VTd3Z0SEU3NXBJMzVmeDlPdVpHbE9sWEhsUFVycmpiCm5JOVY3amtSWUhWdi9iSlJxSzZOYmc3dXF0c0VuN0FwVDVadFdXWWw5amM0Ykg4OEpDb2Rka0wrLy9zYllwRzkKa2hQaGo0NWhESVkzUTg4cW0vU0R1Z1dybUNHa0lMaFZHclQrM0RpaWI0ZFpGYVZZKzB4MnBOeWQ3YXdYZW1hQwphd3A0UHlJeEhSbjJuUUxjblZYa0hVVHhnOUJPcTNLWVM0dUsvY0VjclFHeEVNV0pkR0lmUmMxUUx6dzcvT3ArCnRNWDBVdUZLOFFwT2ZBYlIwNjlNdUVhSXlTZzFtMmFTK1FxNjhGV3BOMURBNmloR05CMkU3UUlEQVFBQm80R1QKTUlHUU1Bc0dBMVVkRHdRRUF3SUZvREFVQmdOVkhSRUVEVEFMZ2dsc2IyTmhiR2h2YzNRd0hRWURWUjBsQkJZdwpGQVlJS3dZQkJRVUhBd0VHQ0NzR0FRVUZCd01DTUF3R0ExVWRFd0VCL3dRQ01BQXdIUVlEVlIwT0JCWUVGSUVzClVIenQ5dTRIMzVKNzFsQzJyeUNkdmtabE1COEdBMVVkSXdRWU1CYUFGSG95TXRWMVNheE1UVW4wcm42UVkwNUwKM28rd01BMEdDU3FHU0liM0RRRUJDd1VBQTRJQkFRQnpVY0l6OHozQVNOVXg5MnYxYlc2OFFsdlpCditROHJZUQo1M0JYKzFYOSsyVENUVE91bUxabkloZnFlQVJLbkJIdjE3Lzg4ZFBlQXNPVTdlT2JXWXpwOGU2TVJaSHIvWW45CjJXMXJQakQvLzRERFZaeUduOXZLc1Yzc1RiR1ExZi9HN2UrZVljd0JDSWNKOTNOajlZM292QnRwWUJpWTdFNnoKOUhjT2txSGlWSnJxaS9VWmtLUk1TcjRobU9CeFhvc280ZXhxUUVaU0xocjB1NnNKTWx4REtlMk5CLysxbGNPaQo0bHZ0ODVHbTZONXU1QWJ0amx1Y1ZGNXpKTjFtVlRpd1FaZTdFakUweXA0NkxrVVBkMlRwQ3ROSmRhWmpENVZICmYyUTFIWHpuN0VPSjVvODlHcWhEaExhUi9VVjN0SU4rbEFUUXQwREVIQ1k4TFpBNDZNOGQKLS0tLS1FTkQgQ0VSVElGSUNBVEUtLS0tLQo=",
"encoding": "base64",
"item": [
"cert",
{
"ca": "self-sign",
"dns": [
"localhost"
],
"name": "quadlet_demo"
}
],
"source": "/etc/pki/tls/certs/quadlet_demo.crt"
}
ok: [managed_node1] => (item=['key', {'name': 'quadlet_demo', 'dns': ['localhost'], 'ca': 'self-sign'}]) => {
"ansible_loop_var": "item",
"changed": false,
"content": "LS0tLS1CRUdJTiBQUklWQVRFIEtFWS0tLS0tCk1JSUV2QUlCQURBTkJna3Foa2lHOXcwQkFRRUZBQVNDQktZd2dnU2lBZ0VBQW9JQkFRQ1RNSjJSTzVXblVyaW8KbVdaZGNVdVlDaStPWlJvNUJjTlFVMVAxamJHenBPOTdFQWFFZTBwWEdpNm9pamU0YzA1NUxDKzBjVHZta2pmbAovSDA2NWthVTZWY2VVOVN1dU51Y2oxWHVPUkZnZFcvOXNsR29ybzF1RHU2cTJ3U2ZzQ2xQbG0xWlppWDJOemhzCmZ6d2tLaDEyUXY3Lyt4dGlrYjJTRStHUGptRU1oamREenlxYjlJTzZCYXVZSWFRZ3VGVWF0UDdjT0tKdmgxa1YKcFZqN1RIYWszSjN0ckJkNlpvSnJDbmcvSWpFZEdmYWRBdHlkVmVRZFJQR0QwRTZyY3BoTGk0cjl3Unl0QWJFUQp4WWwwWWg5RnpWQXZQRHY4Nm42MHhmUlM0VXJ4Q2s1OEJ0SFRyMHk0Um9qSktEV2JacEw1Q3Jyd1ZhazNVTURxCktFWTBIWVR0QWdNQkFBRUNnZ0VBVEFJWDdDNjcybHNpVnBaWXdUcEhVaTA0WDYwZi9JQllXdnVjNkxvU0lQVzAKWVhWbUdEWGVydG9yYkFkc1BNajJwaVpHUUVVVkxMMG5OcEhMVFlZcXZxMDh1c0dKbSt5RUkvWDA1djc3VlRkdgo3SytQOGE2cUJFS2J4d2t3K3hHcVU1b1h2OG03SXNsalRzdnBqN2RiM1gwOVY0MGhtQ1l3b2VXNFhqeWxLeVMvCm1VUUZPK3VpaGd5VUR3NEszdHJEZUhNUXhYVlRQTkNRY3pxSXk4eXB5aFJoVXBQdGtzdGxkbHQ4K1dHa29qNloKdTBDRUNGSkpvK2FEYnRWMUhPbEVqK1NUTllxd3BTSUw0RUlMQ0FyOHAxZ3BKZHE1V1NQNlpaNXNtYmtZVm5Rego2QnkyN2pLb1VuYXRJOW8zM0FDZm5HMkM5Nm1sRmdLNWFTUlBjUGFaRlFLQmdRREQ4dXlPUnJ6c0xTM1czVTBGCnRZM3hneEtsVDNyOWFjeGNIY1NIRi82akdXYXNSZzI5U2JLbTdxZUhiVWp0bHB6bkk1VWN5cTlIYVR3Z1U5OHIKdGVuekRYQTUzbXhJZ3ZtdGRlRlk3NnZ3YzhYbFlaRTEyQUFoa0RpSHVTOEl1M2NuN2tEZ1FUd2UwZnZUWWxEMgpHWWwwYmhtL1JnbzlTbWYyNUtWR0x3and6d0tCZ1FEQVRGT243ZkV6MHdSSE9maWlHbTB6MEhmMUl3RjJFN2ZvClQ4YWxwaXhQMWg2V2dUaFdwUWpvckNJTE1LUHRiN25EdEJTcllwN1o5VGNYMzVnbWFjbUpVRmZpUUQ2Y1l0V1QKTWxad2E5cm5Dalh6QloxcUNqZnNyZUI0c2RReEdpWmxOUFNoWkFYcWdCajd6bjRGY3dUVDFGQ0VpSGhGd0RTYwpaUG5QQzlYRmd3S0JnR1Y3ck5STEFwRjNIS1NKWE5pMkhOdVdQTlB0QkErMithWkRySU1qQkRoSWNZcUxoWlloClV5dHRzc3BLNlZoVGh0T1c5Q1ZETytCS09sK3IwSXQ4M2J1Sm5EUDhaQ3h4MlFuallXQ0VVSkpxQTNZM1A5T1oKR1FoSS8xNGJwS1pySU4xSFljWkRRdlhiWHdjdGhpelBNUWRoOW1lL3lKSFJ0V0tETEszYXNsQzFBb0dBQ0s2cApqRWg2L25zRmsyMnpveVlNVGkyWWJQSUNxK0VpQ1pRb2hPd216Z2Y5NDQrOENuVnpISVpCRno0LytJVjEzcmN2Cnc3azFNT1FZYzdSOXV5WkZtalNOYzBLeFVRQkpZbGNBZmh6d2lrUDB3Wjg4UjBCazNBUFJ3Y0ZOc2xYYTYvRGoKSnJKTWJoVVFpN1g4REtUdno5eWpNZ2l6WFhyYkpVQ040dUpHNlo4Q2dZQVgwOExCd1FuWHJiOHZEVE9scHRHbwpCMHU2NnVKN3pWelFLNmxSMXR0N3MzMk9RSjJLdWt3WkVCbVZZMkFjckkxNW1rc2dqVTNWWFY1d1prWTl2K2tiCmpFaFNJT2RxZGZIa2x4ZnBwWTVneEY2TjA2WUpOLzdJNzNEWWFvZXdIVU9BTVVRSTdxZVB5SEpURDZjbHp3Z2QKeEtpWlFsM3JRMytlOU9qNVFITmlYdz09Ci0tLS0tRU5EIFBSSVZBVEUgS0VZLS0tLS0K",
"encoding": "base64",
"item": [
"key",
{
"ca": "self-sign",
"dns": [
"localhost"
],
"name": "quadlet_demo"
}
],
"source": "/etc/pki/tls/private/quadlet_demo.key"
}
ok: [managed_node1] => (item=['ca', {'name': 'quadlet_demo', 'dns': ['localhost'], 'ca': 'self-sign'}]) => {
"ansible_loop_var": "item",
"changed": false,
"content": "LS0tLS1CRUdJTiBDRVJUSUZJQ0FURS0tLS0tCk1JSURnekNDQW11Z0F3SUJBZ0lSQUxxZnN3V1ZSa25NdTg3clBDa0JaVnN3RFFZSktvWklodmNOQVFFTEJRQXcKVURFZ01CNEdBMVVFQXd3WFRHOWpZV3dnVTJsbmJtbHVaeUJCZFhSb2IzSnBkSGt4TERBcUJnTlZCQU1NSTJKaApPV1ppTXpBMUxUazFORFkwT1dOakxXSmlZMlZsWWpOakxUSTVNREUyTlRVNU1CNFhEVEkwTURneU5ERTJNelF3Ck1sb1hEVEkxTURneU5ERTJNamt3TTFvd0ZERVNNQkFHQTFVRUF4TUpiRzlqWVd4b2IzTjBNSUlCSWpBTkJna3EKaGtpRzl3MEJBUUVGQUFPQ0FROEFNSUlCQ2dLQ0FRRUFrekNka1R1VnAxSzRxSmxtWFhGTG1Bb3ZqbVVhT1FYRApVRk5UOVkyeHM2VHZleEFHaEh0S1Z4b3VxSW8zdUhOT2VTd3Z0SEU3NXBJMzVmeDlPdVpHbE9sWEhsUFVycmpiCm5JOVY3amtSWUhWdi9iSlJxSzZOYmc3dXF0c0VuN0FwVDVadFdXWWw5amM0Ykg4OEpDb2Rka0wrLy9zYllwRzkKa2hQaGo0NWhESVkzUTg4cW0vU0R1Z1dybUNHa0lMaFZHclQrM0RpaWI0ZFpGYVZZKzB4MnBOeWQ3YXdYZW1hQwphd3A0UHlJeEhSbjJuUUxjblZYa0hVVHhnOUJPcTNLWVM0dUsvY0VjclFHeEVNV0pkR0lmUmMxUUx6dzcvT3ArCnRNWDBVdUZLOFFwT2ZBYlIwNjlNdUVhSXlTZzFtMmFTK1FxNjhGV3BOMURBNmloR05CMkU3UUlEQVFBQm80R1QKTUlHUU1Bc0dBMVVkRHdRRUF3SUZvREFVQmdOVkhSRUVEVEFMZ2dsc2IyTmhiR2h2YzNRd0hRWURWUjBsQkJZdwpGQVlJS3dZQkJRVUhBd0VHQ0NzR0FRVUZCd01DTUF3R0ExVWRFd0VCL3dRQ01BQXdIUVlEVlIwT0JCWUVGSUVzClVIenQ5dTRIMzVKNzFsQzJyeUNkdmtabE1COEdBMVVkSXdRWU1CYUFGSG95TXRWMVNheE1UVW4wcm42UVkwNUwKM28rd01BMEdDU3FHU0liM0RRRUJDd1VBQTRJQkFRQnpVY0l6OHozQVNOVXg5MnYxYlc2OFFsdlpCditROHJZUQo1M0JYKzFYOSsyVENUVE91bUxabkloZnFlQVJLbkJIdjE3Lzg4ZFBlQXNPVTdlT2JXWXpwOGU2TVJaSHIvWW45CjJXMXJQakQvLzRERFZaeUduOXZLc1Yzc1RiR1ExZi9HN2UrZVljd0JDSWNKOTNOajlZM292QnRwWUJpWTdFNnoKOUhjT2txSGlWSnJxaS9VWmtLUk1TcjRobU9CeFhvc280ZXhxUUVaU0xocjB1NnNKTWx4REtlMk5CLysxbGNPaQo0bHZ0ODVHbTZONXU1QWJ0amx1Y1ZGNXpKTjFtVlRpd1FaZTdFakUweXA0NkxrVVBkMlRwQ3ROSmRhWmpENVZICmYyUTFIWHpuN0VPSjVvODlHcWhEaExhUi9VVjN0SU4rbEFUUXQwREVIQ1k4TFpBNDZNOGQKLS0tLS1FTkQgQ0VSVElGSUNBVEUtLS0tLQo=",
"encoding": "base64",
"item": [
"ca",
{
"ca": "self-sign",
"dns": [
"localhost"
],
"name": "quadlet_demo"
}
],
"source": "/etc/pki/tls/certs/quadlet_demo.crt"
}
TASK [fedora.linux_system_roles.certificate : Create return data] **************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:160
Saturday 24 August 2024 12:34:04 -0400 (0:00:01.392) 0:00:12.841 *******
ok: [managed_node1] => {
"ansible_facts": {
"certificate_test_certs": {
"quadlet_demo": {
"ca": "/etc/pki/tls/certs/quadlet_demo.crt",
"ca_content": "-----BEGIN CERTIFICATE-----\nMIIDgzCCAmugAwIBAgIRALqfswWVRknMu87rPCkBZVswDQYJKoZIhvcNAQELBQAw\nUDEgMB4GA1UEAwwXTG9jYWwgU2lnbmluZyBBdXRob3JpdHkxLDAqBgNVBAMMI2Jh\nOWZiMzA1LTk1NDY0OWNjLWJiY2VlYjNjLTI5MDE2NTU5MB4XDTI0MDgyNDE2MzQw\nMloXDTI1MDgyNDE2MjkwM1owFDESMBAGA1UEAxMJbG9jYWxob3N0MIIBIjANBgkq\nhkiG9w0BAQEFAAOCAQ8AMIIBCgKCAQEAkzCdkTuVp1K4qJlmXXFLmAovjmUaOQXD\nUFNT9Y2xs6TvexAGhHtKVxouqIo3uHNOeSwvtHE75pI35fx9OuZGlOlXHlPUrrjb\nnI9V7jkRYHVv/bJRqK6Nbg7uqtsEn7ApT5ZtWWYl9jc4bH88JCoddkL+//sbYpG9\nkhPhj45hDIY3Q88qm/SDugWrmCGkILhVGrT+3Diib4dZFaVY+0x2pNyd7awXemaC\nawp4PyIxHRn2nQLcnVXkHUTxg9BOq3KYS4uK/cEcrQGxEMWJdGIfRc1QLzw7/Op+\ntMX0UuFK8QpOfAbR069MuEaIySg1m2aS+Qq68FWpN1DA6ihGNB2E7QIDAQABo4GT\nMIGQMAsGA1UdDwQEAwIFoDAUBgNVHREEDTALgglsb2NhbGhvc3QwHQYDVR0lBBYw\nFAYIKwYBBQUHAwEGCCsGAQUFBwMCMAwGA1UdEwEB/wQCMAAwHQYDVR0OBBYEFIEs\nUHzt9u4H35J71lC2ryCdvkZlMB8GA1UdIwQYMBaAFHoyMtV1SaxMTUn0rn6QY05L\n3o+wMA0GCSqGSIb3DQEBCwUAA4IBAQBzUcIz8z3ASNUx92v1bW68QlvZBv+Q8rYQ\n53BX+1X9+2TCTTOumLZnIhfqeARKnBHv17/88dPeAsOU7eObWYzp8e6MRZHr/Yn9\n2W1rPjD//4DDVZyGn9vKsV3sTbGQ1f/G7e+eYcwBCIcJ93Nj9Y3ovBtpYBiY7E6z\n9HcOkqHiVJrqi/UZkKRMSr4hmOBxXoso4exqQEZSLhr0u6sJMlxDKe2NB/+1lcOi\n4lvt85Gm6N5u5AbtjlucVF5zJN1mVTiwQZe7EjE0yp46LkUPd2TpCtNJdaZjD5VH\nf2Q1HXzn7EOJ5o89GqhDhLaR/UV3tIN+lATQt0DEHCY8LZA46M8d\n-----END CERTIFICATE-----\n",
"cert": "/etc/pki/tls/certs/quadlet_demo.crt",
"cert_content": "-----BEGIN CERTIFICATE-----\nMIIDgzCCAmugAwIBAgIRALqfswWVRknMu87rPCkBZVswDQYJKoZIhvcNAQELBQAw\nUDEgMB4GA1UEAwwXTG9jYWwgU2lnbmluZyBBdXRob3JpdHkxLDAqBgNVBAMMI2Jh\nOWZiMzA1LTk1NDY0OWNjLWJiY2VlYjNjLTI5MDE2NTU5MB4XDTI0MDgyNDE2MzQw\nMloXDTI1MDgyNDE2MjkwM1owFDESMBAGA1UEAxMJbG9jYWxob3N0MIIBIjANBgkq\nhkiG9w0BAQEFAAOCAQ8AMIIBCgKCAQEAkzCdkTuVp1K4qJlmXXFLmAovjmUaOQXD\nUFNT9Y2xs6TvexAGhHtKVxouqIo3uHNOeSwvtHE75pI35fx9OuZGlOlXHlPUrrjb\nnI9V7jkRYHVv/bJRqK6Nbg7uqtsEn7ApT5ZtWWYl9jc4bH88JCoddkL+//sbYpG9\nkhPhj45hDIY3Q88qm/SDugWrmCGkILhVGrT+3Diib4dZFaVY+0x2pNyd7awXemaC\nawp4PyIxHRn2nQLcnVXkHUTxg9BOq3KYS4uK/cEcrQGxEMWJdGIfRc1QLzw7/Op+\ntMX0UuFK8QpOfAbR069MuEaIySg1m2aS+Qq68FWpN1DA6ihGNB2E7QIDAQABo4GT\nMIGQMAsGA1UdDwQEAwIFoDAUBgNVHREEDTALgglsb2NhbGhvc3QwHQYDVR0lBBYw\nFAYIKwYBBQUHAwEGCCsGAQUFBwMCMAwGA1UdEwEB/wQCMAAwHQYDVR0OBBYEFIEs\nUHzt9u4H35J71lC2ryCdvkZlMB8GA1UdIwQYMBaAFHoyMtV1SaxMTUn0rn6QY05L\n3o+wMA0GCSqGSIb3DQEBCwUAA4IBAQBzUcIz8z3ASNUx92v1bW68QlvZBv+Q8rYQ\n53BX+1X9+2TCTTOumLZnIhfqeARKnBHv17/88dPeAsOU7eObWYzp8e6MRZHr/Yn9\n2W1rPjD//4DDVZyGn9vKsV3sTbGQ1f/G7e+eYcwBCIcJ93Nj9Y3ovBtpYBiY7E6z\n9HcOkqHiVJrqi/UZkKRMSr4hmOBxXoso4exqQEZSLhr0u6sJMlxDKe2NB/+1lcOi\n4lvt85Gm6N5u5AbtjlucVF5zJN1mVTiwQZe7EjE0yp46LkUPd2TpCtNJdaZjD5VH\nf2Q1HXzn7EOJ5o89GqhDhLaR/UV3tIN+lATQt0DEHCY8LZA46M8d\n-----END CERTIFICATE-----\n",
"key": "/etc/pki/tls/private/quadlet_demo.key",
"key_content": "-----BEGIN PRIVATE KEY-----\nMIIEvAIBADANBgkqhkiG9w0BAQEFAASCBKYwggSiAgEAAoIBAQCTMJ2RO5WnUrio\nmWZdcUuYCi+OZRo5BcNQU1P1jbGzpO97EAaEe0pXGi6oije4c055LC+0cTvmkjfl\n/H065kaU6VceU9SuuNucj1XuORFgdW/9slGoro1uDu6q2wSfsClPlm1ZZiX2Nzhs\nfzwkKh12Qv7/+xtikb2SE+GPjmEMhjdDzyqb9IO6BauYIaQguFUatP7cOKJvh1kV\npVj7THak3J3trBd6ZoJrCng/IjEdGfadAtydVeQdRPGD0E6rcphLi4r9wRytAbEQ\nxYl0Yh9FzVAvPDv86n60xfRS4UrxCk58BtHTr0y4RojJKDWbZpL5CrrwVak3UMDq\nKEY0HYTtAgMBAAECggEATAIX7C672lsiVpZYwTpHUi04X60f/IBYWvuc6LoSIPW0\nYXVmGDXertorbAdsPMj2piZGQEUVLL0nNpHLTYYqvq08usGJm+yEI/X05v77VTdv\n7K+P8a6qBEKbxwkw+xGqU5oXv8m7IsljTsvpj7db3X09V40hmCYwoeW4XjylKyS/\nmUQFO+uihgyUDw4K3trDeHMQxXVTPNCQczqIy8ypyhRhUpPtkstldlt8+WGkoj6Z\nu0CECFJJo+aDbtV1HOlEj+STNYqwpSIL4EILCAr8p1gpJdq5WSP6ZZ5smbkYVnQz\n6By27jKoUnatI9o33ACfnG2C96mlFgK5aSRPcPaZFQKBgQDD8uyORrzsLS3W3U0F\ntY3xgxKlT3r9acxcHcSHF/6jGWasRg29SbKm7qeHbUjtlpznI5Ucyq9HaTwgU98r\ntenzDXA53mxIgvmtdeFY76vwc8XlYZE12AAhkDiHuS8Iu3cn7kDgQTwe0fvTYlD2\nGYl0bhm/Rgo9Smf25KVGLwjwzwKBgQDATFOn7fEz0wRHOfiiGm0z0Hf1IwF2E7fo\nT8alpixP1h6WgThWpQjorCILMKPtb7nDtBSrYp7Z9TcX35gmacmJUFfiQD6cYtWT\nMlZwa9rnCjXzBZ1qCjfsreB4sdQxGiZlNPShZAXqgBj7zn4FcwTT1FCEiHhFwDSc\nZPnPC9XFgwKBgGV7rNRLApF3HKSJXNi2HNuWPNPtBA+2+aZDrIMjBDhIcYqLhZYh\nUyttsspK6VhThtOW9CVDO+BKOl+r0It83buJnDP8ZCxx2QnjYWCEUJJqA3Y3P9OZ\nGQhI/14bpKZrIN1HYcZDQvXbXwcthizPMQdh9me/yJHRtWKDLK3aslC1AoGACK6p\njEh6/nsFk22zoyYMTi2YbPICq+EiCZQohOwmzgf944+8CnVzHIZBFz4/+IV13rcv\nw7k1MOQYc7R9uyZFmjSNc0KxUQBJYlcAfhzwikP0wZ88R0Bk3APRwcFNslXa6/Dj\nJrJMbhUQi7X8DKTvz9yjMgizXXrbJUCN4uJG6Z8CgYAX08LBwQnXrb8vDTOlptGo\nB0u66uJ7zVzQK6lR1tt7s32OQJ2KukwZEBmVY2AcrI15mksgjU3VXV5wZkY9v+kb\njEhSIOdqdfHklxfppY5gxF6N06YJN/7I73DYaoewHUOAMUQI7qePyHJTD6clzwgd\nxKiZQl3rQ3+e9Oj5QHNiXw==\n-----END PRIVATE KEY-----\n"
}
}
},
"changed": false
}
TASK [fedora.linux_system_roles.certificate : Stop tracking certificates] ******
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:176
Saturday 24 August 2024 12:34:04 -0400 (0:00:00.025) 0:00:12.867 *******
ok: [managed_node1] => (item={'cert': '/etc/pki/tls/certs/quadlet_demo.crt', 'cert_content': '-----BEGIN CERTIFICATE-----\nMIIDgzCCAmugAwIBAgIRALqfswWVRknMu87rPCkBZVswDQYJKoZIhvcNAQELBQAw\nUDEgMB4GA1UEAwwXTG9jYWwgU2lnbmluZyBBdXRob3JpdHkxLDAqBgNVBAMMI2Jh\nOWZiMzA1LTk1NDY0OWNjLWJiY2VlYjNjLTI5MDE2NTU5MB4XDTI0MDgyNDE2MzQw\nMloXDTI1MDgyNDE2MjkwM1owFDESMBAGA1UEAxMJbG9jYWxob3N0MIIBIjANBgkq\nhkiG9w0BAQEFAAOCAQ8AMIIBCgKCAQEAkzCdkTuVp1K4qJlmXXFLmAovjmUaOQXD\nUFNT9Y2xs6TvexAGhHtKVxouqIo3uHNOeSwvtHE75pI35fx9OuZGlOlXHlPUrrjb\nnI9V7jkRYHVv/bJRqK6Nbg7uqtsEn7ApT5ZtWWYl9jc4bH88JCoddkL+//sbYpG9\nkhPhj45hDIY3Q88qm/SDugWrmCGkILhVGrT+3Diib4dZFaVY+0x2pNyd7awXemaC\nawp4PyIxHRn2nQLcnVXkHUTxg9BOq3KYS4uK/cEcrQGxEMWJdGIfRc1QLzw7/Op+\ntMX0UuFK8QpOfAbR069MuEaIySg1m2aS+Qq68FWpN1DA6ihGNB2E7QIDAQABo4GT\nMIGQMAsGA1UdDwQEAwIFoDAUBgNVHREEDTALgglsb2NhbGhvc3QwHQYDVR0lBBYw\nFAYIKwYBBQUHAwEGCCsGAQUFBwMCMAwGA1UdEwEB/wQCMAAwHQYDVR0OBBYEFIEs\nUHzt9u4H35J71lC2ryCdvkZlMB8GA1UdIwQYMBaAFHoyMtV1SaxMTUn0rn6QY05L\n3o+wMA0GCSqGSIb3DQEBCwUAA4IBAQBzUcIz8z3ASNUx92v1bW68QlvZBv+Q8rYQ\n53BX+1X9+2TCTTOumLZnIhfqeARKnBHv17/88dPeAsOU7eObWYzp8e6MRZHr/Yn9\n2W1rPjD//4DDVZyGn9vKsV3sTbGQ1f/G7e+eYcwBCIcJ93Nj9Y3ovBtpYBiY7E6z\n9HcOkqHiVJrqi/UZkKRMSr4hmOBxXoso4exqQEZSLhr0u6sJMlxDKe2NB/+1lcOi\n4lvt85Gm6N5u5AbtjlucVF5zJN1mVTiwQZe7EjE0yp46LkUPd2TpCtNJdaZjD5VH\nf2Q1HXzn7EOJ5o89GqhDhLaR/UV3tIN+lATQt0DEHCY8LZA46M8d\n-----END CERTIFICATE-----\n', 'key': '/etc/pki/tls/private/quadlet_demo.key', 'key_content': '-----BEGIN PRIVATE KEY-----\nMIIEvAIBADANBgkqhkiG9w0BAQEFAASCBKYwggSiAgEAAoIBAQCTMJ2RO5WnUrio\nmWZdcUuYCi+OZRo5BcNQU1P1jbGzpO97EAaEe0pXGi6oije4c055LC+0cTvmkjfl\n/H065kaU6VceU9SuuNucj1XuORFgdW/9slGoro1uDu6q2wSfsClPlm1ZZiX2Nzhs\nfzwkKh12Qv7/+xtikb2SE+GPjmEMhjdDzyqb9IO6BauYIaQguFUatP7cOKJvh1kV\npVj7THak3J3trBd6ZoJrCng/IjEdGfadAtydVeQdRPGD0E6rcphLi4r9wRytAbEQ\nxYl0Yh9FzVAvPDv86n60xfRS4UrxCk58BtHTr0y4RojJKDWbZpL5CrrwVak3UMDq\nKEY0HYTtAgMBAAECggEATAIX7C672lsiVpZYwTpHUi04X60f/IBYWvuc6LoSIPW0\nYXVmGDXertorbAdsPMj2piZGQEUVLL0nNpHLTYYqvq08usGJm+yEI/X05v77VTdv\n7K+P8a6qBEKbxwkw+xGqU5oXv8m7IsljTsvpj7db3X09V40hmCYwoeW4XjylKyS/\nmUQFO+uihgyUDw4K3trDeHMQxXVTPNCQczqIy8ypyhRhUpPtkstldlt8+WGkoj6Z\nu0CECFJJo+aDbtV1HOlEj+STNYqwpSIL4EILCAr8p1gpJdq5WSP6ZZ5smbkYVnQz\n6By27jKoUnatI9o33ACfnG2C96mlFgK5aSRPcPaZFQKBgQDD8uyORrzsLS3W3U0F\ntY3xgxKlT3r9acxcHcSHF/6jGWasRg29SbKm7qeHbUjtlpznI5Ucyq9HaTwgU98r\ntenzDXA53mxIgvmtdeFY76vwc8XlYZE12AAhkDiHuS8Iu3cn7kDgQTwe0fvTYlD2\nGYl0bhm/Rgo9Smf25KVGLwjwzwKBgQDATFOn7fEz0wRHOfiiGm0z0Hf1IwF2E7fo\nT8alpixP1h6WgThWpQjorCILMKPtb7nDtBSrYp7Z9TcX35gmacmJUFfiQD6cYtWT\nMlZwa9rnCjXzBZ1qCjfsreB4sdQxGiZlNPShZAXqgBj7zn4FcwTT1FCEiHhFwDSc\nZPnPC9XFgwKBgGV7rNRLApF3HKSJXNi2HNuWPNPtBA+2+aZDrIMjBDhIcYqLhZYh\nUyttsspK6VhThtOW9CVDO+BKOl+r0It83buJnDP8ZCxx2QnjYWCEUJJqA3Y3P9OZ\nGQhI/14bpKZrIN1HYcZDQvXbXwcthizPMQdh9me/yJHRtWKDLK3aslC1AoGACK6p\njEh6/nsFk22zoyYMTi2YbPICq+EiCZQohOwmzgf944+8CnVzHIZBFz4/+IV13rcv\nw7k1MOQYc7R9uyZFmjSNc0KxUQBJYlcAfhzwikP0wZ88R0Bk3APRwcFNslXa6/Dj\nJrJMbhUQi7X8DKTvz9yjMgizXXrbJUCN4uJG6Z8CgYAX08LBwQnXrb8vDTOlptGo\nB0u66uJ7zVzQK6lR1tt7s32OQJ2KukwZEBmVY2AcrI15mksgjU3VXV5wZkY9v+kb\njEhSIOdqdfHklxfppY5gxF6N06YJN/7I73DYaoewHUOAMUQI7qePyHJTD6clzwgd\nxKiZQl3rQ3+e9Oj5QHNiXw==\n-----END PRIVATE KEY-----\n', 'ca': '/etc/pki/tls/certs/quadlet_demo.crt', 'ca_content': '-----BEGIN CERTIFICATE-----\nMIIDgzCCAmugAwIBAgIRALqfswWVRknMu87rPCkBZVswDQYJKoZIhvcNAQELBQAw\nUDEgMB4GA1UEAwwXTG9jYWwgU2lnbmluZyBBdXRob3JpdHkxLDAqBgNVBAMMI2Jh\nOWZiMzA1LTk1NDY0OWNjLWJiY2VlYjNjLTI5MDE2NTU5MB4XDTI0MDgyNDE2MzQw\nMloXDTI1MDgyNDE2MjkwM1owFDESMBAGA1UEAxMJbG9jYWxob3N0MIIBIjANBgkq\nhkiG9w0BAQEFAAOCAQ8AMIIBCgKCAQEAkzCdkTuVp1K4qJlmXXFLmAovjmUaOQXD\nUFNT9Y2xs6TvexAGhHtKVxouqIo3uHNOeSwvtHE75pI35fx9OuZGlOlXHlPUrrjb\nnI9V7jkRYHVv/bJRqK6Nbg7uqtsEn7ApT5ZtWWYl9jc4bH88JCoddkL+//sbYpG9\nkhPhj45hDIY3Q88qm/SDugWrmCGkILhVGrT+3Diib4dZFaVY+0x2pNyd7awXemaC\nawp4PyIxHRn2nQLcnVXkHUTxg9BOq3KYS4uK/cEcrQGxEMWJdGIfRc1QLzw7/Op+\ntMX0UuFK8QpOfAbR069MuEaIySg1m2aS+Qq68FWpN1DA6ihGNB2E7QIDAQABo4GT\nMIGQMAsGA1UdDwQEAwIFoDAUBgNVHREEDTALgglsb2NhbGhvc3QwHQYDVR0lBBYw\nFAYIKwYBBQUHAwEGCCsGAQUFBwMCMAwGA1UdEwEB/wQCMAAwHQYDVR0OBBYEFIEs\nUHzt9u4H35J71lC2ryCdvkZlMB8GA1UdIwQYMBaAFHoyMtV1SaxMTUn0rn6QY05L\n3o+wMA0GCSqGSIb3DQEBCwUAA4IBAQBzUcIz8z3ASNUx92v1bW68QlvZBv+Q8rYQ\n53BX+1X9+2TCTTOumLZnIhfqeARKnBHv17/88dPeAsOU7eObWYzp8e6MRZHr/Yn9\n2W1rPjD//4DDVZyGn9vKsV3sTbGQ1f/G7e+eYcwBCIcJ93Nj9Y3ovBtpYBiY7E6z\n9HcOkqHiVJrqi/UZkKRMSr4hmOBxXoso4exqQEZSLhr0u6sJMlxDKe2NB/+1lcOi\n4lvt85Gm6N5u5AbtjlucVF5zJN1mVTiwQZe7EjE0yp46LkUPd2TpCtNJdaZjD5VH\nf2Q1HXzn7EOJ5o89GqhDhLaR/UV3tIN+lATQt0DEHCY8LZA46M8d\n-----END CERTIFICATE-----\n'}) => {
"ansible_loop_var": "item",
"changed": false,
"cmd": [
"getcert",
"stop-tracking",
"-f",
"/etc/pki/tls/certs/quadlet_demo.crt"
],
"delta": "0:00:00.030775",
"end": "2024-08-24 12:34:04.757078",
"item": {
"ca": "/etc/pki/tls/certs/quadlet_demo.crt",
"ca_content": "-----BEGIN CERTIFICATE-----\nMIIDgzCCAmugAwIBAgIRALqfswWVRknMu87rPCkBZVswDQYJKoZIhvcNAQELBQAw\nUDEgMB4GA1UEAwwXTG9jYWwgU2lnbmluZyBBdXRob3JpdHkxLDAqBgNVBAMMI2Jh\nOWZiMzA1LTk1NDY0OWNjLWJiY2VlYjNjLTI5MDE2NTU5MB4XDTI0MDgyNDE2MzQw\nMloXDTI1MDgyNDE2MjkwM1owFDESMBAGA1UEAxMJbG9jYWxob3N0MIIBIjANBgkq\nhkiG9w0BAQEFAAOCAQ8AMIIBCgKCAQEAkzCdkTuVp1K4qJlmXXFLmAovjmUaOQXD\nUFNT9Y2xs6TvexAGhHtKVxouqIo3uHNOeSwvtHE75pI35fx9OuZGlOlXHlPUrrjb\nnI9V7jkRYHVv/bJRqK6Nbg7uqtsEn7ApT5ZtWWYl9jc4bH88JCoddkL+//sbYpG9\nkhPhj45hDIY3Q88qm/SDugWrmCGkILhVGrT+3Diib4dZFaVY+0x2pNyd7awXemaC\nawp4PyIxHRn2nQLcnVXkHUTxg9BOq3KYS4uK/cEcrQGxEMWJdGIfRc1QLzw7/Op+\ntMX0UuFK8QpOfAbR069MuEaIySg1m2aS+Qq68FWpN1DA6ihGNB2E7QIDAQABo4GT\nMIGQMAsGA1UdDwQEAwIFoDAUBgNVHREEDTALgglsb2NhbGhvc3QwHQYDVR0lBBYw\nFAYIKwYBBQUHAwEGCCsGAQUFBwMCMAwGA1UdEwEB/wQCMAAwHQYDVR0OBBYEFIEs\nUHzt9u4H35J71lC2ryCdvkZlMB8GA1UdIwQYMBaAFHoyMtV1SaxMTUn0rn6QY05L\n3o+wMA0GCSqGSIb3DQEBCwUAA4IBAQBzUcIz8z3ASNUx92v1bW68QlvZBv+Q8rYQ\n53BX+1X9+2TCTTOumLZnIhfqeARKnBHv17/88dPeAsOU7eObWYzp8e6MRZHr/Yn9\n2W1rPjD//4DDVZyGn9vKsV3sTbGQ1f/G7e+eYcwBCIcJ93Nj9Y3ovBtpYBiY7E6z\n9HcOkqHiVJrqi/UZkKRMSr4hmOBxXoso4exqQEZSLhr0u6sJMlxDKe2NB/+1lcOi\n4lvt85Gm6N5u5AbtjlucVF5zJN1mVTiwQZe7EjE0yp46LkUPd2TpCtNJdaZjD5VH\nf2Q1HXzn7EOJ5o89GqhDhLaR/UV3tIN+lATQt0DEHCY8LZA46M8d\n-----END CERTIFICATE-----\n",
"cert": "/etc/pki/tls/certs/quadlet_demo.crt",
"cert_content": "-----BEGIN CERTIFICATE-----\nMIIDgzCCAmugAwIBAgIRALqfswWVRknMu87rPCkBZVswDQYJKoZIhvcNAQELBQAw\nUDEgMB4GA1UEAwwXTG9jYWwgU2lnbmluZyBBdXRob3JpdHkxLDAqBgNVBAMMI2Jh\nOWZiMzA1LTk1NDY0OWNjLWJiY2VlYjNjLTI5MDE2NTU5MB4XDTI0MDgyNDE2MzQw\nMloXDTI1MDgyNDE2MjkwM1owFDESMBAGA1UEAxMJbG9jYWxob3N0MIIBIjANBgkq\nhkiG9w0BAQEFAAOCAQ8AMIIBCgKCAQEAkzCdkTuVp1K4qJlmXXFLmAovjmUaOQXD\nUFNT9Y2xs6TvexAGhHtKVxouqIo3uHNOeSwvtHE75pI35fx9OuZGlOlXHlPUrrjb\nnI9V7jkRYHVv/bJRqK6Nbg7uqtsEn7ApT5ZtWWYl9jc4bH88JCoddkL+//sbYpG9\nkhPhj45hDIY3Q88qm/SDugWrmCGkILhVGrT+3Diib4dZFaVY+0x2pNyd7awXemaC\nawp4PyIxHRn2nQLcnVXkHUTxg9BOq3KYS4uK/cEcrQGxEMWJdGIfRc1QLzw7/Op+\ntMX0UuFK8QpOfAbR069MuEaIySg1m2aS+Qq68FWpN1DA6ihGNB2E7QIDAQABo4GT\nMIGQMAsGA1UdDwQEAwIFoDAUBgNVHREEDTALgglsb2NhbGhvc3QwHQYDVR0lBBYw\nFAYIKwYBBQUHAwEGCCsGAQUFBwMCMAwGA1UdEwEB/wQCMAAwHQYDVR0OBBYEFIEs\nUHzt9u4H35J71lC2ryCdvkZlMB8GA1UdIwQYMBaAFHoyMtV1SaxMTUn0rn6QY05L\n3o+wMA0GCSqGSIb3DQEBCwUAA4IBAQBzUcIz8z3ASNUx92v1bW68QlvZBv+Q8rYQ\n53BX+1X9+2TCTTOumLZnIhfqeARKnBHv17/88dPeAsOU7eObWYzp8e6MRZHr/Yn9\n2W1rPjD//4DDVZyGn9vKsV3sTbGQ1f/G7e+eYcwBCIcJ93Nj9Y3ovBtpYBiY7E6z\n9HcOkqHiVJrqi/UZkKRMSr4hmOBxXoso4exqQEZSLhr0u6sJMlxDKe2NB/+1lcOi\n4lvt85Gm6N5u5AbtjlucVF5zJN1mVTiwQZe7EjE0yp46LkUPd2TpCtNJdaZjD5VH\nf2Q1HXzn7EOJ5o89GqhDhLaR/UV3tIN+lATQt0DEHCY8LZA46M8d\n-----END CERTIFICATE-----\n",
"key": "/etc/pki/tls/private/quadlet_demo.key",
"key_content": "-----BEGIN PRIVATE KEY-----\nMIIEvAIBADANBgkqhkiG9w0BAQEFAASCBKYwggSiAgEAAoIBAQCTMJ2RO5WnUrio\nmWZdcUuYCi+OZRo5BcNQU1P1jbGzpO97EAaEe0pXGi6oije4c055LC+0cTvmkjfl\n/H065kaU6VceU9SuuNucj1XuORFgdW/9slGoro1uDu6q2wSfsClPlm1ZZiX2Nzhs\nfzwkKh12Qv7/+xtikb2SE+GPjmEMhjdDzyqb9IO6BauYIaQguFUatP7cOKJvh1kV\npVj7THak3J3trBd6ZoJrCng/IjEdGfadAtydVeQdRPGD0E6rcphLi4r9wRytAbEQ\nxYl0Yh9FzVAvPDv86n60xfRS4UrxCk58BtHTr0y4RojJKDWbZpL5CrrwVak3UMDq\nKEY0HYTtAgMBAAECggEATAIX7C672lsiVpZYwTpHUi04X60f/IBYWvuc6LoSIPW0\nYXVmGDXertorbAdsPMj2piZGQEUVLL0nNpHLTYYqvq08usGJm+yEI/X05v77VTdv\n7K+P8a6qBEKbxwkw+xGqU5oXv8m7IsljTsvpj7db3X09V40hmCYwoeW4XjylKyS/\nmUQFO+uihgyUDw4K3trDeHMQxXVTPNCQczqIy8ypyhRhUpPtkstldlt8+WGkoj6Z\nu0CECFJJo+aDbtV1HOlEj+STNYqwpSIL4EILCAr8p1gpJdq5WSP6ZZ5smbkYVnQz\n6By27jKoUnatI9o33ACfnG2C96mlFgK5aSRPcPaZFQKBgQDD8uyORrzsLS3W3U0F\ntY3xgxKlT3r9acxcHcSHF/6jGWasRg29SbKm7qeHbUjtlpznI5Ucyq9HaTwgU98r\ntenzDXA53mxIgvmtdeFY76vwc8XlYZE12AAhkDiHuS8Iu3cn7kDgQTwe0fvTYlD2\nGYl0bhm/Rgo9Smf25KVGLwjwzwKBgQDATFOn7fEz0wRHOfiiGm0z0Hf1IwF2E7fo\nT8alpixP1h6WgThWpQjorCILMKPtb7nDtBSrYp7Z9TcX35gmacmJUFfiQD6cYtWT\nMlZwa9rnCjXzBZ1qCjfsreB4sdQxGiZlNPShZAXqgBj7zn4FcwTT1FCEiHhFwDSc\nZPnPC9XFgwKBgGV7rNRLApF3HKSJXNi2HNuWPNPtBA+2+aZDrIMjBDhIcYqLhZYh\nUyttsspK6VhThtOW9CVDO+BKOl+r0It83buJnDP8ZCxx2QnjYWCEUJJqA3Y3P9OZ\nGQhI/14bpKZrIN1HYcZDQvXbXwcthizPMQdh9me/yJHRtWKDLK3aslC1AoGACK6p\njEh6/nsFk22zoyYMTi2YbPICq+EiCZQohOwmzgf944+8CnVzHIZBFz4/+IV13rcv\nw7k1MOQYc7R9uyZFmjSNc0KxUQBJYlcAfhzwikP0wZ88R0Bk3APRwcFNslXa6/Dj\nJrJMbhUQi7X8DKTvz9yjMgizXXrbJUCN4uJG6Z8CgYAX08LBwQnXrb8vDTOlptGo\nB0u66uJ7zVzQK6lR1tt7s32OQJ2KukwZEBmVY2AcrI15mksgjU3VXV5wZkY9v+kb\njEhSIOdqdfHklxfppY5gxF6N06YJN/7I73DYaoewHUOAMUQI7qePyHJTD6clzwgd\nxKiZQl3rQ3+e9Oj5QHNiXw==\n-----END PRIVATE KEY-----\n"
},
"rc": 0,
"start": "2024-08-24 12:34:04.726303"
}
STDOUT:
Request "20240824163402" removed.
TASK [fedora.linux_system_roles.certificate : Remove files] ********************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:181
Saturday 24 August 2024 12:34:04 -0400 (0:00:00.562) 0:00:13.429 *******
changed: [managed_node1] => (item=/etc/pki/tls/certs/quadlet_demo.crt) => {
"ansible_loop_var": "item",
"changed": true,
"item": "/etc/pki/tls/certs/quadlet_demo.crt",
"path": "/etc/pki/tls/certs/quadlet_demo.crt",
"state": "absent"
}
changed: [managed_node1] => (item=/etc/pki/tls/private/quadlet_demo.key) => {
"ansible_loop_var": "item",
"changed": true,
"item": "/etc/pki/tls/private/quadlet_demo.key",
"path": "/etc/pki/tls/private/quadlet_demo.key",
"state": "absent"
}
ok: [managed_node1] => (item=/etc/pki/tls/certs/quadlet_demo.crt) => {
"ansible_loop_var": "item",
"changed": false,
"item": "/etc/pki/tls/certs/quadlet_demo.crt",
"path": "/etc/pki/tls/certs/quadlet_demo.crt",
"state": "absent"
}
TASK [Run the role] ************************************************************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_demo.yml:62
Saturday 24 August 2024 12:34:06 -0400 (0:00:01.714) 0:00:15.143 *******
TASK [fedora.linux_system_roles.podman : Set platform/version specific variables] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:3
Saturday 24 August 2024 12:34:06 -0400 (0:00:00.117) 0:00:15.260 *******
included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml for managed_node1
TASK [fedora.linux_system_roles.podman : Ensure ansible_facts used by role] ****
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:3
Saturday 24 August 2024 12:34:06 -0400 (0:00:00.029) 0:00:15.290 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Check if system is ostree] ************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:11
Saturday 24 August 2024 12:34:06 -0400 (0:00:00.023) 0:00:15.314 *******
ok: [managed_node1] => {
"changed": false,
"stat": {
"exists": false
}
}
TASK [fedora.linux_system_roles.podman : Set flag to indicate system is ostree] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:16
Saturday 24 August 2024 12:34:07 -0400 (0:00:00.419) 0:00:15.733 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_is_ostree": false
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Check if transactional-update exists in /sbin] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:23
Saturday 24 August 2024 12:34:07 -0400 (0:00:00.031) 0:00:15.764 *******
ok: [managed_node1] => {
"changed": false,
"stat": {
"exists": false
}
}
TASK [fedora.linux_system_roles.podman : Set flag if transactional-update exists] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:28
Saturday 24 August 2024 12:34:07 -0400 (0:00:00.376) 0:00:16.141 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_is_transactional": false
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set platform/version specific variables] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:32
Saturday 24 August 2024 12:34:07 -0400 (0:00:00.052) 0:00:16.193 *******
ok: [managed_node1] => (item=RedHat.yml) => {
"ansible_facts": {
"__podman_packages": [
"podman",
"shadow-utils-subid"
]
},
"ansible_included_var_files": [
"/tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/vars/RedHat.yml"
],
"ansible_loop_var": "item",
"changed": false,
"item": "RedHat.yml"
}
skipping: [managed_node1] => (item=CentOS.yml) => {
"ansible_loop_var": "item",
"changed": false,
"item": "CentOS.yml",
"skip_reason": "Conditional result was False"
}
ok: [managed_node1] => (item=CentOS_8.yml) => {
"ansible_facts": {
"__podman_packages": [
"crun",
"podman",
"podman-plugins",
"shadow-utils-subid"
]
},
"ansible_included_var_files": [
"/tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/vars/CentOS_8.yml"
],
"ansible_loop_var": "item",
"changed": false,
"item": "CentOS_8.yml"
}
ok: [managed_node1] => (item=CentOS_8.yml) => {
"ansible_facts": {
"__podman_packages": [
"crun",
"podman",
"podman-plugins",
"shadow-utils-subid"
]
},
"ansible_included_var_files": [
"/tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/vars/CentOS_8.yml"
],
"ansible_loop_var": "item",
"changed": false,
"item": "CentOS_8.yml"
}
TASK [fedora.linux_system_roles.podman : Gather the package facts] *************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:6
Saturday 24 August 2024 12:34:07 -0400 (0:00:00.065) 0:00:16.258 *******
ok: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : Enable copr if requested] *************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:10
Saturday 24 August 2024 12:34:09 -0400 (0:00:01.878) 0:00:18.137 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Ensure required packages are installed] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:14
Saturday 24 August 2024 12:34:09 -0400 (0:00:00.050) 0:00:18.188 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Notify user that reboot is needed to apply changes] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:28
Saturday 24 August 2024 12:34:09 -0400 (0:00:00.081) 0:00:18.270 *******
skipping: [managed_node1] => {}
TASK [fedora.linux_system_roles.podman : Reboot transactional update systems] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:33
Saturday 24 August 2024 12:34:09 -0400 (0:00:00.094) 0:00:18.364 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Fail if reboot is needed and not set] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:38
Saturday 24 August 2024 12:34:09 -0400 (0:00:00.089) 0:00:18.453 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Get podman version] *******************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:46
Saturday 24 August 2024 12:34:09 -0400 (0:00:00.073) 0:00:18.527 *******
ok: [managed_node1] => {
"changed": false,
"cmd": [
"podman",
"--version"
],
"delta": "0:00:00.027889",
"end": "2024-08-24 12:34:10.333993",
"rc": 0,
"start": "2024-08-24 12:34:10.306104"
}
STDOUT:
podman version 4.9.4-dev
TASK [fedora.linux_system_roles.podman : Set podman version] *******************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:52
Saturday 24 August 2024 12:34:10 -0400 (0:00:00.494) 0:00:19.021 *******
ok: [managed_node1] => {
"ansible_facts": {
"podman_version": "4.9.4-dev"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Podman package version must be 4.2 or later] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:56
Saturday 24 August 2024 12:34:10 -0400 (0:00:00.072) 0:00:19.094 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Podman package version must be 4.4 or later for quadlet, secrets] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:63
Saturday 24 August 2024 12:34:10 -0400 (0:00:00.051) 0:00:19.145 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
META: end_host conditional evaluated to false, continuing execution for managed_node1
TASK [fedora.linux_system_roles.podman : Check user and group information] *****
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:80
Saturday 24 August 2024 12:34:10 -0400 (0:00:00.264) 0:00:19.410 *******
included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for managed_node1
TASK [fedora.linux_system_roles.podman : Get user information] *****************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2
Saturday 24 August 2024 12:34:10 -0400 (0:00:00.087) 0:00:19.497 *******
ok: [managed_node1] => {
"ansible_facts": {
"getent_passwd": {
"root": [
"x",
"0",
"0",
"root",
"/root",
"/bin/bash"
]
}
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Fail if user does not exist] **********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9
Saturday 24 August 2024 12:34:11 -0400 (0:00:00.848) 0:00:20.346 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set group for podman user] ************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16
Saturday 24 August 2024 12:34:11 -0400 (0:00:00.110) 0:00:20.456 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_group": "0"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Get group information] ****************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:28
Saturday 24 August 2024 12:34:11 -0400 (0:00:00.112) 0:00:20.569 *******
ok: [managed_node1] => {
"ansible_facts": {
"getent_group": {
"root": [
"x",
"0",
""
]
}
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set group name] ***********************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:35
Saturday 24 August 2024 12:34:12 -0400 (0:00:00.670) 0:00:21.239 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_group_name": "root"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : See if getsubids exists] **************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:39
Saturday 24 August 2024 12:34:12 -0400 (0:00:00.066) 0:00:21.305 *******
ok: [managed_node1] => {
"changed": false,
"stat": {
"atime": 1724516929.7168462,
"attr_flags": "",
"attributes": [],
"block_size": 4096,
"blocks": 32,
"charset": "binary",
"checksum": "bb5b46ffbafcaa8c4021f3c8b3cb8594f48ef34b",
"ctime": 1724516896.725572,
"dev": 51713,
"device_type": 0,
"executable": true,
"exists": true,
"gid": 0,
"gr_name": "root",
"inode": 6986657,
"isblk": false,
"ischr": false,
"isdir": false,
"isfifo": false,
"isgid": false,
"islnk": false,
"isreg": true,
"issock": false,
"isuid": false,
"mimetype": "application/x-sharedlib",
"mode": "0755",
"mtime": 1700557386.0,
"nlink": 1,
"path": "/usr/bin/getsubids",
"pw_name": "root",
"readable": true,
"rgrp": true,
"roth": true,
"rusr": true,
"size": 12640,
"uid": 0,
"version": "1255679238",
"wgrp": false,
"woth": false,
"writeable": true,
"wusr": true,
"xgrp": true,
"xoth": true,
"xusr": true
}
}
TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:50
Saturday 24 August 2024 12:34:13 -0400 (0:00:00.469) 0:00:21.774 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Check group with getsubids] ***********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:55
Saturday 24 August 2024 12:34:13 -0400 (0:00:00.037) 0:00:21.812 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ******
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:60
Saturday 24 August 2024 12:34:13 -0400 (0:00:00.044) 0:00:21.856 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Get subuid file] **********************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:74
Saturday 24 August 2024 12:34:13 -0400 (0:00:00.044) 0:00:21.901 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Get subgid file] **********************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:79
Saturday 24 August 2024 12:34:13 -0400 (0:00:00.035) 0:00:21.936 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ******
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:84
Saturday 24 August 2024 12:34:13 -0400 (0:00:00.032) 0:00:21.968 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ******
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:94
Saturday 24 August 2024 12:34:13 -0400 (0:00:00.031) 0:00:21.999 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] *****
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:101
Saturday 24 August 2024 12:34:13 -0400 (0:00:00.034) 0:00:22.033 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set config file paths] ****************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:86
Saturday 24 August 2024 12:34:13 -0400 (0:00:00.034) 0:00:22.067 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_container_conf_file": "/etc/containers/containers.conf.d/50-systemroles.conf",
"__podman_policy_json_file": "/etc/containers/policy.json",
"__podman_registries_conf_file": "/etc/containers/registries.conf.d/50-systemroles.conf",
"__podman_storage_conf_file": "/etc/containers/storage.conf"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Handle container.conf.d] **************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:95
Saturday 24 August 2024 12:34:13 -0400 (0:00:00.067) 0:00:22.135 *******
included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_container_conf_d.yml for managed_node1
TASK [fedora.linux_system_roles.podman : Ensure containers.d exists] ***********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_container_conf_d.yml:5
Saturday 24 August 2024 12:34:13 -0400 (0:00:00.062) 0:00:22.197 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Update container config file] *********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_container_conf_d.yml:13
Saturday 24 August 2024 12:34:13 -0400 (0:00:00.037) 0:00:22.235 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Handle registries.conf.d] *************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:98
Saturday 24 August 2024 12:34:13 -0400 (0:00:00.052) 0:00:22.287 *******
included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_registries_conf_d.yml for managed_node1
TASK [fedora.linux_system_roles.podman : Ensure registries.d exists] ***********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_registries_conf_d.yml:5
Saturday 24 August 2024 12:34:13 -0400 (0:00:00.128) 0:00:22.416 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Update registries config file] ********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_registries_conf_d.yml:13
Saturday 24 August 2024 12:34:13 -0400 (0:00:00.041) 0:00:22.458 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Handle storage.conf] ******************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:101
Saturday 24 August 2024 12:34:13 -0400 (0:00:00.040) 0:00:22.498 *******
included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_storage_conf.yml for managed_node1
TASK [fedora.linux_system_roles.podman : Ensure storage.conf parent dir exists] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_storage_conf.yml:5
Saturday 24 August 2024 12:34:13 -0400 (0:00:00.062) 0:00:22.561 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Update storage config file] ***********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_storage_conf.yml:13
Saturday 24 August 2024 12:34:13 -0400 (0:00:00.032) 0:00:22.594 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Handle policy.json] *******************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:104
Saturday 24 August 2024 12:34:14 -0400 (0:00:00.033) 0:00:22.628 *******
included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml for managed_node1
TASK [fedora.linux_system_roles.podman : Ensure policy.json parent dir exists] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:6
Saturday 24 August 2024 12:34:14 -0400 (0:00:00.078) 0:00:22.706 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Stat the policy.json file] ************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:14
Saturday 24 August 2024 12:34:14 -0400 (0:00:00.051) 0:00:22.758 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Get the existing policy.json] *********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:19
Saturday 24 August 2024 12:34:14 -0400 (0:00:00.049) 0:00:22.808 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Write new policy.json file] ***********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:25
Saturday 24 August 2024 12:34:14 -0400 (0:00:00.049) 0:00:22.857 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [Manage firewall for specified ports] *************************************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:110
Saturday 24 August 2024 12:34:14 -0400 (0:00:00.038) 0:00:22.895 *******
TASK [fedora.linux_system_roles.firewall : Setup firewalld] ********************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:2
Saturday 24 August 2024 12:34:14 -0400 (0:00:00.117) 0:00:23.012 *******
included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml for managed_node1
TASK [fedora.linux_system_roles.firewall : Ensure ansible_facts used by role] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:2
Saturday 24 August 2024 12:34:14 -0400 (0:00:00.059) 0:00:23.072 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.firewall : Check if system is ostree] **********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:10
Saturday 24 August 2024 12:34:14 -0400 (0:00:00.111) 0:00:23.184 *******
ok: [managed_node1] => {
"changed": false,
"stat": {
"exists": false
}
}
TASK [fedora.linux_system_roles.firewall : Set flag to indicate system is ostree] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:15
Saturday 24 August 2024 12:34:14 -0400 (0:00:00.430) 0:00:23.614 *******
ok: [managed_node1] => {
"ansible_facts": {
"__firewall_is_ostree": false
},
"changed": false
}
TASK [fedora.linux_system_roles.firewall : Check if transactional-update exists in /sbin] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:22
Saturday 24 August 2024 12:34:15 -0400 (0:00:00.062) 0:00:23.676 *******
ok: [managed_node1] => {
"changed": false,
"stat": {
"exists": false
}
}
TASK [fedora.linux_system_roles.firewall : Set flag if transactional-update exists] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:27
Saturday 24 August 2024 12:34:15 -0400 (0:00:00.428) 0:00:24.105 *******
ok: [managed_node1] => {
"ansible_facts": {
"__firewall_is_transactional": false
},
"changed": false
}
TASK [fedora.linux_system_roles.firewall : Install firewalld] ******************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:31
Saturday 24 August 2024 12:34:15 -0400 (0:00:00.060) 0:00:24.166 *******
ok: [managed_node1] => {
"changed": false,
"rc": 0,
"results": []
}
MSG:
Nothing to do
lsrpackages: firewalld
TASK [fedora.linux_system_roles.firewall : Notify user that reboot is needed to apply changes] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:43
Saturday 24 August 2024 12:34:18 -0400 (0:00:02.791) 0:00:26.957 *******
skipping: [managed_node1] => {}
TASK [fedora.linux_system_roles.firewall : Reboot transactional update systems] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:48
Saturday 24 August 2024 12:34:18 -0400 (0:00:00.066) 0:00:27.024 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.firewall : Fail if reboot is needed and not set] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:53
Saturday 24 August 2024 12:34:18 -0400 (0:00:00.033) 0:00:27.057 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.firewall : Collect service facts] **************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:5
Saturday 24 August 2024 12:34:18 -0400 (0:00:00.034) 0:00:27.092 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.firewall : Attempt to stop and disable conflicting services] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:9
Saturday 24 August 2024 12:34:18 -0400 (0:00:00.038) 0:00:27.131 *******
skipping: [managed_node1] => (item=nftables) => {
"ansible_loop_var": "item",
"changed": false,
"item": "nftables",
"skip_reason": "Conditional result was False"
}
skipping: [managed_node1] => (item=iptables) => {
"ansible_loop_var": "item",
"changed": false,
"item": "iptables",
"skip_reason": "Conditional result was False"
}
skipping: [managed_node1] => (item=ufw) => {
"ansible_loop_var": "item",
"changed": false,
"item": "ufw",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.firewall : Unmask firewalld service] ***********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:22
Saturday 24 August 2024 12:34:18 -0400 (0:00:00.069) 0:00:27.200 *******
ok: [managed_node1] => {
"changed": false,
"name": "firewalld",
"status": {
"ActiveEnterTimestampMonotonic": "0",
"ActiveExitTimestampMonotonic": "0",
"ActiveState": "inactive",
"After": "polkit.service system.slice dbus.service basic.target dbus.socket sysinit.target",
"AllowIsolate": "no",
"AllowedCPUs": "",
"AllowedMemoryNodes": "",
"AmbientCapabilities": "",
"AssertResult": "no",
"AssertTimestampMonotonic": "0",
"Before": "shutdown.target network-pre.target",
"BlockIOAccounting": "no",
"BlockIOWeight": "[not set]",
"BusName": "org.fedoraproject.FirewallD1",
"CPUAccounting": "no",
"CPUAffinity": "",
"CPUAffinityFromNUMA": "no",
"CPUQuotaPerSecUSec": "infinity",
"CPUQuotaPeriodUSec": "infinity",
"CPUSchedulingPolicy": "0",
"CPUSchedulingPriority": "0",
"CPUSchedulingResetOnFork": "no",
"CPUShares": "[not set]",
"CPUUsageNSec": "[not set]",
"CPUWeight": "[not set]",
"CacheDirectoryMode": "0755",
"CanFreeze": "yes",
"CanIsolate": "no",
"CanReload": "yes",
"CanStart": "yes",
"CanStop": "yes",
"CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf",
"CollectMode": "inactive",
"ConditionResult": "no",
"ConditionTimestampMonotonic": "0",
"ConfigurationDirectoryMode": "0755",
"Conflicts": "shutdown.target ebtables.service iptables.service ip6tables.service ipset.service nftables.service",
"ControlPID": "0",
"DefaultDependencies": "yes",
"DefaultMemoryLow": "0",
"DefaultMemoryMin": "0",
"Delegate": "no",
"Description": "firewalld - dynamic firewall daemon",
"DevicePolicy": "auto",
"Documentation": "man:firewalld(1)",
"DynamicUser": "no",
"EffectiveCPUs": "",
"EffectiveMemoryNodes": "",
"EnvironmentFiles": "/etc/sysconfig/firewalld (ignore_errors=yes)",
"ExecMainCode": "0",
"ExecMainExitTimestampMonotonic": "0",
"ExecMainPID": "0",
"ExecMainStartTimestampMonotonic": "0",
"ExecMainStatus": "0",
"ExecReload": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"ExecStart": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"FailureAction": "none",
"FileDescriptorStoreMax": "0",
"FragmentPath": "/usr/lib/systemd/system/firewalld.service",
"FreezerState": "running",
"GID": "[not set]",
"GuessMainPID": "yes",
"IOAccounting": "no",
"IOSchedulingClass": "0",
"IOSchedulingPriority": "0",
"IOWeight": "[not set]",
"IPAccounting": "no",
"IPEgressBytes": "18446744073709551615",
"IPEgressPackets": "18446744073709551615",
"IPIngressBytes": "18446744073709551615",
"IPIngressPackets": "18446744073709551615",
"Id": "firewalld.service",
"IgnoreOnIsolate": "no",
"IgnoreSIGPIPE": "yes",
"InactiveEnterTimestampMonotonic": "0",
"InactiveExitTimestampMonotonic": "0",
"JobRunningTimeoutUSec": "infinity",
"JobTimeoutAction": "none",
"JobTimeoutUSec": "infinity",
"KeyringMode": "private",
"KillMode": "mixed",
"KillSignal": "15",
"LimitAS": "infinity",
"LimitASSoft": "infinity",
"LimitCORE": "infinity",
"LimitCORESoft": "0",
"LimitCPU": "infinity",
"LimitCPUSoft": "infinity",
"LimitDATA": "infinity",
"LimitDATASoft": "infinity",
"LimitFSIZE": "infinity",
"LimitFSIZESoft": "infinity",
"LimitLOCKS": "infinity",
"LimitLOCKSSoft": "infinity",
"LimitMEMLOCK": "65536",
"LimitMEMLOCKSoft": "65536",
"LimitMSGQUEUE": "819200",
"LimitMSGQUEUESoft": "819200",
"LimitNICE": "0",
"LimitNICESoft": "0",
"LimitNOFILE": "262144",
"LimitNOFILESoft": "1024",
"LimitNPROC": "14003",
"LimitNPROCSoft": "14003",
"LimitRSS": "infinity",
"LimitRSSSoft": "infinity",
"LimitRTPRIO": "0",
"LimitRTPRIOSoft": "0",
"LimitRTTIME": "infinity",
"LimitRTTIMESoft": "infinity",
"LimitSIGPENDING": "14003",
"LimitSIGPENDINGSoft": "14003",
"LimitSTACK": "infinity",
"LimitSTACKSoft": "8388608",
"LoadState": "loaded",
"LockPersonality": "no",
"LogLevelMax": "-1",
"LogRateLimitBurst": "0",
"LogRateLimitIntervalUSec": "0",
"LogsDirectoryMode": "0755",
"MainPID": "0",
"MemoryAccounting": "yes",
"MemoryCurrent": "[not set]",
"MemoryDenyWriteExecute": "no",
"MemoryHigh": "infinity",
"MemoryLimit": "infinity",
"MemoryLow": "0",
"MemoryMax": "infinity",
"MemoryMin": "0",
"MemorySwapMax": "infinity",
"MountAPIVFS": "no",
"MountFlags": "",
"NFileDescriptorStore": "0",
"NRestarts": "0",
"NUMAMask": "",
"NUMAPolicy": "n/a",
"Names": "firewalld.service",
"NeedDaemonReload": "no",
"Nice": "0",
"NoNewPrivileges": "no",
"NonBlocking": "no",
"NotifyAccess": "none",
"OOMScoreAdjust": "0",
"OnFailureJobMode": "replace",
"PermissionsStartOnly": "no",
"Perpetual": "no",
"PrivateDevices": "no",
"PrivateMounts": "no",
"PrivateNetwork": "no",
"PrivateTmp": "no",
"PrivateUsers": "no",
"ProtectControlGroups": "no",
"ProtectHome": "no",
"ProtectKernelModules": "no",
"ProtectKernelTunables": "no",
"ProtectSystem": "no",
"RefuseManualStart": "no",
"RefuseManualStop": "no",
"RemainAfterExit": "no",
"RemoveIPC": "no",
"Requires": "system.slice sysinit.target dbus.socket",
"Restart": "no",
"RestartUSec": "100ms",
"RestrictNamespaces": "no",
"RestrictRealtime": "no",
"RestrictSUIDSGID": "no",
"Result": "success",
"RootDirectoryStartOnly": "no",
"RuntimeDirectoryMode": "0755",
"RuntimeDirectoryPreserve": "no",
"RuntimeMaxUSec": "infinity",
"SameProcessGroup": "no",
"SecureBits": "0",
"SendSIGHUP": "no",
"SendSIGKILL": "yes",
"Slice": "system.slice",
"StandardError": "null",
"StandardInput": "null",
"StandardInputData": "",
"StandardOutput": "null",
"StartLimitAction": "none",
"StartLimitBurst": "5",
"StartLimitIntervalUSec": "10s",
"StartupBlockIOWeight": "[not set]",
"StartupCPUShares": "[not set]",
"StartupCPUWeight": "[not set]",
"StartupIOWeight": "[not set]",
"StateChangeTimestampMonotonic": "0",
"StateDirectoryMode": "0755",
"StatusErrno": "0",
"StopWhenUnneeded": "no",
"SubState": "dead",
"SuccessAction": "none",
"SyslogFacility": "3",
"SyslogLevel": "6",
"SyslogLevelPrefix": "yes",
"SyslogPriority": "30",
"SystemCallErrorNumber": "0",
"TTYReset": "no",
"TTYVHangup": "no",
"TTYVTDisallocate": "no",
"TasksAccounting": "yes",
"TasksCurrent": "[not set]",
"TasksMax": "22405",
"TimeoutStartUSec": "1min 30s",
"TimeoutStopUSec": "1min 30s",
"TimerSlackNSec": "50000",
"Transient": "no",
"Type": "dbus",
"UID": "[not set]",
"UMask": "0022",
"UnitFilePreset": "enabled",
"UnitFileState": "disabled",
"UtmpMode": "init",
"Wants": "network-pre.target",
"WatchdogTimestampMonotonic": "0",
"WatchdogUSec": "0"
}
}
TASK [fedora.linux_system_roles.firewall : Enable and start firewalld service] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:28
Saturday 24 August 2024 12:34:19 -0400 (0:00:00.592) 0:00:27.793 *******
changed: [managed_node1] => {
"changed": true,
"enabled": true,
"name": "firewalld",
"state": "started",
"status": {
"ActiveEnterTimestampMonotonic": "0",
"ActiveExitTimestampMonotonic": "0",
"ActiveState": "inactive",
"After": "polkit.service sysinit.target system.slice basic.target dbus.service dbus.socket",
"AllowIsolate": "no",
"AllowedCPUs": "",
"AllowedMemoryNodes": "",
"AmbientCapabilities": "",
"AssertResult": "no",
"AssertTimestampMonotonic": "0",
"Before": "shutdown.target network-pre.target",
"BlockIOAccounting": "no",
"BlockIOWeight": "[not set]",
"BusName": "org.fedoraproject.FirewallD1",
"CPUAccounting": "no",
"CPUAffinity": "",
"CPUAffinityFromNUMA": "no",
"CPUQuotaPerSecUSec": "infinity",
"CPUQuotaPeriodUSec": "infinity",
"CPUSchedulingPolicy": "0",
"CPUSchedulingPriority": "0",
"CPUSchedulingResetOnFork": "no",
"CPUShares": "[not set]",
"CPUUsageNSec": "[not set]",
"CPUWeight": "[not set]",
"CacheDirectoryMode": "0755",
"CanFreeze": "yes",
"CanIsolate": "no",
"CanReload": "yes",
"CanStart": "yes",
"CanStop": "yes",
"CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf",
"CollectMode": "inactive",
"ConditionResult": "no",
"ConditionTimestampMonotonic": "0",
"ConfigurationDirectoryMode": "0755",
"Conflicts": "shutdown.target iptables.service ebtables.service ip6tables.service nftables.service ipset.service",
"ControlPID": "0",
"DefaultDependencies": "yes",
"DefaultMemoryLow": "0",
"DefaultMemoryMin": "0",
"Delegate": "no",
"Description": "firewalld - dynamic firewall daemon",
"DevicePolicy": "auto",
"Documentation": "man:firewalld(1)",
"DynamicUser": "no",
"EffectiveCPUs": "",
"EffectiveMemoryNodes": "",
"EnvironmentFiles": "/etc/sysconfig/firewalld (ignore_errors=yes)",
"ExecMainCode": "0",
"ExecMainExitTimestampMonotonic": "0",
"ExecMainPID": "0",
"ExecMainStartTimestampMonotonic": "0",
"ExecMainStatus": "0",
"ExecReload": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"ExecStart": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"FailureAction": "none",
"FileDescriptorStoreMax": "0",
"FragmentPath": "/usr/lib/systemd/system/firewalld.service",
"FreezerState": "running",
"GID": "[not set]",
"GuessMainPID": "yes",
"IOAccounting": "no",
"IOSchedulingClass": "0",
"IOSchedulingPriority": "0",
"IOWeight": "[not set]",
"IPAccounting": "no",
"IPEgressBytes": "18446744073709551615",
"IPEgressPackets": "18446744073709551615",
"IPIngressBytes": "18446744073709551615",
"IPIngressPackets": "18446744073709551615",
"Id": "firewalld.service",
"IgnoreOnIsolate": "no",
"IgnoreSIGPIPE": "yes",
"InactiveEnterTimestampMonotonic": "0",
"InactiveExitTimestampMonotonic": "0",
"JobRunningTimeoutUSec": "infinity",
"JobTimeoutAction": "none",
"JobTimeoutUSec": "infinity",
"KeyringMode": "private",
"KillMode": "mixed",
"KillSignal": "15",
"LimitAS": "infinity",
"LimitASSoft": "infinity",
"LimitCORE": "infinity",
"LimitCORESoft": "0",
"LimitCPU": "infinity",
"LimitCPUSoft": "infinity",
"LimitDATA": "infinity",
"LimitDATASoft": "infinity",
"LimitFSIZE": "infinity",
"LimitFSIZESoft": "infinity",
"LimitLOCKS": "infinity",
"LimitLOCKSSoft": "infinity",
"LimitMEMLOCK": "65536",
"LimitMEMLOCKSoft": "65536",
"LimitMSGQUEUE": "819200",
"LimitMSGQUEUESoft": "819200",
"LimitNICE": "0",
"LimitNICESoft": "0",
"LimitNOFILE": "262144",
"LimitNOFILESoft": "1024",
"LimitNPROC": "14003",
"LimitNPROCSoft": "14003",
"LimitRSS": "infinity",
"LimitRSSSoft": "infinity",
"LimitRTPRIO": "0",
"LimitRTPRIOSoft": "0",
"LimitRTTIME": "infinity",
"LimitRTTIMESoft": "infinity",
"LimitSIGPENDING": "14003",
"LimitSIGPENDINGSoft": "14003",
"LimitSTACK": "infinity",
"LimitSTACKSoft": "8388608",
"LoadState": "loaded",
"LockPersonality": "no",
"LogLevelMax": "-1",
"LogRateLimitBurst": "0",
"LogRateLimitIntervalUSec": "0",
"LogsDirectoryMode": "0755",
"MainPID": "0",
"MemoryAccounting": "yes",
"MemoryCurrent": "[not set]",
"MemoryDenyWriteExecute": "no",
"MemoryHigh": "infinity",
"MemoryLimit": "infinity",
"MemoryLow": "0",
"MemoryMax": "infinity",
"MemoryMin": "0",
"MemorySwapMax": "infinity",
"MountAPIVFS": "no",
"MountFlags": "",
"NFileDescriptorStore": "0",
"NRestarts": "0",
"NUMAMask": "",
"NUMAPolicy": "n/a",
"Names": "firewalld.service",
"NeedDaemonReload": "no",
"Nice": "0",
"NoNewPrivileges": "no",
"NonBlocking": "no",
"NotifyAccess": "none",
"OOMScoreAdjust": "0",
"OnFailureJobMode": "replace",
"PermissionsStartOnly": "no",
"Perpetual": "no",
"PrivateDevices": "no",
"PrivateMounts": "no",
"PrivateNetwork": "no",
"PrivateTmp": "no",
"PrivateUsers": "no",
"ProtectControlGroups": "no",
"ProtectHome": "no",
"ProtectKernelModules": "no",
"ProtectKernelTunables": "no",
"ProtectSystem": "no",
"RefuseManualStart": "no",
"RefuseManualStop": "no",
"RemainAfterExit": "no",
"RemoveIPC": "no",
"Requires": "dbus.socket system.slice sysinit.target",
"Restart": "no",
"RestartUSec": "100ms",
"RestrictNamespaces": "no",
"RestrictRealtime": "no",
"RestrictSUIDSGID": "no",
"Result": "success",
"RootDirectoryStartOnly": "no",
"RuntimeDirectoryMode": "0755",
"RuntimeDirectoryPreserve": "no",
"RuntimeMaxUSec": "infinity",
"SameProcessGroup": "no",
"SecureBits": "0",
"SendSIGHUP": "no",
"SendSIGKILL": "yes",
"Slice": "system.slice",
"StandardError": "null",
"StandardInput": "null",
"StandardInputData": "",
"StandardOutput": "null",
"StartLimitAction": "none",
"StartLimitBurst": "5",
"StartLimitIntervalUSec": "10s",
"StartupBlockIOWeight": "[not set]",
"StartupCPUShares": "[not set]",
"StartupCPUWeight": "[not set]",
"StartupIOWeight": "[not set]",
"StateChangeTimestampMonotonic": "0",
"StateDirectoryMode": "0755",
"StatusErrno": "0",
"StopWhenUnneeded": "no",
"SubState": "dead",
"SuccessAction": "none",
"SyslogFacility": "3",
"SyslogLevel": "6",
"SyslogLevelPrefix": "yes",
"SyslogPriority": "30",
"SystemCallErrorNumber": "0",
"TTYReset": "no",
"TTYVHangup": "no",
"TTYVTDisallocate": "no",
"TasksAccounting": "yes",
"TasksCurrent": "[not set]",
"TasksMax": "22405",
"TimeoutStartUSec": "1min 30s",
"TimeoutStopUSec": "1min 30s",
"TimerSlackNSec": "50000",
"Transient": "no",
"Type": "dbus",
"UID": "[not set]",
"UMask": "0022",
"UnitFilePreset": "enabled",
"UnitFileState": "disabled",
"UtmpMode": "init",
"Wants": "network-pre.target",
"WatchdogTimestampMonotonic": "0",
"WatchdogUSec": "0"
}
}
TASK [fedora.linux_system_roles.firewall : Check if previous replaced is defined] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:34
Saturday 24 August 2024 12:34:20 -0400 (0:00:01.206) 0:00:28.999 *******
ok: [managed_node1] => {
"ansible_facts": {
"__firewall_previous_replaced": false,
"__firewall_python_cmd": "/usr/libexec/platform-python",
"__firewall_report_changed": true
},
"changed": false
}
TASK [fedora.linux_system_roles.firewall : Get config files, checksums before and remove] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:43
Saturday 24 August 2024 12:34:20 -0400 (0:00:00.049) 0:00:29.049 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.firewall : Tell firewall module it is able to report changed] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:55
Saturday 24 August 2024 12:34:20 -0400 (0:00:00.054) 0:00:29.104 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.firewall : Configure firewall] *****************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:71
Saturday 24 August 2024 12:34:20 -0400 (0:00:00.057) 0:00:29.162 *******
changed: [managed_node1] => (item={'port': '8000/tcp', 'state': 'enabled'}) => {
"__firewall_changed": true,
"ansible_loop_var": "item",
"changed": true,
"item": {
"port": "8000/tcp",
"state": "enabled"
}
}
changed: [managed_node1] => (item={'port': '9000/tcp', 'state': 'enabled'}) => {
"__firewall_changed": true,
"ansible_loop_var": "item",
"changed": true,
"item": {
"port": "9000/tcp",
"state": "enabled"
}
}
TASK [fedora.linux_system_roles.firewall : Gather firewall config information] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:120
Saturday 24 August 2024 12:34:22 -0400 (0:00:01.498) 0:00:30.660 *******
skipping: [managed_node1] => (item={'port': '8000/tcp', 'state': 'enabled'}) => {
"ansible_loop_var": "item",
"changed": false,
"item": {
"port": "8000/tcp",
"state": "enabled"
},
"skip_reason": "Conditional result was False"
}
skipping: [managed_node1] => (item={'port': '9000/tcp', 'state': 'enabled'}) => {
"ansible_loop_var": "item",
"changed": false,
"item": {
"port": "9000/tcp",
"state": "enabled"
},
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.firewall : Update firewalld_config fact] *******
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:130
Saturday 24 August 2024 12:34:22 -0400 (0:00:00.090) 0:00:30.750 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.firewall : Gather firewall config if no arguments] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:139
Saturday 24 August 2024 12:34:22 -0400 (0:00:00.060) 0:00:30.811 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.firewall : Update firewalld_config fact] *******
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:144
Saturday 24 August 2024 12:34:22 -0400 (0:00:00.054) 0:00:30.865 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.firewall : Get config files, checksums after] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:153
Saturday 24 August 2024 12:34:22 -0400 (0:00:00.057) 0:00:30.922 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.firewall : Calculate what has changed] *********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:163
Saturday 24 August 2024 12:34:22 -0400 (0:00:00.053) 0:00:30.976 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.firewall : Show diffs] *************************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:169
Saturday 24 August 2024 12:34:22 -0400 (0:00:00.050) 0:00:31.026 *******
skipping: [managed_node1] => {}
TASK [Manage selinux for specified ports] **************************************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:117
Saturday 24 August 2024 12:34:22 -0400 (0:00:00.053) 0:00:31.080 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Keep track of users that need to cancel linger] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:124
Saturday 24 August 2024 12:34:22 -0400 (0:00:00.052) 0:00:31.132 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_cancel_user_linger": []
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Handle certs.d files - present] *******
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:128
Saturday 24 August 2024 12:34:22 -0400 (0:00:00.049) 0:00:31.182 *******
skipping: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : Handle credential files - present] ****
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:137
Saturday 24 August 2024 12:34:22 -0400 (0:00:00.034) 0:00:31.217 *******
skipping: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : Handle secrets] ***********************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:146
Saturday 24 August 2024 12:34:22 -0400 (0:00:00.033) 0:00:31.251 *******
included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml for managed_node1
included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml for managed_node1
included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml for managed_node1
TASK [fedora.linux_system_roles.podman : Set variables part 1] *****************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:3
Saturday 24 August 2024 12:34:22 -0400 (0:00:00.142) 0:00:31.393 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_user": "root"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set variables part 2] *****************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:7
Saturday 24 August 2024 12:34:22 -0400 (0:00:00.047) 0:00:31.441 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_rootless": false,
"__podman_xdg_runtime_dir": "/run/user/0"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Manage linger] ************************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:13
Saturday 24 August 2024 12:34:22 -0400 (0:00:00.066) 0:00:31.508 *******
included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for managed_node1
TASK [fedora.linux_system_roles.podman : Enable linger if needed] **************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12
Saturday 24 August 2024 12:34:23 -0400 (0:00:00.128) 0:00:31.637 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18
Saturday 24 August 2024 12:34:23 -0400 (0:00:00.050) 0:00:31.687 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22
Saturday 24 August 2024 12:34:23 -0400 (0:00:00.040) 0:00:31.727 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Stat XDG_RUNTIME_DIR] *****************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:18
Saturday 24 August 2024 12:34:23 -0400 (0:00:00.037) 0:00:31.765 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Manage each secret] *******************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:34
Saturday 24 August 2024 12:34:23 -0400 (0:00:00.038) 0:00:31.804 *******
[WARNING]: Using a variable for a task's 'args' is unsafe in some situations
(see
https://docs.ansible.com/ansible/devel/reference_appendices/faq.html#argsplat-
unsafe)
changed: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": true
}
TASK [fedora.linux_system_roles.podman : Set variables part 1] *****************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:3
Saturday 24 August 2024 12:34:23 -0400 (0:00:00.632) 0:00:32.436 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_user": "root"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set variables part 2] *****************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:7
Saturday 24 August 2024 12:34:23 -0400 (0:00:00.051) 0:00:32.487 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_rootless": false,
"__podman_xdg_runtime_dir": "/run/user/0"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Manage linger] ************************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:13
Saturday 24 August 2024 12:34:23 -0400 (0:00:00.046) 0:00:32.533 *******
included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for managed_node1
TASK [fedora.linux_system_roles.podman : Enable linger if needed] **************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12
Saturday 24 August 2024 12:34:23 -0400 (0:00:00.068) 0:00:32.602 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18
Saturday 24 August 2024 12:34:24 -0400 (0:00:00.038) 0:00:32.640 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22
Saturday 24 August 2024 12:34:24 -0400 (0:00:00.033) 0:00:32.674 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Stat XDG_RUNTIME_DIR] *****************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:18
Saturday 24 August 2024 12:34:24 -0400 (0:00:00.032) 0:00:32.707 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Manage each secret] *******************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:34
Saturday 24 August 2024 12:34:24 -0400 (0:00:00.033) 0:00:32.740 *******
[WARNING]: Using a variable for a task's 'args' is unsafe in some situations
(see
https://docs.ansible.com/ansible/devel/reference_appendices/faq.html#argsplat-
unsafe)
changed: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": true
}
TASK [fedora.linux_system_roles.podman : Set variables part 1] *****************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:3
Saturday 24 August 2024 12:34:24 -0400 (0:00:00.555) 0:00:33.296 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_user": "root"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set variables part 2] *****************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:7
Saturday 24 August 2024 12:34:24 -0400 (0:00:00.033) 0:00:33.329 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_rootless": false,
"__podman_xdg_runtime_dir": "/run/user/0"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Manage linger] ************************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:13
Saturday 24 August 2024 12:34:24 -0400 (0:00:00.041) 0:00:33.371 *******
included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for managed_node1
TASK [fedora.linux_system_roles.podman : Enable linger if needed] **************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12
Saturday 24 August 2024 12:34:24 -0400 (0:00:00.066) 0:00:33.438 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18
Saturday 24 August 2024 12:34:24 -0400 (0:00:00.053) 0:00:33.492 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22
Saturday 24 August 2024 12:34:24 -0400 (0:00:00.051) 0:00:33.543 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Stat XDG_RUNTIME_DIR] *****************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:18
Saturday 24 August 2024 12:34:24 -0400 (0:00:00.053) 0:00:33.596 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Manage each secret] *******************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:34
Saturday 24 August 2024 12:34:25 -0400 (0:00:00.051) 0:00:33.648 *******
[WARNING]: Using a variable for a task's 'args' is unsafe in some situations
(see
https://docs.ansible.com/ansible/devel/reference_appendices/faq.html#argsplat-
unsafe)
changed: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": true
}
TASK [fedora.linux_system_roles.podman : Handle Kubernetes specifications] *****
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:153
Saturday 24 August 2024 12:34:25 -0400 (0:00:00.556) 0:00:34.205 *******
skipping: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : Handle Quadlet specifications] ********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:160
Saturday 24 August 2024 12:34:25 -0400 (0:00:00.037) 0:00:34.242 *******
included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for managed_node1
included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for managed_node1
included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for managed_node1
included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for managed_node1
included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for managed_node1
included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for managed_node1
TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14
Saturday 24 August 2024 12:34:25 -0400 (0:00:00.142) 0:00:34.384 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_quadlet_file_src": "quadlet-demo.network",
"__podman_quadlet_spec": {},
"__podman_quadlet_str": "[Network]\nSubnet=192.168.30.0/24\nGateway=192.168.30.1\nLabel=app=wordpress",
"__podman_quadlet_template_src": ""
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25
Saturday 24 August 2024 12:34:25 -0400 (0:00:00.043) 0:00:34.428 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_continue_if_pull_fails": false,
"__podman_pull_image": true,
"__podman_state": "created",
"__podman_systemd_unit_scope": "",
"__podman_user": "root"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] *****
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35
Saturday 24 August 2024 12:34:25 -0400 (0:00:00.040) 0:00:34.469 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48
Saturday 24 August 2024 12:34:25 -0400 (0:00:00.072) 0:00:34.541 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_quadlet_name": "quadlet-demo",
"__podman_quadlet_type": "network",
"__podman_rootless": false
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Check user and group information] *****
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57
Saturday 24 August 2024 12:34:25 -0400 (0:00:00.046) 0:00:34.588 *******
included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for managed_node1
TASK [fedora.linux_system_roles.podman : Get user information] *****************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2
Saturday 24 August 2024 12:34:26 -0400 (0:00:00.062) 0:00:34.650 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Fail if user does not exist] **********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9
Saturday 24 August 2024 12:34:26 -0400 (0:00:00.039) 0:00:34.690 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set group for podman user] ************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16
Saturday 24 August 2024 12:34:26 -0400 (0:00:00.041) 0:00:34.731 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_group": "0"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Get group information] ****************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:28
Saturday 24 August 2024 12:34:26 -0400 (0:00:00.044) 0:00:34.776 *******
ok: [managed_node1] => {
"ansible_facts": {
"getent_group": {
"root": [
"x",
"0",
""
]
}
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set group name] ***********************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:35
Saturday 24 August 2024 12:34:26 -0400 (0:00:00.377) 0:00:35.153 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_group_name": "root"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : See if getsubids exists] **************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:39
Saturday 24 August 2024 12:34:26 -0400 (0:00:00.039) 0:00:35.193 *******
ok: [managed_node1] => {
"changed": false,
"stat": {
"atime": 1724516929.7168462,
"attr_flags": "",
"attributes": [],
"block_size": 4096,
"blocks": 32,
"charset": "binary",
"checksum": "bb5b46ffbafcaa8c4021f3c8b3cb8594f48ef34b",
"ctime": 1724516896.725572,
"dev": 51713,
"device_type": 0,
"executable": true,
"exists": true,
"gid": 0,
"gr_name": "root",
"inode": 6986657,
"isblk": false,
"ischr": false,
"isdir": false,
"isfifo": false,
"isgid": false,
"islnk": false,
"isreg": true,
"issock": false,
"isuid": false,
"mimetype": "application/x-sharedlib",
"mode": "0755",
"mtime": 1700557386.0,
"nlink": 1,
"path": "/usr/bin/getsubids",
"pw_name": "root",
"readable": true,
"rgrp": true,
"roth": true,
"rusr": true,
"size": 12640,
"uid": 0,
"version": "1255679238",
"wgrp": false,
"woth": false,
"writeable": true,
"wusr": true,
"xgrp": true,
"xoth": true,
"xusr": true
}
}
TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:50
Saturday 24 August 2024 12:34:26 -0400 (0:00:00.392) 0:00:35.585 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Check group with getsubids] ***********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:55
Saturday 24 August 2024 12:34:27 -0400 (0:00:00.055) 0:00:35.641 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ******
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:60
Saturday 24 August 2024 12:34:27 -0400 (0:00:00.055) 0:00:35.697 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Get subuid file] **********************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:74
Saturday 24 August 2024 12:34:27 -0400 (0:00:00.053) 0:00:35.751 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Get subgid file] **********************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:79
Saturday 24 August 2024 12:34:27 -0400 (0:00:00.051) 0:00:35.802 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ******
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:84
Saturday 24 August 2024 12:34:27 -0400 (0:00:00.053) 0:00:35.856 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ******
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:94
Saturday 24 August 2024 12:34:27 -0400 (0:00:00.050) 0:00:35.906 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] *****
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:101
Saturday 24 August 2024 12:34:27 -0400 (0:00:00.096) 0:00:36.002 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62
Saturday 24 August 2024 12:34:27 -0400 (0:00:00.039) 0:00:36.042 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_activate_systemd_unit": true,
"__podman_images_found": [],
"__podman_kube_yamls_raw": "",
"__podman_service_name": "quadlet-demo-network.service",
"__podman_systemd_scope": "system",
"__podman_user_home_dir": "/root",
"__podman_xdg_runtime_dir": "/run/user/0"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73
Saturday 24 August 2024 12:34:27 -0400 (0:00:00.066) 0:00:36.109 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_quadlet_path": "/etc/containers/systemd"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Get kube yaml contents] ***************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77
Saturday 24 August 2024 12:34:27 -0400 (0:00:00.037) 0:00:36.146 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87
Saturday 24 August 2024 12:34:27 -0400 (0:00:00.034) 0:00:36.181 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_images": [],
"__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo.network",
"__podman_volumes": []
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 6] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:103
Saturday 24 August 2024 12:34:27 -0400 (0:00:00.075) 0:00:36.257 *******
ok: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : Cleanup quadlets] *********************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:110
Saturday 24 August 2024 12:34:27 -0400 (0:00:00.039) 0:00:36.297 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Create and update quadlets] ***********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:114
Saturday 24 August 2024 12:34:27 -0400 (0:00:00.036) 0:00:36.333 *******
included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml for managed_node1
TASK [fedora.linux_system_roles.podman : Manage linger] ************************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:2
Saturday 24 August 2024 12:34:27 -0400 (0:00:00.118) 0:00:36.452 *******
included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for managed_node1
TASK [fedora.linux_system_roles.podman : Enable linger if needed] **************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12
Saturday 24 August 2024 12:34:27 -0400 (0:00:00.091) 0:00:36.544 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18
Saturday 24 August 2024 12:34:27 -0400 (0:00:00.054) 0:00:36.598 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22
Saturday 24 August 2024 12:34:28 -0400 (0:00:00.052) 0:00:36.651 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Create host directories] **************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:7
Saturday 24 August 2024 12:34:28 -0400 (0:00:00.054) 0:00:36.705 *******
TASK [fedora.linux_system_roles.podman : Ensure container images are present] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:18
Saturday 24 August 2024 12:34:28 -0400 (0:00:00.049) 0:00:36.754 *******
skipping: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : Ensure the quadlet directory is present] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:39
Saturday 24 August 2024 12:34:28 -0400 (0:00:00.053) 0:00:36.808 *******
ok: [managed_node1] => {
"changed": false,
"gid": 0,
"group": "root",
"mode": "0755",
"owner": "root",
"path": "/etc/containers/systemd",
"secontext": "system_u:object_r:etc_t:s0",
"size": 6,
"state": "directory",
"uid": 0
}
TASK [fedora.linux_system_roles.podman : Ensure quadlet file is copied] ********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:48
Saturday 24 August 2024 12:34:28 -0400 (0:00:00.465) 0:00:37.274 *******
changed: [managed_node1] => {
"changed": true,
"checksum": "e57c08d49aff4bae8daab138d913aeddaa8682a0",
"dest": "/etc/containers/systemd/quadlet-demo.network",
"gid": 0,
"group": "root",
"md5sum": "061f3cf318cbd8ab5794bb1173831fb8",
"mode": "0644",
"owner": "root",
"secontext": "system_u:object_r:etc_t:s0",
"size": 74,
"src": "/root/.ansible/tmp/ansible-tmp-1724517268.7169204-29276-193597116746162/source",
"state": "file",
"uid": 0
}
TASK [fedora.linux_system_roles.podman : Ensure quadlet file content is present] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:58
Saturday 24 August 2024 12:34:29 -0400 (0:00:00.791) 0:00:38.066 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Ensure quadlet file is present] *******
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:70
Saturday 24 August 2024 12:34:29 -0400 (0:00:00.036) 0:00:38.102 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Reload systemctl] *********************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:82
Saturday 24 August 2024 12:34:29 -0400 (0:00:00.035) 0:00:38.138 *******
ok: [managed_node1] => {
"changed": false,
"name": null,
"status": {}
}
TASK [fedora.linux_system_roles.podman : Start service] ************************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:110
Saturday 24 August 2024 12:34:30 -0400 (0:00:00.601) 0:00:38.739 *******
changed: [managed_node1] => {
"changed": true,
"name": "quadlet-demo-network.service",
"state": "started",
"status": {
"ActiveEnterTimestampMonotonic": "0",
"ActiveExitTimestampMonotonic": "0",
"ActiveState": "inactive",
"After": "systemd-journald.socket system.slice -.mount sysinit.target basic.target",
"AllowIsolate": "no",
"AllowedCPUs": "",
"AllowedMemoryNodes": "",
"AmbientCapabilities": "",
"AssertResult": "no",
"AssertTimestampMonotonic": "0",
"Before": "shutdown.target",
"BlockIOAccounting": "no",
"BlockIOWeight": "[not set]",
"CPUAccounting": "no",
"CPUAffinity": "",
"CPUAffinityFromNUMA": "no",
"CPUQuotaPerSecUSec": "infinity",
"CPUQuotaPeriodUSec": "infinity",
"CPUSchedulingPolicy": "0",
"CPUSchedulingPriority": "0",
"CPUSchedulingResetOnFork": "no",
"CPUShares": "[not set]",
"CPUUsageNSec": "[not set]",
"CPUWeight": "[not set]",
"CacheDirectoryMode": "0755",
"CanFreeze": "yes",
"CanIsolate": "no",
"CanReload": "no",
"CanStart": "yes",
"CanStop": "yes",
"CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf",
"CollectMode": "inactive",
"ConditionResult": "no",
"ConditionTimestampMonotonic": "0",
"ConfigurationDirectoryMode": "0755",
"Conflicts": "shutdown.target",
"ControlPID": "0",
"DefaultDependencies": "yes",
"DefaultMemoryLow": "0",
"DefaultMemoryMin": "0",
"Delegate": "no",
"Description": "quadlet-demo-network.service",
"DevicePolicy": "auto",
"DynamicUser": "no",
"EffectiveCPUs": "",
"EffectiveMemoryNodes": "",
"ExecMainCode": "0",
"ExecMainExitTimestampMonotonic": "0",
"ExecMainPID": "0",
"ExecMainStartTimestampMonotonic": "0",
"ExecMainStatus": "0",
"ExecStart": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman network create --ignore --subnet=192.168.30.0/24 --gateway=192.168.30.1 --label app=wordpress systemd-quadlet-demo ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"FailureAction": "none",
"FileDescriptorStoreMax": "0",
"FragmentPath": "/run/systemd/generator/quadlet-demo-network.service",
"FreezerState": "running",
"GID": "[not set]",
"GuessMainPID": "yes",
"IOAccounting": "no",
"IOSchedulingClass": "0",
"IOSchedulingPriority": "0",
"IOWeight": "[not set]",
"IPAccounting": "no",
"IPEgressBytes": "18446744073709551615",
"IPEgressPackets": "18446744073709551615",
"IPIngressBytes": "18446744073709551615",
"IPIngressPackets": "18446744073709551615",
"Id": "quadlet-demo-network.service",
"IgnoreOnIsolate": "no",
"IgnoreSIGPIPE": "yes",
"InactiveEnterTimestampMonotonic": "0",
"InactiveExitTimestampMonotonic": "0",
"JobRunningTimeoutUSec": "infinity",
"JobTimeoutAction": "none",
"JobTimeoutUSec": "infinity",
"KeyringMode": "private",
"KillMode": "control-group",
"KillSignal": "15",
"LimitAS": "infinity",
"LimitASSoft": "infinity",
"LimitCORE": "infinity",
"LimitCORESoft": "0",
"LimitCPU": "infinity",
"LimitCPUSoft": "infinity",
"LimitDATA": "infinity",
"LimitDATASoft": "infinity",
"LimitFSIZE": "infinity",
"LimitFSIZESoft": "infinity",
"LimitLOCKS": "infinity",
"LimitLOCKSSoft": "infinity",
"LimitMEMLOCK": "65536",
"LimitMEMLOCKSoft": "65536",
"LimitMSGQUEUE": "819200",
"LimitMSGQUEUESoft": "819200",
"LimitNICE": "0",
"LimitNICESoft": "0",
"LimitNOFILE": "262144",
"LimitNOFILESoft": "1024",
"LimitNPROC": "14003",
"LimitNPROCSoft": "14003",
"LimitRSS": "infinity",
"LimitRSSSoft": "infinity",
"LimitRTPRIO": "0",
"LimitRTPRIOSoft": "0",
"LimitRTTIME": "infinity",
"LimitRTTIMESoft": "infinity",
"LimitSIGPENDING": "14003",
"LimitSIGPENDINGSoft": "14003",
"LimitSTACK": "infinity",
"LimitSTACKSoft": "8388608",
"LoadState": "loaded",
"LockPersonality": "no",
"LogLevelMax": "-1",
"LogRateLimitBurst": "0",
"LogRateLimitIntervalUSec": "0",
"LogsDirectoryMode": "0755",
"MainPID": "0",
"MemoryAccounting": "yes",
"MemoryCurrent": "[not set]",
"MemoryDenyWriteExecute": "no",
"MemoryHigh": "infinity",
"MemoryLimit": "infinity",
"MemoryLow": "0",
"MemoryMax": "infinity",
"MemoryMin": "0",
"MemorySwapMax": "infinity",
"MountAPIVFS": "no",
"MountFlags": "",
"NFileDescriptorStore": "0",
"NRestarts": "0",
"NUMAMask": "",
"NUMAPolicy": "n/a",
"Names": "quadlet-demo-network.service",
"NeedDaemonReload": "no",
"Nice": "0",
"NoNewPrivileges": "no",
"NonBlocking": "no",
"NotifyAccess": "none",
"OOMScoreAdjust": "0",
"OnFailureJobMode": "replace",
"PermissionsStartOnly": "no",
"Perpetual": "no",
"PrivateDevices": "no",
"PrivateMounts": "no",
"PrivateNetwork": "no",
"PrivateTmp": "no",
"PrivateUsers": "no",
"ProtectControlGroups": "no",
"ProtectHome": "no",
"ProtectKernelModules": "no",
"ProtectKernelTunables": "no",
"ProtectSystem": "no",
"RefuseManualStart": "no",
"RefuseManualStop": "no",
"RemainAfterExit": "yes",
"RemoveIPC": "no",
"Requires": "system.slice -.mount sysinit.target",
"RequiresMountsFor": "/run/containers",
"Restart": "no",
"RestartUSec": "100ms",
"RestrictNamespaces": "no",
"RestrictRealtime": "no",
"RestrictSUIDSGID": "no",
"Result": "success",
"RootDirectoryStartOnly": "no",
"RuntimeDirectoryMode": "0755",
"RuntimeDirectoryPreserve": "no",
"RuntimeMaxUSec": "infinity",
"SameProcessGroup": "no",
"SecureBits": "0",
"SendSIGHUP": "no",
"SendSIGKILL": "yes",
"Slice": "system.slice",
"StandardError": "inherit",
"StandardInput": "null",
"StandardInputData": "",
"StandardOutput": "journal",
"StartLimitAction": "none",
"StartLimitBurst": "5",
"StartLimitIntervalUSec": "10s",
"StartupBlockIOWeight": "[not set]",
"StartupCPUShares": "[not set]",
"StartupCPUWeight": "[not set]",
"StartupIOWeight": "[not set]",
"StateChangeTimestampMonotonic": "0",
"StateDirectoryMode": "0755",
"StatusErrno": "0",
"StopWhenUnneeded": "no",
"SubState": "dead",
"SuccessAction": "none",
"SyslogFacility": "3",
"SyslogIdentifier": "quadlet-demo-network",
"SyslogLevel": "6",
"SyslogLevelPrefix": "yes",
"SyslogPriority": "30",
"SystemCallErrorNumber": "0",
"TTYReset": "no",
"TTYVHangup": "no",
"TTYVTDisallocate": "no",
"TasksAccounting": "yes",
"TasksCurrent": "[not set]",
"TasksMax": "22405",
"TimeoutStartUSec": "infinity",
"TimeoutStopUSec": "1min 30s",
"TimerSlackNSec": "50000",
"Transient": "no",
"Type": "oneshot",
"UID": "[not set]",
"UMask": "0022",
"UnitFilePreset": "disabled",
"UnitFileState": "generated",
"UtmpMode": "init",
"WatchdogTimestampMonotonic": "0",
"WatchdogUSec": "0"
}
}
TASK [fedora.linux_system_roles.podman : Restart service] **********************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:125
Saturday 24 August 2024 12:34:30 -0400 (0:00:00.566) 0:00:39.305 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14
Saturday 24 August 2024 12:34:30 -0400 (0:00:00.035) 0:00:39.341 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_quadlet_file_src": "quadlet-demo-mysql.volume",
"__podman_quadlet_spec": {},
"__podman_quadlet_str": "[Volume]",
"__podman_quadlet_template_src": ""
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25
Saturday 24 August 2024 12:34:30 -0400 (0:00:00.051) 0:00:39.392 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_continue_if_pull_fails": false,
"__podman_pull_image": true,
"__podman_state": "created",
"__podman_systemd_unit_scope": "",
"__podman_user": "root"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] *****
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35
Saturday 24 August 2024 12:34:30 -0400 (0:00:00.058) 0:00:39.451 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48
Saturday 24 August 2024 12:34:30 -0400 (0:00:00.055) 0:00:39.506 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_quadlet_name": "quadlet-demo-mysql",
"__podman_quadlet_type": "volume",
"__podman_rootless": false
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Check user and group information] *****
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57
Saturday 24 August 2024 12:34:31 -0400 (0:00:00.127) 0:00:39.633 *******
included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for managed_node1
TASK [fedora.linux_system_roles.podman : Get user information] *****************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2
Saturday 24 August 2024 12:34:31 -0400 (0:00:00.098) 0:00:39.731 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Fail if user does not exist] **********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9
Saturday 24 August 2024 12:34:31 -0400 (0:00:00.060) 0:00:39.791 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set group for podman user] ************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16
Saturday 24 August 2024 12:34:31 -0400 (0:00:00.064) 0:00:39.856 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_group": "0"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Get group information] ****************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:28
Saturday 24 August 2024 12:34:31 -0400 (0:00:00.068) 0:00:39.924 *******
ok: [managed_node1] => {
"ansible_facts": {
"getent_group": {
"root": [
"x",
"0",
""
]
}
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set group name] ***********************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:35
Saturday 24 August 2024 12:34:31 -0400 (0:00:00.439) 0:00:40.364 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_group_name": "root"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : See if getsubids exists] **************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:39
Saturday 24 August 2024 12:34:31 -0400 (0:00:00.066) 0:00:40.431 *******
ok: [managed_node1] => {
"changed": false,
"stat": {
"atime": 1724516929.7168462,
"attr_flags": "",
"attributes": [],
"block_size": 4096,
"blocks": 32,
"charset": "binary",
"checksum": "bb5b46ffbafcaa8c4021f3c8b3cb8594f48ef34b",
"ctime": 1724516896.725572,
"dev": 51713,
"device_type": 0,
"executable": true,
"exists": true,
"gid": 0,
"gr_name": "root",
"inode": 6986657,
"isblk": false,
"ischr": false,
"isdir": false,
"isfifo": false,
"isgid": false,
"islnk": false,
"isreg": true,
"issock": false,
"isuid": false,
"mimetype": "application/x-sharedlib",
"mode": "0755",
"mtime": 1700557386.0,
"nlink": 1,
"path": "/usr/bin/getsubids",
"pw_name": "root",
"readable": true,
"rgrp": true,
"roth": true,
"rusr": true,
"size": 12640,
"uid": 0,
"version": "1255679238",
"wgrp": false,
"woth": false,
"writeable": true,
"wusr": true,
"xgrp": true,
"xoth": true,
"xusr": true
}
}
TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:50
Saturday 24 August 2024 12:34:32 -0400 (0:00:00.420) 0:00:40.851 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Check group with getsubids] ***********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:55
Saturday 24 August 2024 12:34:32 -0400 (0:00:00.056) 0:00:40.908 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ******
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:60
Saturday 24 August 2024 12:34:32 -0400 (0:00:00.051) 0:00:40.960 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Get subuid file] **********************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:74
Saturday 24 August 2024 12:34:32 -0400 (0:00:00.050) 0:00:41.010 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Get subgid file] **********************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:79
Saturday 24 August 2024 12:34:32 -0400 (0:00:00.060) 0:00:41.071 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ******
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:84
Saturday 24 August 2024 12:34:32 -0400 (0:00:00.058) 0:00:41.129 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ******
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:94
Saturday 24 August 2024 12:34:32 -0400 (0:00:00.051) 0:00:41.181 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] *****
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:101
Saturday 24 August 2024 12:34:32 -0400 (0:00:00.052) 0:00:41.233 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62
Saturday 24 August 2024 12:34:32 -0400 (0:00:00.050) 0:00:41.283 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_activate_systemd_unit": true,
"__podman_images_found": [],
"__podman_kube_yamls_raw": "",
"__podman_service_name": "quadlet-demo-mysql-volume.service",
"__podman_systemd_scope": "system",
"__podman_user_home_dir": "/root",
"__podman_xdg_runtime_dir": "/run/user/0"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73
Saturday 24 August 2024 12:34:32 -0400 (0:00:00.206) 0:00:41.490 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_quadlet_path": "/etc/containers/systemd"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Get kube yaml contents] ***************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77
Saturday 24 August 2024 12:34:32 -0400 (0:00:00.054) 0:00:41.545 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87
Saturday 24 August 2024 12:34:32 -0400 (0:00:00.054) 0:00:41.599 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_images": [],
"__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo-mysql.volume",
"__podman_volumes": []
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 6] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:103
Saturday 24 August 2024 12:34:33 -0400 (0:00:00.099) 0:00:41.699 *******
ok: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : Cleanup quadlets] *********************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:110
Saturday 24 August 2024 12:34:33 -0400 (0:00:00.046) 0:00:41.745 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Create and update quadlets] ***********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:114
Saturday 24 August 2024 12:34:33 -0400 (0:00:00.041) 0:00:41.786 *******
included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml for managed_node1
TASK [fedora.linux_system_roles.podman : Manage linger] ************************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:2
Saturday 24 August 2024 12:34:33 -0400 (0:00:00.072) 0:00:41.858 *******
included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for managed_node1
TASK [fedora.linux_system_roles.podman : Enable linger if needed] **************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12
Saturday 24 August 2024 12:34:33 -0400 (0:00:00.055) 0:00:41.913 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18
Saturday 24 August 2024 12:34:33 -0400 (0:00:00.033) 0:00:41.946 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22
Saturday 24 August 2024 12:34:33 -0400 (0:00:00.038) 0:00:41.985 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Create host directories] **************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:7
Saturday 24 August 2024 12:34:33 -0400 (0:00:00.042) 0:00:42.028 *******
TASK [fedora.linux_system_roles.podman : Ensure container images are present] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:18
Saturday 24 August 2024 12:34:33 -0400 (0:00:00.048) 0:00:42.076 *******
skipping: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : Ensure the quadlet directory is present] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:39
Saturday 24 August 2024 12:34:33 -0400 (0:00:00.058) 0:00:42.135 *******
ok: [managed_node1] => {
"changed": false,
"gid": 0,
"group": "root",
"mode": "0755",
"owner": "root",
"path": "/etc/containers/systemd",
"secontext": "system_u:object_r:etc_t:s0",
"size": 34,
"state": "directory",
"uid": 0
}
TASK [fedora.linux_system_roles.podman : Ensure quadlet file is copied] ********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:48
Saturday 24 August 2024 12:34:33 -0400 (0:00:00.413) 0:00:42.548 *******
changed: [managed_node1] => {
"changed": true,
"checksum": "585f8cbdf0ec73000f9227dcffbef71e9552ea4a",
"dest": "/etc/containers/systemd/quadlet-demo-mysql.volume",
"gid": 0,
"group": "root",
"md5sum": "5ddd03a022aeb4502d9bc8ce436b4233",
"mode": "0644",
"owner": "root",
"secontext": "system_u:object_r:etc_t:s0",
"size": 9,
"src": "/root/.ansible/tmp/ansible-tmp-1724517273.9811473-29551-138782588482123/source",
"state": "file",
"uid": 0
}
TASK [fedora.linux_system_roles.podman : Ensure quadlet file content is present] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:58
Saturday 24 August 2024 12:34:34 -0400 (0:00:00.750) 0:00:43.298 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Ensure quadlet file is present] *******
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:70
Saturday 24 August 2024 12:34:34 -0400 (0:00:00.052) 0:00:43.351 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Reload systemctl] *********************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:82
Saturday 24 August 2024 12:34:34 -0400 (0:00:00.055) 0:00:43.407 *******
ok: [managed_node1] => {
"changed": false,
"name": null,
"status": {}
}
TASK [fedora.linux_system_roles.podman : Start service] ************************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:110
Saturday 24 August 2024 12:34:35 -0400 (0:00:00.671) 0:00:44.079 *******
changed: [managed_node1] => {
"changed": true,
"name": "quadlet-demo-mysql-volume.service",
"state": "started",
"status": {
"ActiveEnterTimestampMonotonic": "0",
"ActiveExitTimestampMonotonic": "0",
"ActiveState": "inactive",
"After": "system.slice sysinit.target systemd-journald.socket -.mount basic.target",
"AllowIsolate": "no",
"AllowedCPUs": "",
"AllowedMemoryNodes": "",
"AmbientCapabilities": "",
"AssertResult": "no",
"AssertTimestampMonotonic": "0",
"Before": "shutdown.target",
"BlockIOAccounting": "no",
"BlockIOWeight": "[not set]",
"CPUAccounting": "no",
"CPUAffinity": "",
"CPUAffinityFromNUMA": "no",
"CPUQuotaPerSecUSec": "infinity",
"CPUQuotaPeriodUSec": "infinity",
"CPUSchedulingPolicy": "0",
"CPUSchedulingPriority": "0",
"CPUSchedulingResetOnFork": "no",
"CPUShares": "[not set]",
"CPUUsageNSec": "[not set]",
"CPUWeight": "[not set]",
"CacheDirectoryMode": "0755",
"CanFreeze": "yes",
"CanIsolate": "no",
"CanReload": "no",
"CanStart": "yes",
"CanStop": "yes",
"CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf",
"CollectMode": "inactive",
"ConditionResult": "no",
"ConditionTimestampMonotonic": "0",
"ConfigurationDirectoryMode": "0755",
"Conflicts": "shutdown.target",
"ControlPID": "0",
"DefaultDependencies": "yes",
"DefaultMemoryLow": "0",
"DefaultMemoryMin": "0",
"Delegate": "no",
"Description": "quadlet-demo-mysql-volume.service",
"DevicePolicy": "auto",
"DynamicUser": "no",
"EffectiveCPUs": "",
"EffectiveMemoryNodes": "",
"ExecMainCode": "0",
"ExecMainExitTimestampMonotonic": "0",
"ExecMainPID": "0",
"ExecMainStartTimestampMonotonic": "0",
"ExecMainStatus": "0",
"ExecStart": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman volume create --ignore systemd-quadlet-demo-mysql ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"FailureAction": "none",
"FileDescriptorStoreMax": "0",
"FragmentPath": "/run/systemd/generator/quadlet-demo-mysql-volume.service",
"FreezerState": "running",
"GID": "[not set]",
"GuessMainPID": "yes",
"IOAccounting": "no",
"IOSchedulingClass": "0",
"IOSchedulingPriority": "0",
"IOWeight": "[not set]",
"IPAccounting": "no",
"IPEgressBytes": "18446744073709551615",
"IPEgressPackets": "18446744073709551615",
"IPIngressBytes": "18446744073709551615",
"IPIngressPackets": "18446744073709551615",
"Id": "quadlet-demo-mysql-volume.service",
"IgnoreOnIsolate": "no",
"IgnoreSIGPIPE": "yes",
"InactiveEnterTimestampMonotonic": "0",
"InactiveExitTimestampMonotonic": "0",
"JobRunningTimeoutUSec": "infinity",
"JobTimeoutAction": "none",
"JobTimeoutUSec": "infinity",
"KeyringMode": "private",
"KillMode": "control-group",
"KillSignal": "15",
"LimitAS": "infinity",
"LimitASSoft": "infinity",
"LimitCORE": "infinity",
"LimitCORESoft": "0",
"LimitCPU": "infinity",
"LimitCPUSoft": "infinity",
"LimitDATA": "infinity",
"LimitDATASoft": "infinity",
"LimitFSIZE": "infinity",
"LimitFSIZESoft": "infinity",
"LimitLOCKS": "infinity",
"LimitLOCKSSoft": "infinity",
"LimitMEMLOCK": "65536",
"LimitMEMLOCKSoft": "65536",
"LimitMSGQUEUE": "819200",
"LimitMSGQUEUESoft": "819200",
"LimitNICE": "0",
"LimitNICESoft": "0",
"LimitNOFILE": "262144",
"LimitNOFILESoft": "1024",
"LimitNPROC": "14003",
"LimitNPROCSoft": "14003",
"LimitRSS": "infinity",
"LimitRSSSoft": "infinity",
"LimitRTPRIO": "0",
"LimitRTPRIOSoft": "0",
"LimitRTTIME": "infinity",
"LimitRTTIMESoft": "infinity",
"LimitSIGPENDING": "14003",
"LimitSIGPENDINGSoft": "14003",
"LimitSTACK": "infinity",
"LimitSTACKSoft": "8388608",
"LoadState": "loaded",
"LockPersonality": "no",
"LogLevelMax": "-1",
"LogRateLimitBurst": "0",
"LogRateLimitIntervalUSec": "0",
"LogsDirectoryMode": "0755",
"MainPID": "0",
"MemoryAccounting": "yes",
"MemoryCurrent": "[not set]",
"MemoryDenyWriteExecute": "no",
"MemoryHigh": "infinity",
"MemoryLimit": "infinity",
"MemoryLow": "0",
"MemoryMax": "infinity",
"MemoryMin": "0",
"MemorySwapMax": "infinity",
"MountAPIVFS": "no",
"MountFlags": "",
"NFileDescriptorStore": "0",
"NRestarts": "0",
"NUMAMask": "",
"NUMAPolicy": "n/a",
"Names": "quadlet-demo-mysql-volume.service",
"NeedDaemonReload": "no",
"Nice": "0",
"NoNewPrivileges": "no",
"NonBlocking": "no",
"NotifyAccess": "none",
"OOMScoreAdjust": "0",
"OnFailureJobMode": "replace",
"PermissionsStartOnly": "no",
"Perpetual": "no",
"PrivateDevices": "no",
"PrivateMounts": "no",
"PrivateNetwork": "no",
"PrivateTmp": "no",
"PrivateUsers": "no",
"ProtectControlGroups": "no",
"ProtectHome": "no",
"ProtectKernelModules": "no",
"ProtectKernelTunables": "no",
"ProtectSystem": "no",
"RefuseManualStart": "no",
"RefuseManualStop": "no",
"RemainAfterExit": "yes",
"RemoveIPC": "no",
"Requires": "system.slice sysinit.target -.mount",
"RequiresMountsFor": "/run/containers",
"Restart": "no",
"RestartUSec": "100ms",
"RestrictNamespaces": "no",
"RestrictRealtime": "no",
"RestrictSUIDSGID": "no",
"Result": "success",
"RootDirectoryStartOnly": "no",
"RuntimeDirectoryMode": "0755",
"RuntimeDirectoryPreserve": "no",
"RuntimeMaxUSec": "infinity",
"SameProcessGroup": "no",
"SecureBits": "0",
"SendSIGHUP": "no",
"SendSIGKILL": "yes",
"Slice": "system.slice",
"StandardError": "inherit",
"StandardInput": "null",
"StandardInputData": "",
"StandardOutput": "journal",
"StartLimitAction": "none",
"StartLimitBurst": "5",
"StartLimitIntervalUSec": "10s",
"StartupBlockIOWeight": "[not set]",
"StartupCPUShares": "[not set]",
"StartupCPUWeight": "[not set]",
"StartupIOWeight": "[not set]",
"StateChangeTimestampMonotonic": "0",
"StateDirectoryMode": "0755",
"StatusErrno": "0",
"StopWhenUnneeded": "no",
"SubState": "dead",
"SuccessAction": "none",
"SyslogFacility": "3",
"SyslogIdentifier": "quadlet-demo-mysql-volume",
"SyslogLevel": "6",
"SyslogLevelPrefix": "yes",
"SyslogPriority": "30",
"SystemCallErrorNumber": "0",
"TTYReset": "no",
"TTYVHangup": "no",
"TTYVTDisallocate": "no",
"TasksAccounting": "yes",
"TasksCurrent": "[not set]",
"TasksMax": "22405",
"TimeoutStartUSec": "infinity",
"TimeoutStopUSec": "1min 30s",
"TimerSlackNSec": "50000",
"Transient": "no",
"Type": "oneshot",
"UID": "[not set]",
"UMask": "0022",
"UnitFilePreset": "disabled",
"UnitFileState": "generated",
"UtmpMode": "init",
"WatchdogTimestampMonotonic": "0",
"WatchdogUSec": "0"
}
}
TASK [fedora.linux_system_roles.podman : Restart service] **********************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:125
Saturday 24 August 2024 12:34:36 -0400 (0:00:00.581) 0:00:44.660 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14
Saturday 24 August 2024 12:34:36 -0400 (0:00:00.034) 0:00:44.695 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_quadlet_file_src": "",
"__podman_quadlet_spec": {},
"__podman_quadlet_str": "[Install]\nWantedBy=default.target\n\n[Container]\nImage=quay.io/linux-system-roles/mysql:5.6\nContainerName=quadlet-demo-mysql\nVolume=quadlet-demo-mysql.volume:/var/lib/mysql\nVolume=/tmp/quadlet_demo:/var/lib/quadlet_demo:Z\nNetwork=quadlet-demo.network\nSecret=mysql-root-password-container,type=env,target=MYSQL_ROOT_PASSWORD\nHealthCmd=/bin/true\nHealthOnFailure=kill\n",
"__podman_quadlet_template_src": "quadlet-demo-mysql.container.j2"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25
Saturday 24 August 2024 12:34:36 -0400 (0:00:00.102) 0:00:44.798 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_continue_if_pull_fails": false,
"__podman_pull_image": true,
"__podman_state": "created",
"__podman_systemd_unit_scope": "",
"__podman_user": "root"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] *****
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35
Saturday 24 August 2024 12:34:36 -0400 (0:00:00.053) 0:00:44.851 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48
Saturday 24 August 2024 12:34:36 -0400 (0:00:00.053) 0:00:44.904 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_quadlet_name": "quadlet-demo-mysql",
"__podman_quadlet_type": "container",
"__podman_rootless": false
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Check user and group information] *****
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57
Saturday 24 August 2024 12:34:36 -0400 (0:00:00.050) 0:00:44.955 *******
included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for managed_node1
TASK [fedora.linux_system_roles.podman : Get user information] *****************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2
Saturday 24 August 2024 12:34:36 -0400 (0:00:00.074) 0:00:45.030 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Fail if user does not exist] **********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9
Saturday 24 August 2024 12:34:36 -0400 (0:00:00.089) 0:00:45.119 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set group for podman user] ************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16
Saturday 24 August 2024 12:34:36 -0400 (0:00:00.042) 0:00:45.161 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_group": "0"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Get group information] ****************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:28
Saturday 24 August 2024 12:34:36 -0400 (0:00:00.045) 0:00:45.207 *******
ok: [managed_node1] => {
"ansible_facts": {
"getent_group": {
"root": [
"x",
"0",
""
]
}
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set group name] ***********************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:35
Saturday 24 August 2024 12:34:37 -0400 (0:00:00.432) 0:00:45.640 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_group_name": "root"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : See if getsubids exists] **************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:39
Saturday 24 August 2024 12:34:37 -0400 (0:00:00.065) 0:00:45.706 *******
ok: [managed_node1] => {
"changed": false,
"stat": {
"atime": 1724516929.7168462,
"attr_flags": "",
"attributes": [],
"block_size": 4096,
"blocks": 32,
"charset": "binary",
"checksum": "bb5b46ffbafcaa8c4021f3c8b3cb8594f48ef34b",
"ctime": 1724516896.725572,
"dev": 51713,
"device_type": 0,
"executable": true,
"exists": true,
"gid": 0,
"gr_name": "root",
"inode": 6986657,
"isblk": false,
"ischr": false,
"isdir": false,
"isfifo": false,
"isgid": false,
"islnk": false,
"isreg": true,
"issock": false,
"isuid": false,
"mimetype": "application/x-sharedlib",
"mode": "0755",
"mtime": 1700557386.0,
"nlink": 1,
"path": "/usr/bin/getsubids",
"pw_name": "root",
"readable": true,
"rgrp": true,
"roth": true,
"rusr": true,
"size": 12640,
"uid": 0,
"version": "1255679238",
"wgrp": false,
"woth": false,
"writeable": true,
"wusr": true,
"xgrp": true,
"xoth": true,
"xusr": true
}
}
TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:50
Saturday 24 August 2024 12:34:37 -0400 (0:00:00.421) 0:00:46.127 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Check group with getsubids] ***********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:55
Saturday 24 August 2024 12:34:37 -0400 (0:00:00.046) 0:00:46.173 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ******
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:60
Saturday 24 August 2024 12:34:37 -0400 (0:00:00.035) 0:00:46.209 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Get subuid file] **********************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:74
Saturday 24 August 2024 12:34:37 -0400 (0:00:00.034) 0:00:46.244 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Get subgid file] **********************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:79
Saturday 24 August 2024 12:34:37 -0400 (0:00:00.033) 0:00:46.277 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ******
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:84
Saturday 24 August 2024 12:34:37 -0400 (0:00:00.034) 0:00:46.311 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ******
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:94
Saturday 24 August 2024 12:34:37 -0400 (0:00:00.037) 0:00:46.349 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] *****
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:101
Saturday 24 August 2024 12:34:37 -0400 (0:00:00.051) 0:00:46.400 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62
Saturday 24 August 2024 12:34:37 -0400 (0:00:00.054) 0:00:46.455 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_activate_systemd_unit": true,
"__podman_images_found": [
"quay.io/linux-system-roles/mysql:5.6"
],
"__podman_kube_yamls_raw": "",
"__podman_service_name": "quadlet-demo-mysql.service",
"__podman_systemd_scope": "system",
"__podman_user_home_dir": "/root",
"__podman_xdg_runtime_dir": "/run/user/0"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73
Saturday 24 August 2024 12:34:37 -0400 (0:00:00.076) 0:00:46.532 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_quadlet_path": "/etc/containers/systemd"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Get kube yaml contents] ***************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77
Saturday 24 August 2024 12:34:37 -0400 (0:00:00.043) 0:00:46.575 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87
Saturday 24 August 2024 12:34:38 -0400 (0:00:00.042) 0:00:46.618 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_images": [
"quay.io/linux-system-roles/mysql:5.6"
],
"__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo-mysql.container",
"__podman_volumes": [
"/tmp/quadlet_demo"
]
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 6] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:103
Saturday 24 August 2024 12:34:38 -0400 (0:00:00.161) 0:00:46.779 *******
ok: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : Cleanup quadlets] *********************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:110
Saturday 24 August 2024 12:34:38 -0400 (0:00:00.039) 0:00:46.819 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Create and update quadlets] ***********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:114
Saturday 24 August 2024 12:34:38 -0400 (0:00:00.033) 0:00:46.852 *******
included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml for managed_node1
TASK [fedora.linux_system_roles.podman : Manage linger] ************************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:2
Saturday 24 August 2024 12:34:38 -0400 (0:00:00.100) 0:00:46.953 *******
included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for managed_node1
TASK [fedora.linux_system_roles.podman : Enable linger if needed] **************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12
Saturday 24 August 2024 12:34:38 -0400 (0:00:00.081) 0:00:47.035 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18
Saturday 24 August 2024 12:34:38 -0400 (0:00:00.036) 0:00:47.071 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22
Saturday 24 August 2024 12:34:38 -0400 (0:00:00.037) 0:00:47.109 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Create host directories] **************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:7
Saturday 24 August 2024 12:34:38 -0400 (0:00:00.038) 0:00:47.147 *******
[WARNING]: Using a variable for a task's 'args' is unsafe in some situations
(see
https://docs.ansible.com/ansible/devel/reference_appendices/faq.html#argsplat-
unsafe)
changed: [managed_node1] => (item=/tmp/quadlet_demo) => {
"ansible_loop_var": "item",
"changed": true,
"gid": 0,
"group": "root",
"item": "/tmp/quadlet_demo",
"mode": "0777",
"owner": "root",
"path": "/tmp/quadlet_demo",
"secontext": "unconfined_u:object_r:user_tmp_t:s0",
"size": 6,
"state": "directory",
"uid": 0
}
TASK [fedora.linux_system_roles.podman : Ensure container images are present] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:18
Saturday 24 August 2024 12:34:38 -0400 (0:00:00.414) 0:00:47.562 *******
changed: [managed_node1] => (item=None) => {
"attempts": 1,
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": true
}
changed: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": true
}
TASK [fedora.linux_system_roles.podman : Ensure the quadlet directory is present] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:39
Saturday 24 August 2024 12:34:46 -0400 (0:00:07.231) 0:00:54.793 *******
ok: [managed_node1] => {
"changed": false,
"gid": 0,
"group": "root",
"mode": "0755",
"owner": "root",
"path": "/etc/containers/systemd",
"secontext": "system_u:object_r:etc_t:s0",
"size": 67,
"state": "directory",
"uid": 0
}
TASK [fedora.linux_system_roles.podman : Ensure quadlet file is copied] ********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:48
Saturday 24 August 2024 12:34:46 -0400 (0:00:00.428) 0:00:55.222 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Ensure quadlet file content is present] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:58
Saturday 24 August 2024 12:34:46 -0400 (0:00:00.047) 0:00:55.269 *******
changed: [managed_node1] => {
"changed": true,
"checksum": "ca62b2ad3cc9afb5b5371ebbf797b9bc4fd7edd4",
"dest": "/etc/containers/systemd/quadlet-demo-mysql.container",
"gid": 0,
"group": "root",
"md5sum": "341b473056d2a5dfa35970b0d2e23a5d",
"mode": "0644",
"owner": "root",
"secontext": "system_u:object_r:etc_t:s0",
"size": 363,
"src": "/root/.ansible/tmp/ansible-tmp-1724517286.7033756-30107-189472742671718/source",
"state": "file",
"uid": 0
}
TASK [fedora.linux_system_roles.podman : Ensure quadlet file is present] *******
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:70
Saturday 24 August 2024 12:34:47 -0400 (0:00:00.718) 0:00:55.988 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Reload systemctl] *********************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:82
Saturday 24 August 2024 12:34:47 -0400 (0:00:00.036) 0:00:56.024 *******
ok: [managed_node1] => {
"changed": false,
"name": null,
"status": {}
}
TASK [fedora.linux_system_roles.podman : Start service] ************************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:110
Saturday 24 August 2024 12:34:48 -0400 (0:00:00.648) 0:00:56.672 *******
changed: [managed_node1] => {
"changed": true,
"name": "quadlet-demo-mysql.service",
"state": "started",
"status": {
"ActiveEnterTimestampMonotonic": "0",
"ActiveExitTimestampMonotonic": "0",
"ActiveState": "inactive",
"After": "tmp.mount system.slice basic.target quadlet-demo-mysql-volume.service quadlet-demo-network.service -.mount systemd-journald.socket sysinit.target",
"AllowIsolate": "no",
"AllowedCPUs": "",
"AllowedMemoryNodes": "",
"AmbientCapabilities": "",
"AssertResult": "no",
"AssertTimestampMonotonic": "0",
"Before": "shutdown.target",
"BlockIOAccounting": "no",
"BlockIOWeight": "[not set]",
"CPUAccounting": "no",
"CPUAffinity": "",
"CPUAffinityFromNUMA": "no",
"CPUQuotaPerSecUSec": "infinity",
"CPUQuotaPeriodUSec": "infinity",
"CPUSchedulingPolicy": "0",
"CPUSchedulingPriority": "0",
"CPUSchedulingResetOnFork": "no",
"CPUShares": "[not set]",
"CPUUsageNSec": "[not set]",
"CPUWeight": "[not set]",
"CacheDirectoryMode": "0755",
"CanFreeze": "yes",
"CanIsolate": "no",
"CanReload": "no",
"CanStart": "yes",
"CanStop": "yes",
"CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf",
"CollectMode": "inactive",
"ConditionResult": "no",
"ConditionTimestampMonotonic": "0",
"ConfigurationDirectoryMode": "0755",
"Conflicts": "shutdown.target",
"ControlPID": "0",
"DefaultDependencies": "yes",
"DefaultMemoryLow": "0",
"DefaultMemoryMin": "0",
"Delegate": "yes",
"DelegateControllers": "cpu cpuacct cpuset io blkio memory devices pids",
"Description": "quadlet-demo-mysql.service",
"DevicePolicy": "auto",
"DynamicUser": "no",
"EffectiveCPUs": "",
"EffectiveMemoryNodes": "",
"Environment": "PODMAN_SYSTEMD_UNIT=quadlet-demo-mysql.service",
"ExecMainCode": "0",
"ExecMainExitTimestampMonotonic": "0",
"ExecMainPID": "0",
"ExecMainStartTimestampMonotonic": "0",
"ExecMainStatus": "0",
"ExecStart": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman run --name=quadlet-demo-mysql --cidfile=/run/quadlet-demo-mysql.cid --replace --rm --cgroups=split --network=systemd-quadlet-demo --sdnotify=conmon -d -v systemd-quadlet-demo-mysql:/var/lib/mysql -v /tmp/quadlet_demo:/var/lib/quadlet_demo:Z --secret mysql-root-password-container,type=env,target=MYSQL_ROOT_PASSWORD --health-cmd /bin/true --health-on-failure kill quay.io/linux-system-roles/mysql:5.6 ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"ExecStop": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman rm -v -f -i --cidfile=/run/quadlet-demo-mysql.cid ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"ExecStopPost": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman rm -v -f -i --cidfile=/run/quadlet-demo-mysql.cid ; ignore_errors=yes ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"FailureAction": "none",
"FileDescriptorStoreMax": "0",
"FragmentPath": "/run/systemd/generator/quadlet-demo-mysql.service",
"FreezerState": "running",
"GID": "[not set]",
"GuessMainPID": "yes",
"IOAccounting": "no",
"IOSchedulingClass": "0",
"IOSchedulingPriority": "0",
"IOWeight": "[not set]",
"IPAccounting": "no",
"IPEgressBytes": "18446744073709551615",
"IPEgressPackets": "18446744073709551615",
"IPIngressBytes": "18446744073709551615",
"IPIngressPackets": "18446744073709551615",
"Id": "quadlet-demo-mysql.service",
"IgnoreOnIsolate": "no",
"IgnoreSIGPIPE": "yes",
"InactiveEnterTimestampMonotonic": "0",
"InactiveExitTimestampMonotonic": "0",
"JobRunningTimeoutUSec": "infinity",
"JobTimeoutAction": "none",
"JobTimeoutUSec": "infinity",
"KeyringMode": "private",
"KillMode": "mixed",
"KillSignal": "15",
"LimitAS": "infinity",
"LimitASSoft": "infinity",
"LimitCORE": "infinity",
"LimitCORESoft": "0",
"LimitCPU": "infinity",
"LimitCPUSoft": "infinity",
"LimitDATA": "infinity",
"LimitDATASoft": "infinity",
"LimitFSIZE": "infinity",
"LimitFSIZESoft": "infinity",
"LimitLOCKS": "infinity",
"LimitLOCKSSoft": "infinity",
"LimitMEMLOCK": "65536",
"LimitMEMLOCKSoft": "65536",
"LimitMSGQUEUE": "819200",
"LimitMSGQUEUESoft": "819200",
"LimitNICE": "0",
"LimitNICESoft": "0",
"LimitNOFILE": "262144",
"LimitNOFILESoft": "1024",
"LimitNPROC": "14003",
"LimitNPROCSoft": "14003",
"LimitRSS": "infinity",
"LimitRSSSoft": "infinity",
"LimitRTPRIO": "0",
"LimitRTPRIOSoft": "0",
"LimitRTTIME": "infinity",
"LimitRTTIMESoft": "infinity",
"LimitSIGPENDING": "14003",
"LimitSIGPENDINGSoft": "14003",
"LimitSTACK": "infinity",
"LimitSTACKSoft": "8388608",
"LoadState": "loaded",
"LockPersonality": "no",
"LogLevelMax": "-1",
"LogRateLimitBurst": "0",
"LogRateLimitIntervalUSec": "0",
"LogsDirectoryMode": "0755",
"MainPID": "0",
"MemoryAccounting": "yes",
"MemoryCurrent": "[not set]",
"MemoryDenyWriteExecute": "no",
"MemoryHigh": "infinity",
"MemoryLimit": "infinity",
"MemoryLow": "0",
"MemoryMax": "infinity",
"MemoryMin": "0",
"MemorySwapMax": "infinity",
"MountAPIVFS": "no",
"MountFlags": "",
"NFileDescriptorStore": "0",
"NRestarts": "0",
"NUMAMask": "",
"NUMAPolicy": "n/a",
"Names": "quadlet-demo-mysql.service",
"NeedDaemonReload": "no",
"Nice": "0",
"NoNewPrivileges": "no",
"NonBlocking": "no",
"NotifyAccess": "all",
"OOMScoreAdjust": "0",
"OnFailureJobMode": "replace",
"PermissionsStartOnly": "no",
"Perpetual": "no",
"PrivateDevices": "no",
"PrivateMounts": "no",
"PrivateNetwork": "no",
"PrivateTmp": "no",
"PrivateUsers": "no",
"ProtectControlGroups": "no",
"ProtectHome": "no",
"ProtectKernelModules": "no",
"ProtectKernelTunables": "no",
"ProtectSystem": "no",
"RefuseManualStart": "no",
"RefuseManualStop": "no",
"RemainAfterExit": "no",
"RemoveIPC": "no",
"Requires": "quadlet-demo-mysql-volume.service quadlet-demo-network.service system.slice -.mount sysinit.target",
"RequiresMountsFor": "/tmp/quadlet_demo /run/containers",
"Restart": "no",
"RestartUSec": "100ms",
"RestrictNamespaces": "no",
"RestrictRealtime": "no",
"RestrictSUIDSGID": "no",
"Result": "success",
"RootDirectoryStartOnly": "no",
"RuntimeDirectoryMode": "0755",
"RuntimeDirectoryPreserve": "no",
"RuntimeMaxUSec": "infinity",
"SameProcessGroup": "no",
"SecureBits": "0",
"SendSIGHUP": "no",
"SendSIGKILL": "yes",
"Slice": "system.slice",
"SourcePath": "/etc/containers/systemd/quadlet-demo-mysql.container",
"StandardError": "inherit",
"StandardInput": "null",
"StandardInputData": "",
"StandardOutput": "journal",
"StartLimitAction": "none",
"StartLimitBurst": "5",
"StartLimitIntervalUSec": "10s",
"StartupBlockIOWeight": "[not set]",
"StartupCPUShares": "[not set]",
"StartupCPUWeight": "[not set]",
"StartupIOWeight": "[not set]",
"StateChangeTimestampMonotonic": "0",
"StateDirectoryMode": "0755",
"StatusErrno": "0",
"StopWhenUnneeded": "no",
"SubState": "dead",
"SuccessAction": "none",
"SyslogFacility": "3",
"SyslogIdentifier": "quadlet-demo-mysql",
"SyslogLevel": "6",
"SyslogLevelPrefix": "yes",
"SyslogPriority": "30",
"SystemCallErrorNumber": "0",
"TTYReset": "no",
"TTYVHangup": "no",
"TTYVTDisallocate": "no",
"TasksAccounting": "yes",
"TasksCurrent": "[not set]",
"TasksMax": "22405",
"TimeoutStartUSec": "1min 30s",
"TimeoutStopUSec": "1min 30s",
"TimerSlackNSec": "50000",
"Transient": "no",
"Type": "notify",
"UID": "[not set]",
"UMask": "0022",
"UnitFilePreset": "disabled",
"UnitFileState": "generated",
"UtmpMode": "init",
"WatchdogTimestampMonotonic": "0",
"WatchdogUSec": "0"
}
}
TASK [fedora.linux_system_roles.podman : Restart service] **********************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:125
Saturday 24 August 2024 12:34:49 -0400 (0:00:01.042) 0:00:57.715 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14
Saturday 24 August 2024 12:34:49 -0400 (0:00:00.041) 0:00:57.756 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_quadlet_file_src": "envoy-proxy-configmap.yml",
"__podman_quadlet_spec": {},
"__podman_quadlet_str": "---\napiVersion: v1\nkind: ConfigMap\nmetadata:\n name: envoy-proxy-config\ndata:\n envoy.yaml: |\n admin:\n address:\n socket_address:\n address: 0.0.0.0\n port_value: 9901\n\n static_resources:\n listeners:\n - name: listener_0\n address:\n socket_address:\n address: 0.0.0.0\n port_value: 8080\n filter_chains:\n - filters:\n - name: envoy.filters.network.http_connection_manager\n typed_config:\n \"@type\": type.googleapis.com/envoy.extensions.filters.network.http_connection_manager.v3.HttpConnectionManager\n stat_prefix: ingress_http\n codec_type: AUTO\n route_config:\n name: local_route\n virtual_hosts:\n - name: local_service\n domains: [\"*\"]\n routes:\n - match:\n prefix: \"/\"\n route:\n cluster: backend\n http_filters:\n - name: envoy.filters.http.router\n typed_config:\n \"@type\": type.googleapis.com/envoy.extensions.filters.http.router.v3.Router\n transport_socket:\n name: envoy.transport_sockets.tls\n typed_config:\n \"@type\": type.googleapis.com/envoy.extensions.transport_sockets.tls.v3.DownstreamTlsContext\n common_tls_context:\n tls_certificates:\n - certificate_chain:\n filename: /etc/envoy-certificates/certificate.pem\n private_key:\n filename: /etc/envoy-certificates/certificate.key\n clusters:\n - name: backend\n connect_timeout: 5s\n type: STATIC\n dns_refresh_rate: 1800s\n lb_policy: ROUND_ROBIN\n load_assignment:\n cluster_name: backend\n endpoints:\n - lb_endpoints:\n - endpoint:\n address:\n socket_address:\n address: 127.0.0.1\n port_value: 80",
"__podman_quadlet_template_src": ""
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25
Saturday 24 August 2024 12:34:49 -0400 (0:00:00.048) 0:00:57.805 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_continue_if_pull_fails": false,
"__podman_pull_image": true,
"__podman_state": "created",
"__podman_systemd_unit_scope": "",
"__podman_user": "root"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] *****
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35
Saturday 24 August 2024 12:34:49 -0400 (0:00:00.047) 0:00:57.852 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48
Saturday 24 August 2024 12:34:49 -0400 (0:00:00.035) 0:00:57.888 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_quadlet_name": "envoy-proxy-configmap",
"__podman_quadlet_type": "yml",
"__podman_rootless": false
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Check user and group information] *****
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57
Saturday 24 August 2024 12:34:49 -0400 (0:00:00.046) 0:00:57.934 *******
included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for managed_node1
TASK [fedora.linux_system_roles.podman : Get user information] *****************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2
Saturday 24 August 2024 12:34:49 -0400 (0:00:00.060) 0:00:57.994 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Fail if user does not exist] **********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9
Saturday 24 August 2024 12:34:49 -0400 (0:00:00.071) 0:00:58.066 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set group for podman user] ************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16
Saturday 24 August 2024 12:34:49 -0400 (0:00:00.051) 0:00:58.117 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_group": "0"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Get group information] ****************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:28
Saturday 24 August 2024 12:34:49 -0400 (0:00:00.067) 0:00:58.184 *******
ok: [managed_node1] => {
"ansible_facts": {
"getent_group": {
"root": [
"x",
"0",
""
]
}
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set group name] ***********************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:35
Saturday 24 August 2024 12:34:49 -0400 (0:00:00.417) 0:00:58.601 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_group_name": "root"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : See if getsubids exists] **************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:39
Saturday 24 August 2024 12:34:50 -0400 (0:00:00.051) 0:00:58.653 *******
ok: [managed_node1] => {
"changed": false,
"stat": {
"atime": 1724516929.7168462,
"attr_flags": "",
"attributes": [],
"block_size": 4096,
"blocks": 32,
"charset": "binary",
"checksum": "bb5b46ffbafcaa8c4021f3c8b3cb8594f48ef34b",
"ctime": 1724516896.725572,
"dev": 51713,
"device_type": 0,
"executable": true,
"exists": true,
"gid": 0,
"gr_name": "root",
"inode": 6986657,
"isblk": false,
"ischr": false,
"isdir": false,
"isfifo": false,
"isgid": false,
"islnk": false,
"isreg": true,
"issock": false,
"isuid": false,
"mimetype": "application/x-sharedlib",
"mode": "0755",
"mtime": 1700557386.0,
"nlink": 1,
"path": "/usr/bin/getsubids",
"pw_name": "root",
"readable": true,
"rgrp": true,
"roth": true,
"rusr": true,
"size": 12640,
"uid": 0,
"version": "1255679238",
"wgrp": false,
"woth": false,
"writeable": true,
"wusr": true,
"xgrp": true,
"xoth": true,
"xusr": true
}
}
TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:50
Saturday 24 August 2024 12:34:50 -0400 (0:00:00.408) 0:00:59.062 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Check group with getsubids] ***********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:55
Saturday 24 August 2024 12:34:50 -0400 (0:00:00.038) 0:00:59.100 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ******
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:60
Saturday 24 August 2024 12:34:50 -0400 (0:00:00.081) 0:00:59.182 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Get subuid file] **********************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:74
Saturday 24 August 2024 12:34:50 -0400 (0:00:00.031) 0:00:59.214 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Get subgid file] **********************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:79
Saturday 24 August 2024 12:34:50 -0400 (0:00:00.031) 0:00:59.245 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ******
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:84
Saturday 24 August 2024 12:34:50 -0400 (0:00:00.032) 0:00:59.278 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ******
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:94
Saturday 24 August 2024 12:34:50 -0400 (0:00:00.034) 0:00:59.312 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] *****
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:101
Saturday 24 August 2024 12:34:50 -0400 (0:00:00.037) 0:00:59.349 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62
Saturday 24 August 2024 12:34:50 -0400 (0:00:00.045) 0:00:59.395 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_activate_systemd_unit": true,
"__podman_images_found": [],
"__podman_kube_yamls_raw": "",
"__podman_service_name": "",
"__podman_systemd_scope": "system",
"__podman_user_home_dir": "/root",
"__podman_xdg_runtime_dir": "/run/user/0"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73
Saturday 24 August 2024 12:34:50 -0400 (0:00:00.092) 0:00:59.488 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_quadlet_path": "/etc/containers/systemd"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Get kube yaml contents] ***************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77
Saturday 24 August 2024 12:34:50 -0400 (0:00:00.053) 0:00:59.541 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87
Saturday 24 August 2024 12:34:50 -0400 (0:00:00.055) 0:00:59.596 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_images": [],
"__podman_quadlet_file": "/etc/containers/systemd/envoy-proxy-configmap.yml",
"__podman_volumes": []
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 6] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:103
Saturday 24 August 2024 12:34:51 -0400 (0:00:00.115) 0:00:59.711 *******
ok: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : Cleanup quadlets] *********************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:110
Saturday 24 August 2024 12:34:51 -0400 (0:00:00.059) 0:00:59.770 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Create and update quadlets] ***********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:114
Saturday 24 August 2024 12:34:51 -0400 (0:00:00.051) 0:00:59.822 *******
included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml for managed_node1
TASK [fedora.linux_system_roles.podman : Manage linger] ************************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:2
Saturday 24 August 2024 12:34:51 -0400 (0:00:00.110) 0:00:59.932 *******
included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for managed_node1
TASK [fedora.linux_system_roles.podman : Enable linger if needed] **************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12
Saturday 24 August 2024 12:34:51 -0400 (0:00:00.089) 0:01:00.022 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18
Saturday 24 August 2024 12:34:51 -0400 (0:00:00.053) 0:01:00.075 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22
Saturday 24 August 2024 12:34:51 -0400 (0:00:00.050) 0:01:00.126 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Create host directories] **************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:7
Saturday 24 August 2024 12:34:51 -0400 (0:00:00.101) 0:01:00.227 *******
TASK [fedora.linux_system_roles.podman : Ensure container images are present] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:18
Saturday 24 August 2024 12:34:51 -0400 (0:00:00.050) 0:01:00.278 *******
skipping: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : Ensure the quadlet directory is present] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:39
Saturday 24 August 2024 12:34:51 -0400 (0:00:00.059) 0:01:00.337 *******
ok: [managed_node1] => {
"changed": false,
"gid": 0,
"group": "root",
"mode": "0755",
"owner": "root",
"path": "/etc/containers/systemd",
"secontext": "system_u:object_r:etc_t:s0",
"size": 103,
"state": "directory",
"uid": 0
}
TASK [fedora.linux_system_roles.podman : Ensure quadlet file is copied] ********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:48
Saturday 24 August 2024 12:34:52 -0400 (0:00:00.416) 0:01:00.753 *******
changed: [managed_node1] => {
"changed": true,
"checksum": "d681c7d56f912150d041873e880818b22a90c188",
"dest": "/etc/containers/systemd/envoy-proxy-configmap.yml",
"gid": 0,
"group": "root",
"md5sum": "aec75d972c231aac004e1338934544cf",
"mode": "0644",
"owner": "root",
"secontext": "system_u:object_r:etc_t:s0",
"size": 2102,
"src": "/root/.ansible/tmp/ansible-tmp-1724517292.2012258-30396-55356854144861/source",
"state": "file",
"uid": 0
}
TASK [fedora.linux_system_roles.podman : Ensure quadlet file content is present] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:58
Saturday 24 August 2024 12:34:52 -0400 (0:00:00.740) 0:01:01.493 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Ensure quadlet file is present] *******
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:70
Saturday 24 August 2024 12:34:52 -0400 (0:00:00.034) 0:01:01.528 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Reload systemctl] *********************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:82
Saturday 24 August 2024 12:34:52 -0400 (0:00:00.041) 0:01:01.570 *******
ok: [managed_node1] => {
"changed": false,
"name": null,
"status": {}
}
TASK [fedora.linux_system_roles.podman : Start service] ************************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:110
Saturday 24 August 2024 12:34:53 -0400 (0:00:00.660) 0:01:02.230 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Restart service] **********************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:125
Saturday 24 August 2024 12:34:53 -0400 (0:00:00.054) 0:01:02.284 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14
Saturday 24 August 2024 12:34:53 -0400 (0:00:00.068) 0:01:02.353 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_quadlet_file_src": "",
"__podman_quadlet_spec": {},
"__podman_quadlet_str": "---\napiVersion: v1\nkind: PersistentVolumeClaim\nmetadata:\n name: wp-pv-claim\n labels:\n app: wordpress\nspec:\n accessModes:\n - ReadWriteOnce\n resources:\n requests:\n storage: 20Gi\n---\napiVersion: v1\nkind: Pod\nmetadata:\n name: quadlet-demo\nspec:\n containers:\n - name: wordpress\n image: quay.io/linux-system-roles/wordpress:4.8-apache\n env:\n - name: WORDPRESS_DB_HOST\n value: quadlet-demo-mysql\n - name: WORDPRESS_DB_PASSWORD\n valueFrom:\n secretKeyRef:\n name: mysql-root-password-kube\n key: password\n volumeMounts:\n - name: wordpress-persistent-storage\n mountPath: /var/www/html\n resources:\n requests:\n memory: \"64Mi\"\n cpu: \"250m\"\n limits:\n memory: \"128Mi\"\n cpu: \"500m\"\n - name: envoy\n image: quay.io/linux-system-roles/envoyproxy:v1.25.0\n volumeMounts:\n - name: config-volume\n mountPath: /etc/envoy\n - name: certificates\n mountPath: /etc/envoy-certificates\n env:\n - name: ENVOY_UID\n value: \"0\"\n resources:\n requests:\n memory: \"64Mi\"\n cpu: \"250m\"\n limits:\n memory: \"128Mi\"\n cpu: \"500m\"\n volumes:\n - name: config-volume\n configMap:\n name: envoy-proxy-config\n - name: certificates\n secret:\n secretName: envoy-certificates\n - name: wordpress-persistent-storage\n persistentVolumeClaim:\n claimName: wp-pv-claim\n - name: www # not used - for testing hostpath\n hostPath:\n path: /tmp/httpd3\n - name: create # not used - for testing hostpath\n hostPath:\n path: /tmp/httpd3-create\n",
"__podman_quadlet_template_src": "quadlet-demo.yml.j2"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25
Saturday 24 August 2024 12:34:53 -0400 (0:00:00.201) 0:01:02.555 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_continue_if_pull_fails": false,
"__podman_pull_image": true,
"__podman_state": "created",
"__podman_systemd_unit_scope": "",
"__podman_user": "root"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] *****
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35
Saturday 24 August 2024 12:34:54 -0400 (0:00:00.061) 0:01:02.616 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48
Saturday 24 August 2024 12:34:54 -0400 (0:00:00.105) 0:01:02.721 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_quadlet_name": "quadlet-demo",
"__podman_quadlet_type": "yml",
"__podman_rootless": false
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Check user and group information] *****
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57
Saturday 24 August 2024 12:34:54 -0400 (0:00:00.071) 0:01:02.793 *******
included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for managed_node1
TASK [fedora.linux_system_roles.podman : Get user information] *****************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2
Saturday 24 August 2024 12:34:54 -0400 (0:00:00.097) 0:01:02.890 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Fail if user does not exist] **********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9
Saturday 24 August 2024 12:34:54 -0400 (0:00:00.062) 0:01:02.953 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set group for podman user] ************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16
Saturday 24 August 2024 12:34:54 -0400 (0:00:00.061) 0:01:03.014 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_group": "0"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Get group information] ****************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:28
Saturday 24 August 2024 12:34:54 -0400 (0:00:00.067) 0:01:03.082 *******
ok: [managed_node1] => {
"ansible_facts": {
"getent_group": {
"root": [
"x",
"0",
""
]
}
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set group name] ***********************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:35
Saturday 24 August 2024 12:34:54 -0400 (0:00:00.418) 0:01:03.501 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_group_name": "root"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : See if getsubids exists] **************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:39
Saturday 24 August 2024 12:34:54 -0400 (0:00:00.062) 0:01:03.564 *******
ok: [managed_node1] => {
"changed": false,
"stat": {
"atime": 1724516929.7168462,
"attr_flags": "",
"attributes": [],
"block_size": 4096,
"blocks": 32,
"charset": "binary",
"checksum": "bb5b46ffbafcaa8c4021f3c8b3cb8594f48ef34b",
"ctime": 1724516896.725572,
"dev": 51713,
"device_type": 0,
"executable": true,
"exists": true,
"gid": 0,
"gr_name": "root",
"inode": 6986657,
"isblk": false,
"ischr": false,
"isdir": false,
"isfifo": false,
"isgid": false,
"islnk": false,
"isreg": true,
"issock": false,
"isuid": false,
"mimetype": "application/x-sharedlib",
"mode": "0755",
"mtime": 1700557386.0,
"nlink": 1,
"path": "/usr/bin/getsubids",
"pw_name": "root",
"readable": true,
"rgrp": true,
"roth": true,
"rusr": true,
"size": 12640,
"uid": 0,
"version": "1255679238",
"wgrp": false,
"woth": false,
"writeable": true,
"wusr": true,
"xgrp": true,
"xoth": true,
"xusr": true
}
}
TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:50
Saturday 24 August 2024 12:34:55 -0400 (0:00:00.478) 0:01:04.043 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Check group with getsubids] ***********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:55
Saturday 24 August 2024 12:34:55 -0400 (0:00:00.037) 0:01:04.080 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ******
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:60
Saturday 24 August 2024 12:34:55 -0400 (0:00:00.037) 0:01:04.117 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Get subuid file] **********************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:74
Saturday 24 August 2024 12:34:55 -0400 (0:00:00.033) 0:01:04.151 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Get subgid file] **********************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:79
Saturday 24 August 2024 12:34:55 -0400 (0:00:00.030) 0:01:04.181 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ******
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:84
Saturday 24 August 2024 12:34:55 -0400 (0:00:00.031) 0:01:04.213 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ******
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:94
Saturday 24 August 2024 12:34:55 -0400 (0:00:00.030) 0:01:04.243 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] *****
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:101
Saturday 24 August 2024 12:34:55 -0400 (0:00:00.078) 0:01:04.322 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62
Saturday 24 August 2024 12:34:55 -0400 (0:00:00.060) 0:01:04.382 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_activate_systemd_unit": true,
"__podman_images_found": [],
"__podman_kube_yamls_raw": "",
"__podman_service_name": "",
"__podman_systemd_scope": "system",
"__podman_user_home_dir": "/root",
"__podman_xdg_runtime_dir": "/run/user/0"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73
Saturday 24 August 2024 12:34:55 -0400 (0:00:00.067) 0:01:04.449 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_quadlet_path": "/etc/containers/systemd"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Get kube yaml contents] ***************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77
Saturday 24 August 2024 12:34:55 -0400 (0:00:00.040) 0:01:04.490 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87
Saturday 24 August 2024 12:34:55 -0400 (0:00:00.041) 0:01:04.532 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_images": [],
"__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo.yml",
"__podman_volumes": []
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 6] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:103
Saturday 24 August 2024 12:34:55 -0400 (0:00:00.073) 0:01:04.605 *******
ok: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : Cleanup quadlets] *********************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:110
Saturday 24 August 2024 12:34:56 -0400 (0:00:00.037) 0:01:04.643 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Create and update quadlets] ***********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:114
Saturday 24 August 2024 12:34:56 -0400 (0:00:00.031) 0:01:04.674 *******
included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml for managed_node1
TASK [fedora.linux_system_roles.podman : Manage linger] ************************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:2
Saturday 24 August 2024 12:34:56 -0400 (0:00:00.088) 0:01:04.763 *******
included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for managed_node1
TASK [fedora.linux_system_roles.podman : Enable linger if needed] **************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12
Saturday 24 August 2024 12:34:56 -0400 (0:00:00.087) 0:01:04.851 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18
Saturday 24 August 2024 12:34:56 -0400 (0:00:00.054) 0:01:04.905 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22
Saturday 24 August 2024 12:34:56 -0400 (0:00:00.050) 0:01:04.956 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Create host directories] **************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:7
Saturday 24 August 2024 12:34:56 -0400 (0:00:00.050) 0:01:05.006 *******
TASK [fedora.linux_system_roles.podman : Ensure container images are present] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:18
Saturday 24 August 2024 12:34:56 -0400 (0:00:00.048) 0:01:05.055 *******
skipping: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : Ensure the quadlet directory is present] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:39
Saturday 24 August 2024 12:34:56 -0400 (0:00:00.048) 0:01:05.103 *******
ok: [managed_node1] => {
"changed": false,
"gid": 0,
"group": "root",
"mode": "0755",
"owner": "root",
"path": "/etc/containers/systemd",
"secontext": "system_u:object_r:etc_t:s0",
"size": 136,
"state": "directory",
"uid": 0
}
TASK [fedora.linux_system_roles.podman : Ensure quadlet file is copied] ********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:48
Saturday 24 August 2024 12:34:56 -0400 (0:00:00.480) 0:01:05.583 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Ensure quadlet file content is present] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:58
Saturday 24 August 2024 12:34:57 -0400 (0:00:00.037) 0:01:05.621 *******
changed: [managed_node1] => {
"changed": true,
"checksum": "998dccde0483b1654327a46ddd89cbaa47650370",
"dest": "/etc/containers/systemd/quadlet-demo.yml",
"gid": 0,
"group": "root",
"md5sum": "fd890594adfc24339cb9cdc5e7b19a66",
"mode": "0644",
"owner": "root",
"secontext": "system_u:object_r:etc_t:s0",
"size": 1605,
"src": "/root/.ansible/tmp/ansible-tmp-1724517297.0551069-30667-219028600000855/source",
"state": "file",
"uid": 0
}
TASK [fedora.linux_system_roles.podman : Ensure quadlet file is present] *******
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:70
Saturday 24 August 2024 12:34:57 -0400 (0:00:00.737) 0:01:06.359 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Reload systemctl] *********************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:82
Saturday 24 August 2024 12:34:57 -0400 (0:00:00.058) 0:01:06.417 *******
ok: [managed_node1] => {
"changed": false,
"name": null,
"status": {}
}
TASK [fedora.linux_system_roles.podman : Start service] ************************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:110
Saturday 24 August 2024 12:34:58 -0400 (0:00:00.636) 0:01:07.054 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Restart service] **********************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:125
Saturday 24 August 2024 12:34:58 -0400 (0:00:00.058) 0:01:07.112 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14
Saturday 24 August 2024 12:34:58 -0400 (0:00:00.057) 0:01:07.170 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_quadlet_file_src": "quadlet-demo.kube",
"__podman_quadlet_spec": {},
"__podman_quadlet_str": "[Install]\nWantedBy=default.target\n\n[Unit]\nRequires=quadlet-demo-mysql.service\nAfter=quadlet-demo-mysql.service\n\n[Kube]\n# Point to the yaml file in the same directory\nYaml=quadlet-demo.yml\n# Use the quadlet-demo network\nNetwork=quadlet-demo.network\n# Publish the envoy proxy data port\nPublishPort=8000:8080\n# Publish the envoy proxy admin port\nPublishPort=9000:9901\n# Use the envoy proxy config map in the same directory\nConfigMap=envoy-proxy-configmap.yml",
"__podman_quadlet_template_src": ""
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25
Saturday 24 August 2024 12:34:58 -0400 (0:00:00.069) 0:01:07.239 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_continue_if_pull_fails": false,
"__podman_pull_image": true,
"__podman_state": "created",
"__podman_systemd_unit_scope": "",
"__podman_user": "root"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] *****
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35
Saturday 24 August 2024 12:34:58 -0400 (0:00:00.059) 0:01:07.299 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48
Saturday 24 August 2024 12:34:58 -0400 (0:00:00.055) 0:01:07.355 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_quadlet_name": "quadlet-demo",
"__podman_quadlet_type": "kube",
"__podman_rootless": false
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Check user and group information] *****
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57
Saturday 24 August 2024 12:34:58 -0400 (0:00:00.071) 0:01:07.426 *******
included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for managed_node1
TASK [fedora.linux_system_roles.podman : Get user information] *****************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2
Saturday 24 August 2024 12:34:58 -0400 (0:00:00.107) 0:01:07.534 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Fail if user does not exist] **********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9
Saturday 24 August 2024 12:34:58 -0400 (0:00:00.059) 0:01:07.594 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set group for podman user] ************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16
Saturday 24 August 2024 12:34:59 -0400 (0:00:00.161) 0:01:07.756 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_group": "0"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Get group information] ****************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:28
Saturday 24 August 2024 12:34:59 -0400 (0:00:00.071) 0:01:07.827 *******
ok: [managed_node1] => {
"ansible_facts": {
"getent_group": {
"root": [
"x",
"0",
""
]
}
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set group name] ***********************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:35
Saturday 24 August 2024 12:34:59 -0400 (0:00:00.405) 0:01:08.233 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_group_name": "root"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : See if getsubids exists] **************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:39
Saturday 24 August 2024 12:34:59 -0400 (0:00:00.040) 0:01:08.274 *******
ok: [managed_node1] => {
"changed": false,
"stat": {
"atime": 1724516929.7168462,
"attr_flags": "",
"attributes": [],
"block_size": 4096,
"blocks": 32,
"charset": "binary",
"checksum": "bb5b46ffbafcaa8c4021f3c8b3cb8594f48ef34b",
"ctime": 1724516896.725572,
"dev": 51713,
"device_type": 0,
"executable": true,
"exists": true,
"gid": 0,
"gr_name": "root",
"inode": 6986657,
"isblk": false,
"ischr": false,
"isdir": false,
"isfifo": false,
"isgid": false,
"islnk": false,
"isreg": true,
"issock": false,
"isuid": false,
"mimetype": "application/x-sharedlib",
"mode": "0755",
"mtime": 1700557386.0,
"nlink": 1,
"path": "/usr/bin/getsubids",
"pw_name": "root",
"readable": true,
"rgrp": true,
"roth": true,
"rusr": true,
"size": 12640,
"uid": 0,
"version": "1255679238",
"wgrp": false,
"woth": false,
"writeable": true,
"wusr": true,
"xgrp": true,
"xoth": true,
"xusr": true
}
}
TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:50
Saturday 24 August 2024 12:35:00 -0400 (0:00:00.408) 0:01:08.682 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Check group with getsubids] ***********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:55
Saturday 24 August 2024 12:35:00 -0400 (0:00:00.055) 0:01:08.738 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ******
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:60
Saturday 24 August 2024 12:35:00 -0400 (0:00:00.052) 0:01:08.790 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Get subuid file] **********************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:74
Saturday 24 August 2024 12:35:00 -0400 (0:00:00.055) 0:01:08.846 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Get subgid file] **********************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:79
Saturday 24 August 2024 12:35:00 -0400 (0:00:00.041) 0:01:08.887 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ******
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:84
Saturday 24 August 2024 12:35:00 -0400 (0:00:00.036) 0:01:08.924 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ******
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:94
Saturday 24 August 2024 12:35:00 -0400 (0:00:00.037) 0:01:08.962 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] *****
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:101
Saturday 24 August 2024 12:35:00 -0400 (0:00:00.043) 0:01:09.005 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62
Saturday 24 August 2024 12:35:00 -0400 (0:00:00.038) 0:01:09.043 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_activate_systemd_unit": true,
"__podman_images_found": [],
"__podman_kube_yamls_raw": [
"quadlet-demo.yml"
],
"__podman_service_name": "quadlet-demo.service",
"__podman_systemd_scope": "system",
"__podman_user_home_dir": "/root",
"__podman_xdg_runtime_dir": "/run/user/0"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73
Saturday 24 August 2024 12:35:00 -0400 (0:00:00.057) 0:01:09.100 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_quadlet_path": "/etc/containers/systemd"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Get kube yaml contents] ***************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77
Saturday 24 August 2024 12:35:00 -0400 (0:00:00.033) 0:01:09.134 *******
ok: [managed_node1] => {
"changed": false,
"content": "LS0tCmFwaVZlcnNpb246IHYxCmtpbmQ6IFBlcnNpc3RlbnRWb2x1bWVDbGFpbQptZXRhZGF0YToKICBuYW1lOiB3cC1wdi1jbGFpbQogIGxhYmVsczoKICAgIGFwcDogd29yZHByZXNzCnNwZWM6CiAgYWNjZXNzTW9kZXM6CiAgLSBSZWFkV3JpdGVPbmNlCiAgcmVzb3VyY2VzOgogICAgcmVxdWVzdHM6CiAgICAgIHN0b3JhZ2U6IDIwR2kKLS0tCmFwaVZlcnNpb246IHYxCmtpbmQ6IFBvZAptZXRhZGF0YToKICBuYW1lOiBxdWFkbGV0LWRlbW8Kc3BlYzoKICBjb250YWluZXJzOgogIC0gbmFtZTogd29yZHByZXNzCiAgICBpbWFnZTogcXVheS5pby9saW51eC1zeXN0ZW0tcm9sZXMvd29yZHByZXNzOjQuOC1hcGFjaGUKICAgIGVudjoKICAgIC0gbmFtZTogV09SRFBSRVNTX0RCX0hPU1QKICAgICAgdmFsdWU6IHF1YWRsZXQtZGVtby1teXNxbAogICAgLSBuYW1lOiBXT1JEUFJFU1NfREJfUEFTU1dPUkQKICAgICAgdmFsdWVGcm9tOgogICAgICAgIHNlY3JldEtleVJlZjoKICAgICAgICAgIG5hbWU6IG15c3FsLXJvb3QtcGFzc3dvcmQta3ViZQogICAgICAgICAga2V5OiBwYXNzd29yZAogICAgdm9sdW1lTW91bnRzOgogICAgLSBuYW1lOiB3b3JkcHJlc3MtcGVyc2lzdGVudC1zdG9yYWdlCiAgICAgIG1vdW50UGF0aDogL3Zhci93d3cvaHRtbAogICAgcmVzb3VyY2VzOgogICAgICByZXF1ZXN0czoKICAgICAgICBtZW1vcnk6ICI2NE1pIgogICAgICAgIGNwdTogIjI1MG0iCiAgICAgIGxpbWl0czoKICAgICAgICBtZW1vcnk6ICIxMjhNaSIKICAgICAgICBjcHU6ICI1MDBtIgogIC0gbmFtZTogZW52b3kKICAgIGltYWdlOiBxdWF5LmlvL2xpbnV4LXN5c3RlbS1yb2xlcy9lbnZveXByb3h5OnYxLjI1LjAKICAgIHZvbHVtZU1vdW50czoKICAgIC0gbmFtZTogY29uZmlnLXZvbHVtZQogICAgICBtb3VudFBhdGg6IC9ldGMvZW52b3kKICAgIC0gbmFtZTogY2VydGlmaWNhdGVzCiAgICAgIG1vdW50UGF0aDogL2V0Yy9lbnZveS1jZXJ0aWZpY2F0ZXMKICAgIGVudjoKICAgIC0gbmFtZTogRU5WT1lfVUlECiAgICAgIHZhbHVlOiAiMCIKICAgIHJlc291cmNlczoKICAgICAgcmVxdWVzdHM6CiAgICAgICAgbWVtb3J5OiAiNjRNaSIKICAgICAgICBjcHU6ICIyNTBtIgogICAgICBsaW1pdHM6CiAgICAgICAgbWVtb3J5OiAiMTI4TWkiCiAgICAgICAgY3B1OiAiNTAwbSIKICB2b2x1bWVzOgogIC0gbmFtZTogY29uZmlnLXZvbHVtZQogICAgY29uZmlnTWFwOgogICAgICBuYW1lOiBlbnZveS1wcm94eS1jb25maWcKICAtIG5hbWU6IGNlcnRpZmljYXRlcwogICAgc2VjcmV0OgogICAgICBzZWNyZXROYW1lOiBlbnZveS1jZXJ0aWZpY2F0ZXMKICAtIG5hbWU6IHdvcmRwcmVzcy1wZXJzaXN0ZW50LXN0b3JhZ2UKICAgIHBlcnNpc3RlbnRWb2x1bWVDbGFpbToKICAgICAgY2xhaW1OYW1lOiB3cC1wdi1jbGFpbQogIC0gbmFtZTogd3d3ICAjIG5vdCB1c2VkIC0gZm9yIHRlc3RpbmcgaG9zdHBhdGgKICAgIGhvc3RQYXRoOgogICAgICBwYXRoOiAvdG1wL2h0dHBkMwogIC0gbmFtZTogY3JlYXRlICAjIG5vdCB1c2VkIC0gZm9yIHRlc3RpbmcgaG9zdHBhdGgKICAgIGhvc3RQYXRoOgogICAgICBwYXRoOiAvdG1wL2h0dHBkMy1jcmVhdGUK",
"encoding": "base64",
"source": "/etc/containers/systemd/quadlet-demo.yml"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87
Saturday 24 August 2024 12:35:00 -0400 (0:00:00.429) 0:01:09.563 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_images": [
"quay.io/linux-system-roles/wordpress:4.8-apache",
"quay.io/linux-system-roles/envoyproxy:v1.25.0"
],
"__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo.kube",
"__podman_volumes": [
"/tmp/httpd3-create",
"/tmp/httpd3"
]
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 6] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:103
Saturday 24 August 2024 12:35:01 -0400 (0:00:00.195) 0:01:09.759 *******
ok: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : Cleanup quadlets] *********************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:110
Saturday 24 August 2024 12:35:01 -0400 (0:00:00.057) 0:01:09.817 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Create and update quadlets] ***********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:114
Saturday 24 August 2024 12:35:01 -0400 (0:00:00.054) 0:01:09.871 *******
included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml for managed_node1
TASK [fedora.linux_system_roles.podman : Manage linger] ************************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:2
Saturday 24 August 2024 12:35:01 -0400 (0:00:00.115) 0:01:09.986 *******
included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for managed_node1
TASK [fedora.linux_system_roles.podman : Enable linger if needed] **************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12
Saturday 24 August 2024 12:35:01 -0400 (0:00:00.095) 0:01:10.081 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18
Saturday 24 August 2024 12:35:01 -0400 (0:00:00.036) 0:01:10.118 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22
Saturday 24 August 2024 12:35:01 -0400 (0:00:00.041) 0:01:10.160 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Create host directories] **************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:7
Saturday 24 August 2024 12:35:01 -0400 (0:00:00.042) 0:01:10.202 *******
[WARNING]: Using a variable for a task's 'args' is unsafe in some situations
(see
https://docs.ansible.com/ansible/devel/reference_appendices/faq.html#argsplat-
unsafe)
changed: [managed_node1] => (item=/tmp/httpd3-create) => {
"ansible_loop_var": "item",
"changed": true,
"gid": 0,
"group": "root",
"item": "/tmp/httpd3-create",
"mode": "0755",
"owner": "root",
"path": "/tmp/httpd3-create",
"secontext": "unconfined_u:object_r:user_tmp_t:s0",
"size": 6,
"state": "directory",
"uid": 0
}
changed: [managed_node1] => (item=/tmp/httpd3) => {
"ansible_loop_var": "item",
"changed": true,
"gid": 0,
"group": "root",
"item": "/tmp/httpd3",
"mode": "0755",
"owner": "root",
"path": "/tmp/httpd3",
"secontext": "unconfined_u:object_r:user_tmp_t:s0",
"size": 6,
"state": "directory",
"uid": 0
}
TASK [fedora.linux_system_roles.podman : Ensure container images are present] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:18
Saturday 24 August 2024 12:35:02 -0400 (0:00:00.733) 0:01:10.936 *******
changed: [managed_node1] => (item=None) => {
"attempts": 1,
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": true
}
changed: [managed_node1] => (item=None) => {
"attempts": 1,
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": true
}
changed: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": true
}
TASK [fedora.linux_system_roles.podman : Ensure the quadlet directory is present] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:39
Saturday 24 August 2024 12:35:20 -0400 (0:00:18.295) 0:01:29.231 *******
ok: [managed_node1] => {
"changed": false,
"gid": 0,
"group": "root",
"mode": "0755",
"owner": "root",
"path": "/etc/containers/systemd",
"secontext": "system_u:object_r:etc_t:s0",
"size": 160,
"state": "directory",
"uid": 0
}
TASK [fedora.linux_system_roles.podman : Ensure quadlet file is copied] ********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:48
Saturday 24 August 2024 12:35:20 -0400 (0:00:00.379) 0:01:29.611 *******
changed: [managed_node1] => {
"changed": true,
"checksum": "7a5c73a5d935a42431c87bcdbeb8a04ed0909dc7",
"dest": "/etc/containers/systemd/quadlet-demo.kube",
"gid": 0,
"group": "root",
"md5sum": "da53c88f92b68b0487aa209f795b6bb3",
"mode": "0644",
"owner": "root",
"secontext": "system_u:object_r:etc_t:s0",
"size": 456,
"src": "/root/.ansible/tmp/ansible-tmp-1724517321.03834-31464-224102551766944/source",
"state": "file",
"uid": 0
}
TASK [fedora.linux_system_roles.podman : Ensure quadlet file content is present] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:58
Saturday 24 August 2024 12:35:21 -0400 (0:00:00.691) 0:01:30.302 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Ensure quadlet file is present] *******
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:70
Saturday 24 August 2024 12:35:21 -0400 (0:00:00.047) 0:01:30.350 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Reload systemctl] *********************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:82
Saturday 24 August 2024 12:35:21 -0400 (0:00:00.038) 0:01:30.388 *******
ok: [managed_node1] => {
"changed": false,
"name": null,
"status": {}
}
TASK [fedora.linux_system_roles.podman : Start service] ************************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:110
Saturday 24 August 2024 12:35:22 -0400 (0:00:00.599) 0:01:30.988 *******
changed: [managed_node1] => {
"changed": true,
"name": "quadlet-demo.service",
"state": "started",
"status": {
"ActiveEnterTimestampMonotonic": "0",
"ActiveExitTimestampMonotonic": "0",
"ActiveState": "inactive",
"After": "quadlet-demo-mysql.service basic.target -.mount sysinit.target system.slice quadlet-demo-network.service systemd-journald.socket",
"AllowIsolate": "no",
"AllowedCPUs": "",
"AllowedMemoryNodes": "",
"AmbientCapabilities": "",
"AssertResult": "no",
"AssertTimestampMonotonic": "0",
"Before": "shutdown.target",
"BlockIOAccounting": "no",
"BlockIOWeight": "[not set]",
"CPUAccounting": "no",
"CPUAffinity": "",
"CPUAffinityFromNUMA": "no",
"CPUQuotaPerSecUSec": "infinity",
"CPUQuotaPeriodUSec": "infinity",
"CPUSchedulingPolicy": "0",
"CPUSchedulingPriority": "0",
"CPUSchedulingResetOnFork": "no",
"CPUShares": "[not set]",
"CPUUsageNSec": "[not set]",
"CPUWeight": "[not set]",
"CacheDirectoryMode": "0755",
"CanFreeze": "yes",
"CanIsolate": "no",
"CanReload": "no",
"CanStart": "yes",
"CanStop": "yes",
"CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf",
"CollectMode": "inactive",
"ConditionResult": "no",
"ConditionTimestampMonotonic": "0",
"ConfigurationDirectoryMode": "0755",
"Conflicts": "shutdown.target",
"ControlPID": "0",
"DefaultDependencies": "yes",
"DefaultMemoryLow": "0",
"DefaultMemoryMin": "0",
"Delegate": "no",
"Description": "quadlet-demo.service",
"DevicePolicy": "auto",
"DynamicUser": "no",
"EffectiveCPUs": "",
"EffectiveMemoryNodes": "",
"Environment": "PODMAN_SYSTEMD_UNIT=quadlet-demo.service",
"ExecMainCode": "0",
"ExecMainExitTimestampMonotonic": "0",
"ExecMainPID": "0",
"ExecMainStartTimestampMonotonic": "0",
"ExecMainStatus": "0",
"ExecStart": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman kube play --replace --service-container=true --network=systemd-quadlet-demo --configmap /etc/containers/systemd/envoy-proxy-configmap.yml --publish 8000:8080 --publish 9000:9901 /etc/containers/systemd/quadlet-demo.yml ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"ExecStopPost": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman kube down /etc/containers/systemd/quadlet-demo.yml ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"FailureAction": "none",
"FileDescriptorStoreMax": "0",
"FragmentPath": "/run/systemd/generator/quadlet-demo.service",
"FreezerState": "running",
"GID": "[not set]",
"GuessMainPID": "yes",
"IOAccounting": "no",
"IOSchedulingClass": "0",
"IOSchedulingPriority": "0",
"IOWeight": "[not set]",
"IPAccounting": "no",
"IPEgressBytes": "18446744073709551615",
"IPEgressPackets": "18446744073709551615",
"IPIngressBytes": "18446744073709551615",
"IPIngressPackets": "18446744073709551615",
"Id": "quadlet-demo.service",
"IgnoreOnIsolate": "no",
"IgnoreSIGPIPE": "yes",
"InactiveEnterTimestampMonotonic": "0",
"InactiveExitTimestampMonotonic": "0",
"JobRunningTimeoutUSec": "infinity",
"JobTimeoutAction": "none",
"JobTimeoutUSec": "infinity",
"KeyringMode": "private",
"KillMode": "mixed",
"KillSignal": "15",
"LimitAS": "infinity",
"LimitASSoft": "infinity",
"LimitCORE": "infinity",
"LimitCORESoft": "0",
"LimitCPU": "infinity",
"LimitCPUSoft": "infinity",
"LimitDATA": "infinity",
"LimitDATASoft": "infinity",
"LimitFSIZE": "infinity",
"LimitFSIZESoft": "infinity",
"LimitLOCKS": "infinity",
"LimitLOCKSSoft": "infinity",
"LimitMEMLOCK": "65536",
"LimitMEMLOCKSoft": "65536",
"LimitMSGQUEUE": "819200",
"LimitMSGQUEUESoft": "819200",
"LimitNICE": "0",
"LimitNICESoft": "0",
"LimitNOFILE": "262144",
"LimitNOFILESoft": "1024",
"LimitNPROC": "14003",
"LimitNPROCSoft": "14003",
"LimitRSS": "infinity",
"LimitRSSSoft": "infinity",
"LimitRTPRIO": "0",
"LimitRTPRIOSoft": "0",
"LimitRTTIME": "infinity",
"LimitRTTIMESoft": "infinity",
"LimitSIGPENDING": "14003",
"LimitSIGPENDINGSoft": "14003",
"LimitSTACK": "infinity",
"LimitSTACKSoft": "8388608",
"LoadState": "loaded",
"LockPersonality": "no",
"LogLevelMax": "-1",
"LogRateLimitBurst": "0",
"LogRateLimitIntervalUSec": "0",
"LogsDirectoryMode": "0755",
"MainPID": "0",
"MemoryAccounting": "yes",
"MemoryCurrent": "[not set]",
"MemoryDenyWriteExecute": "no",
"MemoryHigh": "infinity",
"MemoryLimit": "infinity",
"MemoryLow": "0",
"MemoryMax": "infinity",
"MemoryMin": "0",
"MemorySwapMax": "infinity",
"MountAPIVFS": "no",
"MountFlags": "",
"NFileDescriptorStore": "0",
"NRestarts": "0",
"NUMAMask": "",
"NUMAPolicy": "n/a",
"Names": "quadlet-demo.service",
"NeedDaemonReload": "no",
"Nice": "0",
"NoNewPrivileges": "no",
"NonBlocking": "no",
"NotifyAccess": "all",
"OOMScoreAdjust": "0",
"OnFailureJobMode": "replace",
"PermissionsStartOnly": "no",
"Perpetual": "no",
"PrivateDevices": "no",
"PrivateMounts": "no",
"PrivateNetwork": "no",
"PrivateTmp": "no",
"PrivateUsers": "no",
"ProtectControlGroups": "no",
"ProtectHome": "no",
"ProtectKernelModules": "no",
"ProtectKernelTunables": "no",
"ProtectSystem": "no",
"RefuseManualStart": "no",
"RefuseManualStop": "no",
"RemainAfterExit": "no",
"RemoveIPC": "no",
"Requires": "quadlet-demo-network.service system.slice sysinit.target quadlet-demo-mysql.service -.mount",
"RequiresMountsFor": "/run/containers",
"Restart": "no",
"RestartUSec": "100ms",
"RestrictNamespaces": "no",
"RestrictRealtime": "no",
"RestrictSUIDSGID": "no",
"Result": "success",
"RootDirectoryStartOnly": "no",
"RuntimeDirectoryMode": "0755",
"RuntimeDirectoryPreserve": "no",
"RuntimeMaxUSec": "infinity",
"SameProcessGroup": "no",
"SecureBits": "0",
"SendSIGHUP": "no",
"SendSIGKILL": "yes",
"Slice": "system.slice",
"SourcePath": "/etc/containers/systemd/quadlet-demo.kube",
"StandardError": "inherit",
"StandardInput": "null",
"StandardInputData": "",
"StandardOutput": "journal",
"StartLimitAction": "none",
"StartLimitBurst": "5",
"StartLimitIntervalUSec": "10s",
"StartupBlockIOWeight": "[not set]",
"StartupCPUShares": "[not set]",
"StartupCPUWeight": "[not set]",
"StartupIOWeight": "[not set]",
"StateChangeTimestampMonotonic": "0",
"StateDirectoryMode": "0755",
"StatusErrno": "0",
"StopWhenUnneeded": "no",
"SubState": "dead",
"SuccessAction": "none",
"SyslogFacility": "3",
"SyslogIdentifier": "quadlet-demo",
"SyslogLevel": "6",
"SyslogLevelPrefix": "yes",
"SyslogPriority": "30",
"SystemCallErrorNumber": "0",
"TTYReset": "no",
"TTYVHangup": "no",
"TTYVTDisallocate": "no",
"TasksAccounting": "yes",
"TasksCurrent": "[not set]",
"TasksMax": "22405",
"TimeoutStartUSec": "1min 30s",
"TimeoutStopUSec": "1min 30s",
"TimerSlackNSec": "50000",
"Transient": "no",
"Type": "notify",
"UID": "[not set]",
"UMask": "0022",
"UnitFilePreset": "disabled",
"UnitFileState": "generated",
"UtmpMode": "init",
"WatchdogTimestampMonotonic": "0",
"WatchdogUSec": "0"
}
}
TASK [fedora.linux_system_roles.podman : Restart service] **********************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:125
Saturday 24 August 2024 12:35:23 -0400 (0:00:01.479) 0:01:32.467 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Cancel linger] ************************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:167
Saturday 24 August 2024 12:35:23 -0400 (0:00:00.034) 0:01:32.501 *******
TASK [fedora.linux_system_roles.podman : Handle credential files - absent] *****
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:173
Saturday 24 August 2024 12:35:23 -0400 (0:00:00.029) 0:01:32.531 *******
skipping: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : Handle certs.d files - absent] ********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:182
Saturday 24 August 2024 12:35:23 -0400 (0:00:00.030) 0:01:32.561 *******
skipping: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [Check quadlet files] *****************************************************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_demo.yml:96
Saturday 24 August 2024 12:35:23 -0400 (0:00:00.029) 0:01:32.590 *******
ok: [managed_node1] => {
"changed": false,
"cmd": [
"ls",
"-alrtF",
"/etc/containers/systemd"
],
"delta": "0:00:00.005668",
"end": "2024-08-24 12:35:24.342396",
"rc": 0,
"start": "2024-08-24 12:35:24.336728"
}
STDOUT:
total 24
drwxr-xr-x. 8 root root 182 Aug 24 12:30 ../
-rw-r--r--. 1 root root 74 Aug 24 12:34 quadlet-demo.network
-rw-r--r--. 1 root root 9 Aug 24 12:34 quadlet-demo-mysql.volume
-rw-r--r--. 1 root root 363 Aug 24 12:34 quadlet-demo-mysql.container
-rw-r--r--. 1 root root 2102 Aug 24 12:34 envoy-proxy-configmap.yml
-rw-r--r--. 1 root root 1605 Aug 24 12:34 quadlet-demo.yml
-rw-r--r--. 1 root root 456 Aug 24 12:35 quadlet-demo.kube
drwxr-xr-x. 2 root root 185 Aug 24 12:35 ./
TASK [Check containers] ********************************************************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_demo.yml:100
Saturday 24 August 2024 12:35:24 -0400 (0:00:00.441) 0:01:33.032 *******
ok: [managed_node1] => {
"changed": false,
"cmd": [
"podman",
"ps",
"-a"
],
"delta": "0:00:00.051368",
"end": "2024-08-24 12:35:24.808241",
"failed_when_result": false,
"rc": 0,
"start": "2024-08-24 12:35:24.756873"
}
STDOUT:
CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES
68250d5fc021 quay.io/linux-system-roles/mysql:5.6 mysqld 36 seconds ago Up 36 seconds (healthy) quadlet-demo-mysql
a3a21fc2dc26 localhost/podman-pause:4.9.4-dev-1708535009 1 second ago Up 1 second a96f3a51b8d1-service
c14ac2390f14 localhost/podman-pause:4.9.4-dev-1708535009 1 second ago Up 1 second 0.0.0.0:8000->8080/tcp, 0.0.0.0:9000->9901/tcp bbca104112c4-infra
3ecc5b442f00 quay.io/linux-system-roles/wordpress:4.8-apache apache2-foregroun... 1 second ago Up 1 second 0.0.0.0:8000->8080/tcp, 0.0.0.0:9000->9901/tcp quadlet-demo-wordpress
14fe7e5a1bc4 quay.io/linux-system-roles/envoyproxy:v1.25.0 envoy -c /etc/env... 1 second ago Up 1 second 0.0.0.0:8000->8080/tcp, 0.0.0.0:9000->9901/tcp quadlet-demo-envoy
TASK [Check volumes] ***********************************************************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_demo.yml:105
Saturday 24 August 2024 12:35:24 -0400 (0:00:00.463) 0:01:33.496 *******
ok: [managed_node1] => {
"changed": false,
"cmd": [
"podman",
"volume",
"ls"
],
"delta": "0:00:00.028920",
"end": "2024-08-24 12:35:25.231088",
"failed_when_result": false,
"rc": 0,
"start": "2024-08-24 12:35:25.202168"
}
STDOUT:
DRIVER VOLUME NAME
local systemd-quadlet-demo-mysql
local wp-pv-claim
local envoy-proxy-config
local envoy-certificates
TASK [Check pods] **************************************************************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_demo.yml:110
Saturday 24 August 2024 12:35:25 -0400 (0:00:00.424) 0:01:33.920 *******
ok: [managed_node1] => {
"changed": false,
"cmd": [
"podman",
"pod",
"ps",
"--ctr-ids",
"--ctr-names",
"--ctr-status"
],
"delta": "0:00:00.037163",
"end": "2024-08-24 12:35:25.657541",
"failed_when_result": false,
"rc": 0,
"start": "2024-08-24 12:35:25.620378"
}
STDOUT:
POD ID NAME STATUS CREATED INFRA ID IDS NAMES STATUS
bbca104112c4 quadlet-demo Running 2 seconds ago c14ac2390f14 c14ac2390f14,3ecc5b442f00,14fe7e5a1bc4 bbca104112c4-infra,quadlet-demo-wordpress,quadlet-demo-envoy running,running,running
TASK [Check systemd] ***********************************************************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_demo.yml:115
Saturday 24 August 2024 12:35:25 -0400 (0:00:00.416) 0:01:34.337 *******
ok: [managed_node1] => {
"changed": false,
"cmd": "set -euo pipefail; systemctl list-units | grep quadlet",
"delta": "0:00:00.011411",
"end": "2024-08-24 12:35:26.031591",
"failed_when_result": false,
"rc": 0,
"start": "2024-08-24 12:35:26.020180"
}
STDOUT:
quadlet-demo-mysql-volume.service loaded active exited quadlet-demo-mysql-volume.service
quadlet-demo-mysql.service loaded active running quadlet-demo-mysql.service
quadlet-demo-network.service loaded active exited quadlet-demo-network.service
quadlet-demo.service loaded active running quadlet-demo.service
TASK [Check web] ***************************************************************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_demo.yml:121
Saturday 24 August 2024 12:35:26 -0400 (0:00:00.373) 0:01:34.711 *******
changed: [managed_node1] => {
"attempts": 1,
"changed": true,
"checksum_dest": null,
"checksum_src": "9dbc8c154daaed74c52aa781ab9994c0246a012f",
"dest": "/run/out",
"elapsed": 0,
"gid": 0,
"group": "root",
"md5sum": "f59bccb13786ba6aa549360a06d6b1b4",
"mode": "0600",
"owner": "root",
"secontext": "system_u:object_r:var_run_t:s0",
"size": 11666,
"src": "/root/.ansible/tmp/ansible-tmp-1724517326.132896-31692-693845049303/tmps517tyl0",
"state": "file",
"status_code": 200,
"uid": 0,
"url": "https://localhost:8000"
}
MSG:
OK (unknown bytes)
TASK [Show web] ****************************************************************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_demo.yml:132
Saturday 24 August 2024 12:35:27 -0400 (0:00:01.083) 0:01:35.794 *******
ok: [managed_node1] => {
"changed": false,
"cmd": [
"cat",
"/run/out"
],
"delta": "0:00:00.002511",
"end": "2024-08-24 12:35:27.473382",
"rc": 0,
"start": "2024-08-24 12:35:27.470871"
}
STDOUT:
WordPress › Installation
WordPress
TASK [Error] *******************************************************************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_demo.yml:137
Saturday 24 August 2024 12:35:27 -0400 (0:00:00.360) 0:01:36.155 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [Check] *******************************************************************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_demo.yml:148
Saturday 24 August 2024 12:35:27 -0400 (0:00:00.030) 0:01:36.186 *******
ok: [managed_node1] => {
"changed": false,
"cmd": [
"podman",
"ps",
"-a"
],
"delta": "0:00:00.036747",
"end": "2024-08-24 12:35:27.893535",
"rc": 0,
"start": "2024-08-24 12:35:27.856788"
}
STDOUT:
CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES
68250d5fc021 quay.io/linux-system-roles/mysql:5.6 mysqld 39 seconds ago Up 39 seconds (healthy) quadlet-demo-mysql
a3a21fc2dc26 localhost/podman-pause:4.9.4-dev-1708535009 5 seconds ago Up 4 seconds a96f3a51b8d1-service
c14ac2390f14 localhost/podman-pause:4.9.4-dev-1708535009 4 seconds ago Up 4 seconds 0.0.0.0:8000->8080/tcp, 0.0.0.0:9000->9901/tcp bbca104112c4-infra
3ecc5b442f00 quay.io/linux-system-roles/wordpress:4.8-apache apache2-foregroun... 4 seconds ago Up 4 seconds 0.0.0.0:8000->8080/tcp, 0.0.0.0:9000->9901/tcp quadlet-demo-wordpress
14fe7e5a1bc4 quay.io/linux-system-roles/envoyproxy:v1.25.0 envoy -c /etc/env... 4 seconds ago Up 4 seconds 0.0.0.0:8000->8080/tcp, 0.0.0.0:9000->9901/tcp quadlet-demo-envoy
TASK [Check pods] **************************************************************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_demo.yml:152
Saturday 24 August 2024 12:35:27 -0400 (0:00:00.410) 0:01:36.597 *******
ok: [managed_node1] => {
"changed": false,
"cmd": [
"podman",
"pod",
"ps",
"--ctr-ids",
"--ctr-names",
"--ctr-status"
],
"delta": "0:00:00.036935",
"end": "2024-08-24 12:35:28.336457",
"failed_when_result": false,
"rc": 0,
"start": "2024-08-24 12:35:28.299522"
}
STDOUT:
POD ID NAME STATUS CREATED INFRA ID IDS NAMES STATUS
bbca104112c4 quadlet-demo Running 5 seconds ago c14ac2390f14 c14ac2390f14,3ecc5b442f00,14fe7e5a1bc4 bbca104112c4-infra,quadlet-demo-wordpress,quadlet-demo-envoy running,running,running
TASK [Check systemd] ***********************************************************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_demo.yml:157
Saturday 24 August 2024 12:35:28 -0400 (0:00:00.427) 0:01:37.025 *******
ok: [managed_node1] => {
"changed": false,
"cmd": "set -euo pipefail; systemctl list-units --all | grep quadlet",
"delta": "0:00:00.010186",
"end": "2024-08-24 12:35:28.707551",
"failed_when_result": false,
"rc": 0,
"start": "2024-08-24 12:35:28.697365"
}
STDOUT:
quadlet-demo-mysql-volume.service loaded active exited quadlet-demo-mysql-volume.service
quadlet-demo-mysql.service loaded active running quadlet-demo-mysql.service
quadlet-demo-network.service loaded active exited quadlet-demo-network.service
quadlet-demo.service loaded active running quadlet-demo.service
TASK [LS] **********************************************************************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_demo.yml:165
Saturday 24 August 2024 12:35:28 -0400 (0:00:00.379) 0:01:37.404 *******
ok: [managed_node1] => {
"changed": false,
"cmd": [
"ls",
"-alrtF",
"/etc/systemd/system"
],
"delta": "0:00:00.003509",
"end": "2024-08-24 12:35:29.082531",
"failed_when_result": false,
"rc": 0,
"start": "2024-08-24 12:35:29.079022"
}
STDOUT:
total 8
lrwxrwxrwx. 1 root root 9 May 11 2019 systemd-timedated.service -> /dev/null
drwxr-xr-x. 4 root root 169 May 29 03:45 ../
lrwxrwxrwx. 1 root root 39 May 29 03:45 syslog.service -> /usr/lib/systemd/system/rsyslog.service
drwxr-xr-x. 2 root root 32 May 29 03:45 getty.target.wants/
lrwxrwxrwx. 1 root root 37 May 29 03:45 ctrl-alt-del.target -> /usr/lib/systemd/system/reboot.target
lrwxrwxrwx. 1 root root 57 May 29 03:45 dbus-org.freedesktop.nm-dispatcher.service -> /usr/lib/systemd/system/NetworkManager-dispatcher.service
drwxr-xr-x. 2 root root 48 May 29 03:45 network-online.target.wants/
lrwxrwxrwx. 1 root root 41 May 29 03:45 dbus-org.freedesktop.timedate1.service -> /usr/lib/systemd/system/timedatex.service
drwxr-xr-x. 2 root root 61 May 29 03:45 timers.target.wants/
drwxr-xr-x. 2 root root 31 May 29 03:45 basic.target.wants/
drwxr-xr-x. 2 root root 38 May 29 03:45 dev-virtio\x2dports-org.qemu.guest_agent.0.device.wants/
lrwxrwxrwx. 1 root root 41 May 29 03:47 default.target -> /usr/lib/systemd/system/multi-user.target
drwxr-xr-x. 2 root root 51 May 29 03:55 sockets.target.wants/
drwxr-xr-x. 2 root root 31 May 29 03:55 remote-fs.target.wants/
drwxr-xr-x. 2 root root 59 May 29 03:55 sshd-keygen@.service.d/
drwxr-xr-x. 2 root root 119 May 29 03:55 cloud-init.target.wants/
drwxr-xr-x. 2 root root 181 May 29 03:56 sysinit.target.wants/
drwxr-xr-x. 2 root root 4096 Aug 24 12:34 multi-user.target.wants/
lrwxrwxrwx. 1 root root 41 Aug 24 12:34 dbus-org.fedoraproject.FirewallD1.service -> /usr/lib/systemd/system/firewalld.service
drwxr-xr-x. 13 root root 4096 Aug 24 12:34 ./
TASK [Cleanup] *****************************************************************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_demo.yml:172
Saturday 24 August 2024 12:35:29 -0400 (0:00:00.373) 0:01:37.778 *******
TASK [fedora.linux_system_roles.podman : Set platform/version specific variables] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:3
Saturday 24 August 2024 12:35:29 -0400 (0:00:00.110) 0:01:37.888 *******
included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml for managed_node1
TASK [fedora.linux_system_roles.podman : Ensure ansible_facts used by role] ****
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:3
Saturday 24 August 2024 12:35:29 -0400 (0:00:00.089) 0:01:37.977 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Check if system is ostree] ************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:11
Saturday 24 August 2024 12:35:29 -0400 (0:00:00.063) 0:01:38.041 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set flag to indicate system is ostree] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:16
Saturday 24 August 2024 12:35:29 -0400 (0:00:00.053) 0:01:38.094 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Check if transactional-update exists in /sbin] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:23
Saturday 24 August 2024 12:35:29 -0400 (0:00:00.051) 0:01:38.146 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set flag if transactional-update exists] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:28
Saturday 24 August 2024 12:35:29 -0400 (0:00:00.052) 0:01:38.198 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set platform/version specific variables] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:32
Saturday 24 August 2024 12:35:29 -0400 (0:00:00.051) 0:01:38.250 *******
ok: [managed_node1] => (item=RedHat.yml) => {
"ansible_facts": {
"__podman_packages": [
"podman",
"shadow-utils-subid"
]
},
"ansible_included_var_files": [
"/tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/vars/RedHat.yml"
],
"ansible_loop_var": "item",
"changed": false,
"item": "RedHat.yml"
}
skipping: [managed_node1] => (item=CentOS.yml) => {
"ansible_loop_var": "item",
"changed": false,
"item": "CentOS.yml",
"skip_reason": "Conditional result was False"
}
ok: [managed_node1] => (item=CentOS_8.yml) => {
"ansible_facts": {
"__podman_packages": [
"crun",
"podman",
"podman-plugins",
"shadow-utils-subid"
]
},
"ansible_included_var_files": [
"/tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/vars/CentOS_8.yml"
],
"ansible_loop_var": "item",
"changed": false,
"item": "CentOS_8.yml"
}
ok: [managed_node1] => (item=CentOS_8.yml) => {
"ansible_facts": {
"__podman_packages": [
"crun",
"podman",
"podman-plugins",
"shadow-utils-subid"
]
},
"ansible_included_var_files": [
"/tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/vars/CentOS_8.yml"
],
"ansible_loop_var": "item",
"changed": false,
"item": "CentOS_8.yml"
}
TASK [fedora.linux_system_roles.podman : Gather the package facts] *************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:6
Saturday 24 August 2024 12:35:29 -0400 (0:00:00.126) 0:01:38.376 *******
ok: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : Enable copr if requested] *************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:10
Saturday 24 August 2024 12:35:31 -0400 (0:00:01.508) 0:01:39.884 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Ensure required packages are installed] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:14
Saturday 24 August 2024 12:35:31 -0400 (0:00:00.036) 0:01:39.921 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Notify user that reboot is needed to apply changes] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:28
Saturday 24 August 2024 12:35:31 -0400 (0:00:00.046) 0:01:39.967 *******
skipping: [managed_node1] => {}
TASK [fedora.linux_system_roles.podman : Reboot transactional update systems] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:33
Saturday 24 August 2024 12:35:31 -0400 (0:00:00.031) 0:01:39.998 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Fail if reboot is needed and not set] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:38
Saturday 24 August 2024 12:35:31 -0400 (0:00:00.029) 0:01:40.028 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Get podman version] *******************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:46
Saturday 24 August 2024 12:35:31 -0400 (0:00:00.031) 0:01:40.059 *******
ok: [managed_node1] => {
"changed": false,
"cmd": [
"podman",
"--version"
],
"delta": "0:00:00.026652",
"end": "2024-08-24 12:35:31.780172",
"rc": 0,
"start": "2024-08-24 12:35:31.753520"
}
STDOUT:
podman version 4.9.4-dev
TASK [fedora.linux_system_roles.podman : Set podman version] *******************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:52
Saturday 24 August 2024 12:35:31 -0400 (0:00:00.403) 0:01:40.462 *******
ok: [managed_node1] => {
"ansible_facts": {
"podman_version": "4.9.4-dev"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Podman package version must be 4.2 or later] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:56
Saturday 24 August 2024 12:35:31 -0400 (0:00:00.033) 0:01:40.496 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Podman package version must be 4.4 or later for quadlet, secrets] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:63
Saturday 24 August 2024 12:35:31 -0400 (0:00:00.030) 0:01:40.527 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
META: end_host conditional evaluated to false, continuing execution for managed_node1
TASK [fedora.linux_system_roles.podman : Check user and group information] *****
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:80
Saturday 24 August 2024 12:35:31 -0400 (0:00:00.058) 0:01:40.585 *******
included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for managed_node1
TASK [fedora.linux_system_roles.podman : Get user information] *****************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2
Saturday 24 August 2024 12:35:32 -0400 (0:00:00.126) 0:01:40.712 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Fail if user does not exist] **********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9
Saturday 24 August 2024 12:35:32 -0400 (0:00:00.051) 0:01:40.764 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set group for podman user] ************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16
Saturday 24 August 2024 12:35:32 -0400 (0:00:00.053) 0:01:40.817 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_group": "0"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Get group information] ****************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:28
Saturday 24 August 2024 12:35:32 -0400 (0:00:00.051) 0:01:40.868 *******
ok: [managed_node1] => {
"ansible_facts": {
"getent_group": {
"root": [
"x",
"0",
""
]
}
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set group name] ***********************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:35
Saturday 24 August 2024 12:35:32 -0400 (0:00:00.373) 0:01:41.241 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_group_name": "root"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : See if getsubids exists] **************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:39
Saturday 24 August 2024 12:35:32 -0400 (0:00:00.053) 0:01:41.295 *******
ok: [managed_node1] => {
"changed": false,
"stat": {
"atime": 1724516929.7168462,
"attr_flags": "",
"attributes": [],
"block_size": 4096,
"blocks": 32,
"charset": "binary",
"checksum": "bb5b46ffbafcaa8c4021f3c8b3cb8594f48ef34b",
"ctime": 1724516896.725572,
"dev": 51713,
"device_type": 0,
"executable": true,
"exists": true,
"gid": 0,
"gr_name": "root",
"inode": 6986657,
"isblk": false,
"ischr": false,
"isdir": false,
"isfifo": false,
"isgid": false,
"islnk": false,
"isreg": true,
"issock": false,
"isuid": false,
"mimetype": "application/x-sharedlib",
"mode": "0755",
"mtime": 1700557386.0,
"nlink": 1,
"path": "/usr/bin/getsubids",
"pw_name": "root",
"readable": true,
"rgrp": true,
"roth": true,
"rusr": true,
"size": 12640,
"uid": 0,
"version": "1255679238",
"wgrp": false,
"woth": false,
"writeable": true,
"wusr": true,
"xgrp": true,
"xoth": true,
"xusr": true
}
}
TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:50
Saturday 24 August 2024 12:35:33 -0400 (0:00:00.374) 0:01:41.670 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Check group with getsubids] ***********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:55
Saturday 24 August 2024 12:35:33 -0400 (0:00:00.039) 0:01:41.709 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ******
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:60
Saturday 24 August 2024 12:35:33 -0400 (0:00:00.051) 0:01:41.761 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Get subuid file] **********************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:74
Saturday 24 August 2024 12:35:33 -0400 (0:00:00.061) 0:01:41.823 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Get subgid file] **********************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:79
Saturday 24 August 2024 12:35:33 -0400 (0:00:00.052) 0:01:41.875 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ******
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:84
Saturday 24 August 2024 12:35:33 -0400 (0:00:00.049) 0:01:41.925 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ******
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:94
Saturday 24 August 2024 12:35:33 -0400 (0:00:00.051) 0:01:41.977 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] *****
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:101
Saturday 24 August 2024 12:35:33 -0400 (0:00:00.036) 0:01:42.013 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set config file paths] ****************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:86
Saturday 24 August 2024 12:35:33 -0400 (0:00:00.037) 0:01:42.050 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_container_conf_file": "/etc/containers/containers.conf.d/50-systemroles.conf",
"__podman_policy_json_file": "/etc/containers/policy.json",
"__podman_registries_conf_file": "/etc/containers/registries.conf.d/50-systemroles.conf",
"__podman_storage_conf_file": "/etc/containers/storage.conf"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Handle container.conf.d] **************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:95
Saturday 24 August 2024 12:35:33 -0400 (0:00:00.051) 0:01:42.102 *******
included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_container_conf_d.yml for managed_node1
TASK [fedora.linux_system_roles.podman : Ensure containers.d exists] ***********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_container_conf_d.yml:5
Saturday 24 August 2024 12:35:33 -0400 (0:00:00.105) 0:01:42.208 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Update container config file] *********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_container_conf_d.yml:13
Saturday 24 August 2024 12:35:33 -0400 (0:00:00.031) 0:01:42.239 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Handle registries.conf.d] *************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:98
Saturday 24 August 2024 12:35:33 -0400 (0:00:00.032) 0:01:42.271 *******
included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_registries_conf_d.yml for managed_node1
TASK [fedora.linux_system_roles.podman : Ensure registries.d exists] ***********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_registries_conf_d.yml:5
Saturday 24 August 2024 12:35:33 -0400 (0:00:00.060) 0:01:42.332 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Update registries config file] ********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_registries_conf_d.yml:13
Saturday 24 August 2024 12:35:33 -0400 (0:00:00.037) 0:01:42.370 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Handle storage.conf] ******************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:101
Saturday 24 August 2024 12:35:33 -0400 (0:00:00.050) 0:01:42.421 *******
included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_storage_conf.yml for managed_node1
TASK [fedora.linux_system_roles.podman : Ensure storage.conf parent dir exists] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_storage_conf.yml:5
Saturday 24 August 2024 12:35:33 -0400 (0:00:00.105) 0:01:42.526 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Update storage config file] ***********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_storage_conf.yml:13
Saturday 24 August 2024 12:35:33 -0400 (0:00:00.050) 0:01:42.577 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Handle policy.json] *******************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:104
Saturday 24 August 2024 12:35:34 -0400 (0:00:00.057) 0:01:42.634 *******
included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml for managed_node1
TASK [fedora.linux_system_roles.podman : Ensure policy.json parent dir exists] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:6
Saturday 24 August 2024 12:35:34 -0400 (0:00:00.104) 0:01:42.738 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Stat the policy.json file] ************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:14
Saturday 24 August 2024 12:35:34 -0400 (0:00:00.037) 0:01:42.776 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Get the existing policy.json] *********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:19
Saturday 24 August 2024 12:35:34 -0400 (0:00:00.045) 0:01:42.822 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Write new policy.json file] ***********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:25
Saturday 24 August 2024 12:35:34 -0400 (0:00:00.041) 0:01:42.863 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [Manage firewall for specified ports] *************************************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:110
Saturday 24 August 2024 12:35:34 -0400 (0:00:00.038) 0:01:42.901 *******
TASK [fedora.linux_system_roles.firewall : Setup firewalld] ********************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:2
Saturday 24 August 2024 12:35:34 -0400 (0:00:00.121) 0:01:43.023 *******
included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml for managed_node1
TASK [fedora.linux_system_roles.firewall : Ensure ansible_facts used by role] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:2
Saturday 24 August 2024 12:35:34 -0400 (0:00:00.057) 0:01:43.081 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.firewall : Check if system is ostree] **********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:10
Saturday 24 August 2024 12:35:34 -0400 (0:00:00.047) 0:01:43.128 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.firewall : Set flag to indicate system is ostree] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:15
Saturday 24 August 2024 12:35:34 -0400 (0:00:00.050) 0:01:43.179 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.firewall : Check if transactional-update exists in /sbin] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:22
Saturday 24 August 2024 12:35:34 -0400 (0:00:00.041) 0:01:43.220 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.firewall : Set flag if transactional-update exists] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:27
Saturday 24 August 2024 12:35:34 -0400 (0:00:00.036) 0:01:43.257 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.firewall : Install firewalld] ******************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:31
Saturday 24 August 2024 12:35:34 -0400 (0:00:00.036) 0:01:43.294 *******
ok: [managed_node1] => {
"changed": false,
"rc": 0,
"results": []
}
MSG:
Nothing to do
lsrpackages: firewalld
TASK [fedora.linux_system_roles.firewall : Notify user that reboot is needed to apply changes] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:43
Saturday 24 August 2024 12:35:37 -0400 (0:00:02.778) 0:01:46.073 *******
skipping: [managed_node1] => {}
TASK [fedora.linux_system_roles.firewall : Reboot transactional update systems] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:48
Saturday 24 August 2024 12:35:37 -0400 (0:00:00.032) 0:01:46.105 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.firewall : Fail if reboot is needed and not set] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:53
Saturday 24 August 2024 12:35:37 -0400 (0:00:00.032) 0:01:46.138 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.firewall : Collect service facts] **************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:5
Saturday 24 August 2024 12:35:37 -0400 (0:00:00.032) 0:01:46.170 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.firewall : Attempt to stop and disable conflicting services] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:9
Saturday 24 August 2024 12:35:37 -0400 (0:00:00.030) 0:01:46.201 *******
skipping: [managed_node1] => (item=nftables) => {
"ansible_loop_var": "item",
"changed": false,
"item": "nftables",
"skip_reason": "Conditional result was False"
}
skipping: [managed_node1] => (item=iptables) => {
"ansible_loop_var": "item",
"changed": false,
"item": "iptables",
"skip_reason": "Conditional result was False"
}
skipping: [managed_node1] => (item=ufw) => {
"ansible_loop_var": "item",
"changed": false,
"item": "ufw",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.firewall : Unmask firewalld service] ***********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:22
Saturday 24 August 2024 12:35:37 -0400 (0:00:00.041) 0:01:46.243 *******
ok: [managed_node1] => {
"changed": false,
"name": "firewalld",
"status": {
"ActiveEnterTimestamp": "Sat 2024-08-24 12:34:20 EDT",
"ActiveEnterTimestampMonotonic": "663046301",
"ActiveExitTimestampMonotonic": "0",
"ActiveState": "active",
"After": "dbus.service sysinit.target dbus.socket polkit.service system.slice basic.target",
"AllowIsolate": "no",
"AllowedCPUs": "",
"AllowedMemoryNodes": "",
"AmbientCapabilities": "",
"AssertResult": "yes",
"AssertTimestamp": "Sat 2024-08-24 12:34:19 EDT",
"AssertTimestampMonotonic": "662544922",
"Before": "shutdown.target multi-user.target network-pre.target",
"BlockIOAccounting": "no",
"BlockIOWeight": "[not set]",
"BusName": "org.fedoraproject.FirewallD1",
"CPUAccounting": "no",
"CPUAffinity": "",
"CPUAffinityFromNUMA": "no",
"CPUQuotaPerSecUSec": "infinity",
"CPUQuotaPeriodUSec": "infinity",
"CPUSchedulingPolicy": "0",
"CPUSchedulingPriority": "0",
"CPUSchedulingResetOnFork": "no",
"CPUShares": "[not set]",
"CPUUsageNSec": "[not set]",
"CPUWeight": "[not set]",
"CacheDirectoryMode": "0755",
"CanFreeze": "yes",
"CanIsolate": "no",
"CanReload": "yes",
"CanStart": "yes",
"CanStop": "yes",
"CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf",
"CollectMode": "inactive",
"ConditionResult": "yes",
"ConditionTimestamp": "Sat 2024-08-24 12:34:19 EDT",
"ConditionTimestampMonotonic": "662544921",
"ConfigurationDirectoryMode": "0755",
"Conflicts": "shutdown.target ip6tables.service ipset.service ebtables.service iptables.service nftables.service",
"ControlGroup": "/system.slice/firewalld.service",
"ControlPID": "0",
"DefaultDependencies": "yes",
"DefaultMemoryLow": "0",
"DefaultMemoryMin": "0",
"Delegate": "no",
"Description": "firewalld - dynamic firewall daemon",
"DevicePolicy": "auto",
"Documentation": "man:firewalld(1)",
"DynamicUser": "no",
"EffectiveCPUs": "",
"EffectiveMemoryNodes": "",
"EnvironmentFiles": "/etc/sysconfig/firewalld (ignore_errors=yes)",
"ExecMainCode": "0",
"ExecMainExitTimestampMonotonic": "0",
"ExecMainPID": "80148",
"ExecMainStartTimestamp": "Sat 2024-08-24 12:34:19 EDT",
"ExecMainStartTimestampMonotonic": "662546362",
"ExecMainStatus": "0",
"ExecReload": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"ExecStart": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"FailureAction": "none",
"FileDescriptorStoreMax": "0",
"FragmentPath": "/usr/lib/systemd/system/firewalld.service",
"FreezerState": "running",
"GID": "[not set]",
"GuessMainPID": "yes",
"IOAccounting": "no",
"IOSchedulingClass": "0",
"IOSchedulingPriority": "0",
"IOWeight": "[not set]",
"IPAccounting": "no",
"IPEgressBytes": "18446744073709551615",
"IPEgressPackets": "18446744073709551615",
"IPIngressBytes": "18446744073709551615",
"IPIngressPackets": "18446744073709551615",
"Id": "firewalld.service",
"IgnoreOnIsolate": "no",
"IgnoreSIGPIPE": "yes",
"InactiveEnterTimestampMonotonic": "0",
"InactiveExitTimestamp": "Sat 2024-08-24 12:34:19 EDT",
"InactiveExitTimestampMonotonic": "662546400",
"InvocationID": "4700a8e8b9b14f7194db3c9a057ff62b",
"JobRunningTimeoutUSec": "infinity",
"JobTimeoutAction": "none",
"JobTimeoutUSec": "infinity",
"KeyringMode": "private",
"KillMode": "mixed",
"KillSignal": "15",
"LimitAS": "infinity",
"LimitASSoft": "infinity",
"LimitCORE": "infinity",
"LimitCORESoft": "0",
"LimitCPU": "infinity",
"LimitCPUSoft": "infinity",
"LimitDATA": "infinity",
"LimitDATASoft": "infinity",
"LimitFSIZE": "infinity",
"LimitFSIZESoft": "infinity",
"LimitLOCKS": "infinity",
"LimitLOCKSSoft": "infinity",
"LimitMEMLOCK": "65536",
"LimitMEMLOCKSoft": "65536",
"LimitMSGQUEUE": "819200",
"LimitMSGQUEUESoft": "819200",
"LimitNICE": "0",
"LimitNICESoft": "0",
"LimitNOFILE": "262144",
"LimitNOFILESoft": "1024",
"LimitNPROC": "14003",
"LimitNPROCSoft": "14003",
"LimitRSS": "infinity",
"LimitRSSSoft": "infinity",
"LimitRTPRIO": "0",
"LimitRTPRIOSoft": "0",
"LimitRTTIME": "infinity",
"LimitRTTIMESoft": "infinity",
"LimitSIGPENDING": "14003",
"LimitSIGPENDINGSoft": "14003",
"LimitSTACK": "infinity",
"LimitSTACKSoft": "8388608",
"LoadState": "loaded",
"LockPersonality": "no",
"LogLevelMax": "-1",
"LogRateLimitBurst": "0",
"LogRateLimitIntervalUSec": "0",
"LogsDirectoryMode": "0755",
"MainPID": "80148",
"MemoryAccounting": "yes",
"MemoryCurrent": "39976960",
"MemoryDenyWriteExecute": "no",
"MemoryHigh": "infinity",
"MemoryLimit": "infinity",
"MemoryLow": "0",
"MemoryMax": "infinity",
"MemoryMin": "0",
"MemorySwapMax": "infinity",
"MountAPIVFS": "no",
"MountFlags": "",
"NFileDescriptorStore": "0",
"NRestarts": "0",
"NUMAMask": "",
"NUMAPolicy": "n/a",
"Names": "firewalld.service",
"NeedDaemonReload": "no",
"Nice": "0",
"NoNewPrivileges": "no",
"NonBlocking": "no",
"NotifyAccess": "none",
"OOMScoreAdjust": "0",
"OnFailureJobMode": "replace",
"PermissionsStartOnly": "no",
"Perpetual": "no",
"PrivateDevices": "no",
"PrivateMounts": "no",
"PrivateNetwork": "no",
"PrivateTmp": "no",
"PrivateUsers": "no",
"ProtectControlGroups": "no",
"ProtectHome": "no",
"ProtectKernelModules": "no",
"ProtectKernelTunables": "no",
"ProtectSystem": "no",
"RefuseManualStart": "no",
"RefuseManualStop": "no",
"RemainAfterExit": "no",
"RemoveIPC": "no",
"Requires": "dbus.socket system.slice sysinit.target",
"Restart": "no",
"RestartUSec": "100ms",
"RestrictNamespaces": "no",
"RestrictRealtime": "no",
"RestrictSUIDSGID": "no",
"Result": "success",
"RootDirectoryStartOnly": "no",
"RuntimeDirectoryMode": "0755",
"RuntimeDirectoryPreserve": "no",
"RuntimeMaxUSec": "infinity",
"SameProcessGroup": "no",
"SecureBits": "0",
"SendSIGHUP": "no",
"SendSIGKILL": "yes",
"Slice": "system.slice",
"StandardError": "null",
"StandardInput": "null",
"StandardInputData": "",
"StandardOutput": "null",
"StartLimitAction": "none",
"StartLimitBurst": "5",
"StartLimitIntervalUSec": "10s",
"StartupBlockIOWeight": "[not set]",
"StartupCPUShares": "[not set]",
"StartupCPUWeight": "[not set]",
"StartupIOWeight": "[not set]",
"StateChangeTimestamp": "Sat 2024-08-24 12:34:20 EDT",
"StateChangeTimestampMonotonic": "663046301",
"StateDirectoryMode": "0755",
"StatusErrno": "0",
"StopWhenUnneeded": "no",
"SubState": "running",
"SuccessAction": "none",
"SyslogFacility": "3",
"SyslogLevel": "6",
"SyslogLevelPrefix": "yes",
"SyslogPriority": "30",
"SystemCallErrorNumber": "0",
"TTYReset": "no",
"TTYVHangup": "no",
"TTYVTDisallocate": "no",
"TasksAccounting": "yes",
"TasksCurrent": "2",
"TasksMax": "22405",
"TimeoutStartUSec": "1min 30s",
"TimeoutStopUSec": "1min 30s",
"TimerSlackNSec": "50000",
"Transient": "no",
"Type": "dbus",
"UID": "[not set]",
"UMask": "0022",
"UnitFilePreset": "enabled",
"UnitFileState": "enabled",
"UtmpMode": "init",
"WantedBy": "multi-user.target",
"Wants": "network-pre.target",
"WatchdogTimestamp": "Sat 2024-08-24 12:34:20 EDT",
"WatchdogTimestampMonotonic": "663046298",
"WatchdogUSec": "0"
}
}
TASK [fedora.linux_system_roles.firewall : Enable and start firewalld service] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:28
Saturday 24 August 2024 12:35:38 -0400 (0:00:00.545) 0:01:46.788 *******
ok: [managed_node1] => {
"changed": false,
"enabled": true,
"name": "firewalld",
"state": "started",
"status": {
"ActiveEnterTimestamp": "Sat 2024-08-24 12:34:20 EDT",
"ActiveEnterTimestampMonotonic": "663046301",
"ActiveExitTimestampMonotonic": "0",
"ActiveState": "active",
"After": "dbus.service sysinit.target dbus.socket polkit.service system.slice basic.target",
"AllowIsolate": "no",
"AllowedCPUs": "",
"AllowedMemoryNodes": "",
"AmbientCapabilities": "",
"AssertResult": "yes",
"AssertTimestamp": "Sat 2024-08-24 12:34:19 EDT",
"AssertTimestampMonotonic": "662544922",
"Before": "shutdown.target multi-user.target network-pre.target",
"BlockIOAccounting": "no",
"BlockIOWeight": "[not set]",
"BusName": "org.fedoraproject.FirewallD1",
"CPUAccounting": "no",
"CPUAffinity": "",
"CPUAffinityFromNUMA": "no",
"CPUQuotaPerSecUSec": "infinity",
"CPUQuotaPeriodUSec": "infinity",
"CPUSchedulingPolicy": "0",
"CPUSchedulingPriority": "0",
"CPUSchedulingResetOnFork": "no",
"CPUShares": "[not set]",
"CPUUsageNSec": "[not set]",
"CPUWeight": "[not set]",
"CacheDirectoryMode": "0755",
"CanFreeze": "yes",
"CanIsolate": "no",
"CanReload": "yes",
"CanStart": "yes",
"CanStop": "yes",
"CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf",
"CollectMode": "inactive",
"ConditionResult": "yes",
"ConditionTimestamp": "Sat 2024-08-24 12:34:19 EDT",
"ConditionTimestampMonotonic": "662544921",
"ConfigurationDirectoryMode": "0755",
"Conflicts": "shutdown.target ip6tables.service ipset.service ebtables.service iptables.service nftables.service",
"ControlGroup": "/system.slice/firewalld.service",
"ControlPID": "0",
"DefaultDependencies": "yes",
"DefaultMemoryLow": "0",
"DefaultMemoryMin": "0",
"Delegate": "no",
"Description": "firewalld - dynamic firewall daemon",
"DevicePolicy": "auto",
"Documentation": "man:firewalld(1)",
"DynamicUser": "no",
"EffectiveCPUs": "",
"EffectiveMemoryNodes": "",
"EnvironmentFiles": "/etc/sysconfig/firewalld (ignore_errors=yes)",
"ExecMainCode": "0",
"ExecMainExitTimestampMonotonic": "0",
"ExecMainPID": "80148",
"ExecMainStartTimestamp": "Sat 2024-08-24 12:34:19 EDT",
"ExecMainStartTimestampMonotonic": "662546362",
"ExecMainStatus": "0",
"ExecReload": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"ExecStart": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"FailureAction": "none",
"FileDescriptorStoreMax": "0",
"FragmentPath": "/usr/lib/systemd/system/firewalld.service",
"FreezerState": "running",
"GID": "[not set]",
"GuessMainPID": "yes",
"IOAccounting": "no",
"IOSchedulingClass": "0",
"IOSchedulingPriority": "0",
"IOWeight": "[not set]",
"IPAccounting": "no",
"IPEgressBytes": "18446744073709551615",
"IPEgressPackets": "18446744073709551615",
"IPIngressBytes": "18446744073709551615",
"IPIngressPackets": "18446744073709551615",
"Id": "firewalld.service",
"IgnoreOnIsolate": "no",
"IgnoreSIGPIPE": "yes",
"InactiveEnterTimestampMonotonic": "0",
"InactiveExitTimestamp": "Sat 2024-08-24 12:34:19 EDT",
"InactiveExitTimestampMonotonic": "662546400",
"InvocationID": "4700a8e8b9b14f7194db3c9a057ff62b",
"JobRunningTimeoutUSec": "infinity",
"JobTimeoutAction": "none",
"JobTimeoutUSec": "infinity",
"KeyringMode": "private",
"KillMode": "mixed",
"KillSignal": "15",
"LimitAS": "infinity",
"LimitASSoft": "infinity",
"LimitCORE": "infinity",
"LimitCORESoft": "0",
"LimitCPU": "infinity",
"LimitCPUSoft": "infinity",
"LimitDATA": "infinity",
"LimitDATASoft": "infinity",
"LimitFSIZE": "infinity",
"LimitFSIZESoft": "infinity",
"LimitLOCKS": "infinity",
"LimitLOCKSSoft": "infinity",
"LimitMEMLOCK": "65536",
"LimitMEMLOCKSoft": "65536",
"LimitMSGQUEUE": "819200",
"LimitMSGQUEUESoft": "819200",
"LimitNICE": "0",
"LimitNICESoft": "0",
"LimitNOFILE": "262144",
"LimitNOFILESoft": "1024",
"LimitNPROC": "14003",
"LimitNPROCSoft": "14003",
"LimitRSS": "infinity",
"LimitRSSSoft": "infinity",
"LimitRTPRIO": "0",
"LimitRTPRIOSoft": "0",
"LimitRTTIME": "infinity",
"LimitRTTIMESoft": "infinity",
"LimitSIGPENDING": "14003",
"LimitSIGPENDINGSoft": "14003",
"LimitSTACK": "infinity",
"LimitSTACKSoft": "8388608",
"LoadState": "loaded",
"LockPersonality": "no",
"LogLevelMax": "-1",
"LogRateLimitBurst": "0",
"LogRateLimitIntervalUSec": "0",
"LogsDirectoryMode": "0755",
"MainPID": "80148",
"MemoryAccounting": "yes",
"MemoryCurrent": "39976960",
"MemoryDenyWriteExecute": "no",
"MemoryHigh": "infinity",
"MemoryLimit": "infinity",
"MemoryLow": "0",
"MemoryMax": "infinity",
"MemoryMin": "0",
"MemorySwapMax": "infinity",
"MountAPIVFS": "no",
"MountFlags": "",
"NFileDescriptorStore": "0",
"NRestarts": "0",
"NUMAMask": "",
"NUMAPolicy": "n/a",
"Names": "firewalld.service",
"NeedDaemonReload": "no",
"Nice": "0",
"NoNewPrivileges": "no",
"NonBlocking": "no",
"NotifyAccess": "none",
"OOMScoreAdjust": "0",
"OnFailureJobMode": "replace",
"PermissionsStartOnly": "no",
"Perpetual": "no",
"PrivateDevices": "no",
"PrivateMounts": "no",
"PrivateNetwork": "no",
"PrivateTmp": "no",
"PrivateUsers": "no",
"ProtectControlGroups": "no",
"ProtectHome": "no",
"ProtectKernelModules": "no",
"ProtectKernelTunables": "no",
"ProtectSystem": "no",
"RefuseManualStart": "no",
"RefuseManualStop": "no",
"RemainAfterExit": "no",
"RemoveIPC": "no",
"Requires": "dbus.socket system.slice sysinit.target",
"Restart": "no",
"RestartUSec": "100ms",
"RestrictNamespaces": "no",
"RestrictRealtime": "no",
"RestrictSUIDSGID": "no",
"Result": "success",
"RootDirectoryStartOnly": "no",
"RuntimeDirectoryMode": "0755",
"RuntimeDirectoryPreserve": "no",
"RuntimeMaxUSec": "infinity",
"SameProcessGroup": "no",
"SecureBits": "0",
"SendSIGHUP": "no",
"SendSIGKILL": "yes",
"Slice": "system.slice",
"StandardError": "null",
"StandardInput": "null",
"StandardInputData": "",
"StandardOutput": "null",
"StartLimitAction": "none",
"StartLimitBurst": "5",
"StartLimitIntervalUSec": "10s",
"StartupBlockIOWeight": "[not set]",
"StartupCPUShares": "[not set]",
"StartupCPUWeight": "[not set]",
"StartupIOWeight": "[not set]",
"StateChangeTimestamp": "Sat 2024-08-24 12:34:20 EDT",
"StateChangeTimestampMonotonic": "663046301",
"StateDirectoryMode": "0755",
"StatusErrno": "0",
"StopWhenUnneeded": "no",
"SubState": "running",
"SuccessAction": "none",
"SyslogFacility": "3",
"SyslogLevel": "6",
"SyslogLevelPrefix": "yes",
"SyslogPriority": "30",
"SystemCallErrorNumber": "0",
"TTYReset": "no",
"TTYVHangup": "no",
"TTYVTDisallocate": "no",
"TasksAccounting": "yes",
"TasksCurrent": "2",
"TasksMax": "22405",
"TimeoutStartUSec": "1min 30s",
"TimeoutStopUSec": "1min 30s",
"TimerSlackNSec": "50000",
"Transient": "no",
"Type": "dbus",
"UID": "[not set]",
"UMask": "0022",
"UnitFilePreset": "enabled",
"UnitFileState": "enabled",
"UtmpMode": "init",
"WantedBy": "multi-user.target",
"Wants": "network-pre.target",
"WatchdogTimestamp": "Sat 2024-08-24 12:34:20 EDT",
"WatchdogTimestampMonotonic": "663046298",
"WatchdogUSec": "0"
}
}
TASK [fedora.linux_system_roles.firewall : Check if previous replaced is defined] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:34
Saturday 24 August 2024 12:35:38 -0400 (0:00:00.518) 0:01:47.307 *******
ok: [managed_node1] => {
"ansible_facts": {
"__firewall_previous_replaced": false,
"__firewall_python_cmd": "/usr/libexec/platform-python",
"__firewall_report_changed": true
},
"changed": false
}
TASK [fedora.linux_system_roles.firewall : Get config files, checksums before and remove] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:43
Saturday 24 August 2024 12:35:38 -0400 (0:00:00.044) 0:01:47.351 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.firewall : Tell firewall module it is able to report changed] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:55
Saturday 24 August 2024 12:35:38 -0400 (0:00:00.039) 0:01:47.391 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.firewall : Configure firewall] *****************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:71
Saturday 24 August 2024 12:35:38 -0400 (0:00:00.037) 0:01:47.429 *******
ok: [managed_node1] => (item={'port': '8000/tcp', 'state': 'enabled'}) => {
"__firewall_changed": false,
"ansible_loop_var": "item",
"changed": false,
"item": {
"port": "8000/tcp",
"state": "enabled"
}
}
ok: [managed_node1] => (item={'port': '9000/tcp', 'state': 'enabled'}) => {
"__firewall_changed": false,
"ansible_loop_var": "item",
"changed": false,
"item": {
"port": "9000/tcp",
"state": "enabled"
}
}
TASK [fedora.linux_system_roles.firewall : Gather firewall config information] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:120
Saturday 24 August 2024 12:35:39 -0400 (0:00:01.142) 0:01:48.571 *******
skipping: [managed_node1] => (item={'port': '8000/tcp', 'state': 'enabled'}) => {
"ansible_loop_var": "item",
"changed": false,
"item": {
"port": "8000/tcp",
"state": "enabled"
},
"skip_reason": "Conditional result was False"
}
skipping: [managed_node1] => (item={'port': '9000/tcp', 'state': 'enabled'}) => {
"ansible_loop_var": "item",
"changed": false,
"item": {
"port": "9000/tcp",
"state": "enabled"
},
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.firewall : Update firewalld_config fact] *******
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:130
Saturday 24 August 2024 12:35:40 -0400 (0:00:00.050) 0:01:48.622 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.firewall : Gather firewall config if no arguments] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:139
Saturday 24 August 2024 12:35:40 -0400 (0:00:00.034) 0:01:48.657 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.firewall : Update firewalld_config fact] *******
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:144
Saturday 24 August 2024 12:35:40 -0400 (0:00:00.034) 0:01:48.691 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.firewall : Get config files, checksums after] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:153
Saturday 24 August 2024 12:35:40 -0400 (0:00:00.033) 0:01:48.724 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.firewall : Calculate what has changed] *********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:163
Saturday 24 August 2024 12:35:40 -0400 (0:00:00.032) 0:01:48.756 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.firewall : Show diffs] *************************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:169
Saturday 24 August 2024 12:35:40 -0400 (0:00:00.031) 0:01:48.788 *******
skipping: [managed_node1] => {}
TASK [Manage selinux for specified ports] **************************************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:117
Saturday 24 August 2024 12:35:40 -0400 (0:00:00.031) 0:01:48.819 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Keep track of users that need to cancel linger] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:124
Saturday 24 August 2024 12:35:40 -0400 (0:00:00.031) 0:01:48.850 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_cancel_user_linger": []
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Handle certs.d files - present] *******
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:128
Saturday 24 August 2024 12:35:40 -0400 (0:00:00.033) 0:01:48.884 *******
skipping: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : Handle credential files - present] ****
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:137
Saturday 24 August 2024 12:35:40 -0400 (0:00:00.029) 0:01:48.914 *******
skipping: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : Handle secrets] ***********************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:146
Saturday 24 August 2024 12:35:40 -0400 (0:00:00.028) 0:01:48.942 *******
included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml for managed_node1
included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml for managed_node1
included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml for managed_node1
TASK [fedora.linux_system_roles.podman : Set variables part 1] *****************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:3
Saturday 24 August 2024 12:35:40 -0400 (0:00:00.131) 0:01:49.074 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_user": "root"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set variables part 2] *****************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:7
Saturday 24 August 2024 12:35:40 -0400 (0:00:00.039) 0:01:49.113 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_rootless": false,
"__podman_xdg_runtime_dir": "/run/user/0"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Manage linger] ************************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:13
Saturday 24 August 2024 12:35:40 -0400 (0:00:00.120) 0:01:49.234 *******
included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for managed_node1
TASK [fedora.linux_system_roles.podman : Enable linger if needed] **************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12
Saturday 24 August 2024 12:35:40 -0400 (0:00:00.057) 0:01:49.291 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18
Saturday 24 August 2024 12:35:40 -0400 (0:00:00.032) 0:01:49.324 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22
Saturday 24 August 2024 12:35:40 -0400 (0:00:00.032) 0:01:49.356 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Stat XDG_RUNTIME_DIR] *****************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:18
Saturday 24 August 2024 12:35:40 -0400 (0:00:00.030) 0:01:49.387 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Manage each secret] *******************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:34
Saturday 24 August 2024 12:35:40 -0400 (0:00:00.035) 0:01:49.422 *******
[WARNING]: Using a variable for a task's 'args' is unsafe in some situations
(see
https://docs.ansible.com/ansible/devel/reference_appendices/faq.html#argsplat-
unsafe)
changed: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": true
}
TASK [fedora.linux_system_roles.podman : Set variables part 1] *****************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:3
Saturday 24 August 2024 12:35:41 -0400 (0:00:00.424) 0:01:49.847 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_user": "root"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set variables part 2] *****************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:7
Saturday 24 August 2024 12:35:41 -0400 (0:00:00.038) 0:01:49.885 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_rootless": false,
"__podman_xdg_runtime_dir": "/run/user/0"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Manage linger] ************************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:13
Saturday 24 August 2024 12:35:41 -0400 (0:00:00.044) 0:01:49.929 *******
included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for managed_node1
TASK [fedora.linux_system_roles.podman : Enable linger if needed] **************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12
Saturday 24 August 2024 12:35:41 -0400 (0:00:00.086) 0:01:50.016 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18
Saturday 24 August 2024 12:35:41 -0400 (0:00:00.049) 0:01:50.066 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22
Saturday 24 August 2024 12:35:41 -0400 (0:00:00.050) 0:01:50.117 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Stat XDG_RUNTIME_DIR] *****************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:18
Saturday 24 August 2024 12:35:41 -0400 (0:00:00.049) 0:01:50.167 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Manage each secret] *******************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:34
Saturday 24 August 2024 12:35:41 -0400 (0:00:00.038) 0:01:50.206 *******
[WARNING]: Using a variable for a task's 'args' is unsafe in some situations
(see
https://docs.ansible.com/ansible/devel/reference_appendices/faq.html#argsplat-
unsafe)
changed: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": true
}
TASK [fedora.linux_system_roles.podman : Set variables part 1] *****************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:3
Saturday 24 August 2024 12:35:42 -0400 (0:00:00.414) 0:01:50.620 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_user": "root"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set variables part 2] *****************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:7
Saturday 24 August 2024 12:35:42 -0400 (0:00:00.042) 0:01:50.662 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_rootless": false,
"__podman_xdg_runtime_dir": "/run/user/0"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Manage linger] ************************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:13
Saturday 24 August 2024 12:35:42 -0400 (0:00:00.055) 0:01:50.718 *******
included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for managed_node1
TASK [fedora.linux_system_roles.podman : Enable linger if needed] **************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12
Saturday 24 August 2024 12:35:42 -0400 (0:00:00.056) 0:01:50.774 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18
Saturday 24 August 2024 12:35:42 -0400 (0:00:00.031) 0:01:50.806 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22
Saturday 24 August 2024 12:35:42 -0400 (0:00:00.031) 0:01:50.837 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Stat XDG_RUNTIME_DIR] *****************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:18
Saturday 24 August 2024 12:35:42 -0400 (0:00:00.030) 0:01:50.868 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Manage each secret] *******************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:34
Saturday 24 August 2024 12:35:42 -0400 (0:00:00.035) 0:01:50.904 *******
[WARNING]: Using a variable for a task's 'args' is unsafe in some situations
(see
https://docs.ansible.com/ansible/devel/reference_appendices/faq.html#argsplat-
unsafe)
changed: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": true
}
TASK [fedora.linux_system_roles.podman : Handle Kubernetes specifications] *****
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:153
Saturday 24 August 2024 12:35:42 -0400 (0:00:00.404) 0:01:51.308 *******
skipping: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : Handle Quadlet specifications] ********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:160
Saturday 24 August 2024 12:35:42 -0400 (0:00:00.030) 0:01:51.339 *******
included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for managed_node1
included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for managed_node1
included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for managed_node1
included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for managed_node1
included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for managed_node1
included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for managed_node1
TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14
Saturday 24 August 2024 12:35:42 -0400 (0:00:00.136) 0:01:51.476 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_quadlet_file_src": "quadlet-demo.kube",
"__podman_quadlet_spec": {},
"__podman_quadlet_str": "[Install]\nWantedBy=default.target\n\n[Unit]\nRequires=quadlet-demo-mysql.service\nAfter=quadlet-demo-mysql.service\n\n[Kube]\n# Point to the yaml file in the same directory\nYaml=quadlet-demo.yml\n# Use the quadlet-demo network\nNetwork=quadlet-demo.network\n# Publish the envoy proxy data port\nPublishPort=8000:8080\n# Publish the envoy proxy admin port\nPublishPort=9000:9901\n# Use the envoy proxy config map in the same directory\nConfigMap=envoy-proxy-configmap.yml",
"__podman_quadlet_template_src": ""
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25
Saturday 24 August 2024 12:35:42 -0400 (0:00:00.045) 0:01:51.521 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_continue_if_pull_fails": false,
"__podman_pull_image": true,
"__podman_state": "absent",
"__podman_systemd_unit_scope": "",
"__podman_user": "root"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] *****
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35
Saturday 24 August 2024 12:35:43 -0400 (0:00:00.099) 0:01:51.621 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48
Saturday 24 August 2024 12:35:43 -0400 (0:00:00.053) 0:01:51.674 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_quadlet_name": "quadlet-demo",
"__podman_quadlet_type": "kube",
"__podman_rootless": false
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Check user and group information] *****
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57
Saturday 24 August 2024 12:35:43 -0400 (0:00:00.074) 0:01:51.748 *******
included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for managed_node1
TASK [fedora.linux_system_roles.podman : Get user information] *****************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2
Saturday 24 August 2024 12:35:43 -0400 (0:00:00.078) 0:01:51.826 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Fail if user does not exist] **********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9
Saturday 24 August 2024 12:35:43 -0400 (0:00:00.048) 0:01:51.874 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set group for podman user] ************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16
Saturday 24 August 2024 12:35:43 -0400 (0:00:00.053) 0:01:51.928 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_group": "0"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Get group information] ****************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:28
Saturday 24 August 2024 12:35:43 -0400 (0:00:00.048) 0:01:51.977 *******
ok: [managed_node1] => {
"ansible_facts": {
"getent_group": {
"root": [
"x",
"0",
""
]
}
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set group name] ***********************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:35
Saturday 24 August 2024 12:35:43 -0400 (0:00:00.389) 0:01:52.366 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_group_name": "root"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : See if getsubids exists] **************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:39
Saturday 24 August 2024 12:35:43 -0400 (0:00:00.066) 0:01:52.432 *******
ok: [managed_node1] => {
"changed": false,
"stat": {
"atime": 1724516929.7168462,
"attr_flags": "",
"attributes": [],
"block_size": 4096,
"blocks": 32,
"charset": "binary",
"checksum": "bb5b46ffbafcaa8c4021f3c8b3cb8594f48ef34b",
"ctime": 1724516896.725572,
"dev": 51713,
"device_type": 0,
"executable": true,
"exists": true,
"gid": 0,
"gr_name": "root",
"inode": 6986657,
"isblk": false,
"ischr": false,
"isdir": false,
"isfifo": false,
"isgid": false,
"islnk": false,
"isreg": true,
"issock": false,
"isuid": false,
"mimetype": "application/x-sharedlib",
"mode": "0755",
"mtime": 1700557386.0,
"nlink": 1,
"path": "/usr/bin/getsubids",
"pw_name": "root",
"readable": true,
"rgrp": true,
"roth": true,
"rusr": true,
"size": 12640,
"uid": 0,
"version": "1255679238",
"wgrp": false,
"woth": false,
"writeable": true,
"wusr": true,
"xgrp": true,
"xoth": true,
"xusr": true
}
}
TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:50
Saturday 24 August 2024 12:35:44 -0400 (0:00:00.377) 0:01:52.810 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Check group with getsubids] ***********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:55
Saturday 24 August 2024 12:35:44 -0400 (0:00:00.046) 0:01:52.857 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ******
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:60
Saturday 24 August 2024 12:35:44 -0400 (0:00:00.052) 0:01:52.910 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Get subuid file] **********************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:74
Saturday 24 August 2024 12:35:44 -0400 (0:00:00.043) 0:01:52.953 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Get subgid file] **********************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:79
Saturday 24 August 2024 12:35:44 -0400 (0:00:00.038) 0:01:52.992 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ******
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:84
Saturday 24 August 2024 12:35:44 -0400 (0:00:00.040) 0:01:53.032 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ******
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:94
Saturday 24 August 2024 12:35:44 -0400 (0:00:00.078) 0:01:53.111 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] *****
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:101
Saturday 24 August 2024 12:35:44 -0400 (0:00:00.032) 0:01:53.143 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62
Saturday 24 August 2024 12:35:44 -0400 (0:00:00.032) 0:01:53.176 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_activate_systemd_unit": true,
"__podman_images_found": [],
"__podman_kube_yamls_raw": [
"quadlet-demo.yml"
],
"__podman_service_name": "quadlet-demo.service",
"__podman_systemd_scope": "system",
"__podman_user_home_dir": "/root",
"__podman_xdg_runtime_dir": "/run/user/0"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73
Saturday 24 August 2024 12:35:44 -0400 (0:00:00.055) 0:01:53.231 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_quadlet_path": "/etc/containers/systemd"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Get kube yaml contents] ***************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77
Saturday 24 August 2024 12:35:44 -0400 (0:00:00.037) 0:01:53.268 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87
Saturday 24 August 2024 12:35:44 -0400 (0:00:00.050) 0:01:53.318 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_images": [],
"__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo.kube",
"__podman_volumes": []
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 6] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:103
Saturday 24 August 2024 12:35:44 -0400 (0:00:00.088) 0:01:53.407 *******
ok: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : Cleanup quadlets] *********************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:110
Saturday 24 August 2024 12:35:44 -0400 (0:00:00.046) 0:01:53.453 *******
included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml for managed_node1
TASK [fedora.linux_system_roles.podman : Stat XDG_RUNTIME_DIR] *****************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:4
Saturday 24 August 2024 12:35:44 -0400 (0:00:00.083) 0:01:53.537 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Stop and disable service] *************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:12
Saturday 24 August 2024 12:35:44 -0400 (0:00:00.031) 0:01:53.568 *******
changed: [managed_node1] => {
"changed": true,
"enabled": false,
"failed_when_result": false,
"name": "quadlet-demo.service",
"state": "stopped",
"status": {
"ActiveEnterTimestamp": "Sat 2024-08-24 12:35:23 EDT",
"ActiveEnterTimestampMonotonic": "726499194",
"ActiveExitTimestampMonotonic": "0",
"ActiveState": "active",
"After": "systemd-journald.socket basic.target -.mount quadlet-demo-network.service system.slice quadlet-demo-mysql.service sysinit.target",
"AllowIsolate": "no",
"AllowedCPUs": "",
"AllowedMemoryNodes": "",
"AmbientCapabilities": "",
"AssertResult": "yes",
"AssertTimestamp": "Sat 2024-08-24 12:35:22 EDT",
"AssertTimestampMonotonic": "725568879",
"Before": "shutdown.target",
"BlockIOAccounting": "no",
"BlockIOWeight": "[not set]",
"CPUAccounting": "no",
"CPUAffinity": "",
"CPUAffinityFromNUMA": "no",
"CPUQuotaPerSecUSec": "infinity",
"CPUQuotaPeriodUSec": "infinity",
"CPUSchedulingPolicy": "0",
"CPUSchedulingPriority": "0",
"CPUSchedulingResetOnFork": "no",
"CPUShares": "[not set]",
"CPUUsageNSec": "[not set]",
"CPUWeight": "[not set]",
"CacheDirectoryMode": "0755",
"CanFreeze": "yes",
"CanIsolate": "no",
"CanReload": "no",
"CanStart": "yes",
"CanStop": "yes",
"CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf",
"CollectMode": "inactive",
"ConditionResult": "yes",
"ConditionTimestamp": "Sat 2024-08-24 12:35:22 EDT",
"ConditionTimestampMonotonic": "725568878",
"ConfigurationDirectoryMode": "0755",
"Conflicts": "shutdown.target",
"ControlGroup": "/system.slice/quadlet-demo.service",
"ControlPID": "0",
"DefaultDependencies": "yes",
"DefaultMemoryLow": "0",
"DefaultMemoryMin": "0",
"Delegate": "no",
"Description": "quadlet-demo.service",
"DevicePolicy": "auto",
"DynamicUser": "no",
"EffectiveCPUs": "",
"EffectiveMemoryNodes": "",
"Environment": "PODMAN_SYSTEMD_UNIT=quadlet-demo.service",
"ExecMainCode": "0",
"ExecMainExitTimestampMonotonic": "0",
"ExecMainPID": "87597",
"ExecMainStartTimestamp": "Sat 2024-08-24 12:35:23 EDT",
"ExecMainStartTimestampMonotonic": "726499171",
"ExecMainStatus": "0",
"ExecStart": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman kube play --replace --service-container=true --network=systemd-quadlet-demo --configmap /etc/containers/systemd/envoy-proxy-configmap.yml --publish 8000:8080 --publish 9000:9901 /etc/containers/systemd/quadlet-demo.yml ; ignore_errors=no ; start_time=[Sat 2024-08-24 12:35:22 EDT] ; stop_time=[n/a] ; pid=87537 ; code=(null) ; status=0/0 }",
"ExecStopPost": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman kube down /etc/containers/systemd/quadlet-demo.yml ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"FailureAction": "none",
"FileDescriptorStoreMax": "0",
"FragmentPath": "/run/systemd/generator/quadlet-demo.service",
"FreezerState": "running",
"GID": "[not set]",
"GuessMainPID": "yes",
"IOAccounting": "no",
"IOSchedulingClass": "0",
"IOSchedulingPriority": "0",
"IOWeight": "[not set]",
"IPAccounting": "no",
"IPEgressBytes": "18446744073709551615",
"IPEgressPackets": "18446744073709551615",
"IPIngressBytes": "18446744073709551615",
"IPIngressPackets": "18446744073709551615",
"Id": "quadlet-demo.service",
"IgnoreOnIsolate": "no",
"IgnoreSIGPIPE": "yes",
"InactiveEnterTimestampMonotonic": "0",
"InactiveExitTimestamp": "Sat 2024-08-24 12:35:22 EDT",
"InactiveExitTimestampMonotonic": "725570197",
"InvocationID": "67d7f03a2403424ab0f3e8a91570703c",
"JobRunningTimeoutUSec": "infinity",
"JobTimeoutAction": "none",
"JobTimeoutUSec": "infinity",
"KeyringMode": "private",
"KillMode": "mixed",
"KillSignal": "15",
"LimitAS": "infinity",
"LimitASSoft": "infinity",
"LimitCORE": "infinity",
"LimitCORESoft": "0",
"LimitCPU": "infinity",
"LimitCPUSoft": "infinity",
"LimitDATA": "infinity",
"LimitDATASoft": "infinity",
"LimitFSIZE": "infinity",
"LimitFSIZESoft": "infinity",
"LimitLOCKS": "infinity",
"LimitLOCKSSoft": "infinity",
"LimitMEMLOCK": "65536",
"LimitMEMLOCKSoft": "65536",
"LimitMSGQUEUE": "819200",
"LimitMSGQUEUESoft": "819200",
"LimitNICE": "0",
"LimitNICESoft": "0",
"LimitNOFILE": "262144",
"LimitNOFILESoft": "1024",
"LimitNPROC": "14003",
"LimitNPROCSoft": "14003",
"LimitRSS": "infinity",
"LimitRSSSoft": "infinity",
"LimitRTPRIO": "0",
"LimitRTPRIOSoft": "0",
"LimitRTTIME": "infinity",
"LimitRTTIMESoft": "infinity",
"LimitSIGPENDING": "14003",
"LimitSIGPENDINGSoft": "14003",
"LimitSTACK": "infinity",
"LimitSTACKSoft": "8388608",
"LoadState": "loaded",
"LockPersonality": "no",
"LogLevelMax": "-1",
"LogRateLimitBurst": "0",
"LogRateLimitIntervalUSec": "0",
"LogsDirectoryMode": "0755",
"MainPID": "87597",
"MemoryAccounting": "yes",
"MemoryCurrent": "5255168",
"MemoryDenyWriteExecute": "no",
"MemoryHigh": "infinity",
"MemoryLimit": "infinity",
"MemoryLow": "0",
"MemoryMax": "infinity",
"MemoryMin": "0",
"MemorySwapMax": "infinity",
"MountAPIVFS": "no",
"MountFlags": "",
"NFileDescriptorStore": "0",
"NRestarts": "0",
"NUMAMask": "",
"NUMAPolicy": "n/a",
"Names": "quadlet-demo.service",
"NeedDaemonReload": "no",
"Nice": "0",
"NoNewPrivileges": "no",
"NonBlocking": "no",
"NotifyAccess": "all",
"OOMScoreAdjust": "0",
"OnFailureJobMode": "replace",
"PermissionsStartOnly": "no",
"Perpetual": "no",
"PrivateDevices": "no",
"PrivateMounts": "no",
"PrivateNetwork": "no",
"PrivateTmp": "no",
"PrivateUsers": "no",
"ProtectControlGroups": "no",
"ProtectHome": "no",
"ProtectKernelModules": "no",
"ProtectKernelTunables": "no",
"ProtectSystem": "no",
"RefuseManualStart": "no",
"RefuseManualStop": "no",
"RemainAfterExit": "no",
"RemoveIPC": "no",
"Requires": "-.mount quadlet-demo-mysql.service sysinit.target quadlet-demo-network.service system.slice",
"RequiresMountsFor": "/run/containers",
"Restart": "no",
"RestartUSec": "100ms",
"RestrictNamespaces": "no",
"RestrictRealtime": "no",
"RestrictSUIDSGID": "no",
"Result": "success",
"RootDirectoryStartOnly": "no",
"RuntimeDirectoryMode": "0755",
"RuntimeDirectoryPreserve": "no",
"RuntimeMaxUSec": "infinity",
"SameProcessGroup": "no",
"SecureBits": "0",
"SendSIGHUP": "no",
"SendSIGKILL": "yes",
"Slice": "system.slice",
"SourcePath": "/etc/containers/systemd/quadlet-demo.kube",
"StandardError": "inherit",
"StandardInput": "null",
"StandardInputData": "",
"StandardOutput": "journal",
"StartLimitAction": "none",
"StartLimitBurst": "5",
"StartLimitIntervalUSec": "10s",
"StartupBlockIOWeight": "[not set]",
"StartupCPUShares": "[not set]",
"StartupCPUWeight": "[not set]",
"StartupIOWeight": "[not set]",
"StateChangeTimestamp": "Sat 2024-08-24 12:35:23 EDT",
"StateChangeTimestampMonotonic": "726499194",
"StateDirectoryMode": "0755",
"StatusErrno": "0",
"StopWhenUnneeded": "no",
"SubState": "running",
"SuccessAction": "none",
"SyslogFacility": "3",
"SyslogIdentifier": "quadlet-demo",
"SyslogLevel": "6",
"SyslogLevelPrefix": "yes",
"SyslogPriority": "30",
"SystemCallErrorNumber": "0",
"TTYReset": "no",
"TTYVHangup": "no",
"TTYVTDisallocate": "no",
"TasksAccounting": "yes",
"TasksCurrent": "4",
"TasksMax": "22405",
"TimeoutStartUSec": "1min 30s",
"TimeoutStopUSec": "1min 30s",
"TimerSlackNSec": "50000",
"Transient": "no",
"Type": "notify",
"UID": "[not set]",
"UMask": "0022",
"UnitFilePreset": "disabled",
"UnitFileState": "generated",
"UtmpMode": "init",
"WatchdogTimestamp": "Sat 2024-08-24 12:35:23 EDT",
"WatchdogTimestampMonotonic": "726499191",
"WatchdogUSec": "0"
}
}
TASK [fedora.linux_system_roles.podman : See if quadlet file exists] ***********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:33
Saturday 24 August 2024 12:35:48 -0400 (0:00:03.207) 0:01:56.776 *******
ok: [managed_node1] => {
"changed": false,
"stat": {
"atime": 1724517322.2370505,
"attr_flags": "",
"attributes": [],
"block_size": 4096,
"blocks": 8,
"charset": "us-ascii",
"checksum": "7a5c73a5d935a42431c87bcdbeb8a04ed0909dc7",
"ctime": 1724517321.5960455,
"dev": 51713,
"device_type": 0,
"executable": false,
"exists": true,
"gid": 0,
"gr_name": "root",
"inode": 497033403,
"isblk": false,
"ischr": false,
"isdir": false,
"isfifo": false,
"isgid": false,
"islnk": false,
"isreg": true,
"issock": false,
"isuid": false,
"mimetype": "text/plain",
"mode": "0644",
"mtime": 1724517321.3340435,
"nlink": 1,
"path": "/etc/containers/systemd/quadlet-demo.kube",
"pw_name": "root",
"readable": true,
"rgrp": true,
"roth": true,
"rusr": true,
"size": 456,
"uid": 0,
"version": "3427625058",
"wgrp": false,
"woth": false,
"writeable": true,
"wusr": true,
"xgrp": false,
"xoth": false,
"xusr": false
}
}
TASK [fedora.linux_system_roles.podman : Parse quadlet file] *******************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:38
Saturday 24 August 2024 12:35:48 -0400 (0:00:00.399) 0:01:57.175 *******
included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml for managed_node1
TASK [fedora.linux_system_roles.podman : Slurp quadlet file] *******************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml:6
Saturday 24 August 2024 12:35:48 -0400 (0:00:00.058) 0:01:57.234 *******
ok: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : Parse quadlet file] *******************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml:12
Saturday 24 August 2024 12:35:49 -0400 (0:00:00.388) 0:01:57.623 *******
ok: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : Parse quadlet yaml file] **************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml:44
Saturday 24 August 2024 12:35:49 -0400 (0:00:00.075) 0:01:57.698 *******
skipping: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : Reset raw variable] *******************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml:52
Saturday 24 August 2024 12:35:49 -0400 (0:00:00.051) 0:01:57.750 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_quadlet_raw": null
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Remove quadlet file] ******************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:42
Saturday 24 August 2024 12:35:49 -0400 (0:00:00.051) 0:01:57.801 *******
changed: [managed_node1] => {
"changed": true,
"path": "/etc/containers/systemd/quadlet-demo.kube",
"state": "absent"
}
TASK [fedora.linux_system_roles.podman : Refresh systemd] **********************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:48
Saturday 24 August 2024 12:35:49 -0400 (0:00:00.451) 0:01:58.252 *******
ok: [managed_node1] => {
"changed": false,
"name": null,
"status": {}
}
TASK [fedora.linux_system_roles.podman : Remove managed resource] **************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:58
Saturday 24 August 2024 12:35:50 -0400 (0:00:00.606) 0:01:58.859 *******
skipping: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : Remove volumes] ***********************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:95
Saturday 24 August 2024 12:35:50 -0400 (0:00:00.041) 0:01:58.900 *******
skipping: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : Clear parsed podman variable] *********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:112
Saturday 24 August 2024 12:35:50 -0400 (0:00:00.054) 0:01:58.955 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_quadlet_parsed": null
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Prune images no longer in use] ********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:116
Saturday 24 August 2024 12:35:50 -0400 (0:00:00.037) 0:01:58.992 *******
changed: [managed_node1] => {
"changed": true,
"cmd": [
"podman",
"image",
"prune",
"--all",
"-f"
],
"delta": "0:00:00.721014",
"end": "2024-08-24 12:35:51.408908",
"rc": 0,
"start": "2024-08-24 12:35:50.687894"
}
STDOUT:
0030ba3d620c647159c935ee778991c68ef3e51a274703753b0bc530104ef5e5
9f9ec7f2fdef9168f74e9d057f307955db14d782cff22ded51d277d74798cb2f
fc3855720f39e33996597ca2164a27fa28c6b03eb8774fdb5249a1c2244ef9fd
fcf3e41b8864a14d75a6d0627d3d02154e28a153aa57e8baa392cd744ffa0d0b
5af2585e22ed1562885d9407efab74010090427be79048c2cd6a226517cc1e1d
TASK [fedora.linux_system_roles.podman : Manage linger] ************************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:127
Saturday 24 August 2024 12:35:51 -0400 (0:00:01.110) 0:02:00.103 *******
included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for managed_node1
TASK [fedora.linux_system_roles.podman : Enable linger if needed] **************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12
Saturday 24 August 2024 12:35:51 -0400 (0:00:00.083) 0:02:00.186 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18
Saturday 24 August 2024 12:35:51 -0400 (0:00:00.039) 0:02:00.225 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22
Saturday 24 August 2024 12:35:51 -0400 (0:00:00.041) 0:02:00.267 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : For testing and debugging - images] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:137
Saturday 24 August 2024 12:35:51 -0400 (0:00:00.035) 0:02:00.303 *******
ok: [managed_node1] => {
"changed": false,
"cmd": [
"podman",
"images",
"-n"
],
"delta": "0:00:00.033505",
"end": "2024-08-24 12:35:52.020143",
"rc": 0,
"start": "2024-08-24 12:35:51.986638"
}
STDOUT:
quay.io/linux-system-roles/mysql 5.6 dd3b2a5dcb48 2 years ago 308 MB
TASK [fedora.linux_system_roles.podman : For testing and debugging - volumes] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:146
Saturday 24 August 2024 12:35:52 -0400 (0:00:00.403) 0:02:00.707 *******
ok: [managed_node1] => {
"changed": false,
"cmd": [
"podman",
"volume",
"ls",
"-n"
],
"delta": "0:00:00.030150",
"end": "2024-08-24 12:35:52.451392",
"rc": 0,
"start": "2024-08-24 12:35:52.421242"
}
STDOUT:
local systemd-quadlet-demo-mysql
local wp-pv-claim
local envoy-proxy-config
local envoy-certificates
TASK [fedora.linux_system_roles.podman : For testing and debugging - containers] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:155
Saturday 24 August 2024 12:35:52 -0400 (0:00:00.425) 0:02:01.132 *******
ok: [managed_node1] => {
"changed": false,
"cmd": [
"podman",
"ps",
"--noheading"
],
"delta": "0:00:00.035408",
"end": "2024-08-24 12:35:52.888353",
"rc": 0,
"start": "2024-08-24 12:35:52.852945"
}
STDOUT:
68250d5fc021 quay.io/linux-system-roles/mysql:5.6 mysqld About a minute ago Up About a minute (healthy) quadlet-demo-mysql
TASK [fedora.linux_system_roles.podman : For testing and debugging - networks] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:164
Saturday 24 August 2024 12:35:52 -0400 (0:00:00.436) 0:02:01.569 *******
ok: [managed_node1] => {
"changed": false,
"cmd": [
"podman",
"network",
"ls",
"-n",
"-q"
],
"delta": "0:00:00.055656",
"end": "2024-08-24 12:35:53.333009",
"rc": 0,
"start": "2024-08-24 12:35:53.277353"
}
STDOUT:
podman
podman-default-kube-network
systemd-quadlet-demo
TASK [fedora.linux_system_roles.podman : For testing and debugging - secrets] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:173
Saturday 24 August 2024 12:35:53 -0400 (0:00:00.454) 0:02:02.024 *******
ok: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : For testing and debugging - services] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:183
Saturday 24 August 2024 12:35:53 -0400 (0:00:00.421) 0:02:02.445 *******
ok: [managed_node1] => {
"ansible_facts": {
"services": {
"68250d5fc0218bc6db3ab2...5ebe15c37a833d3345ad2e2df2c1916dd6cc3.service": {
"name": "68250d5fc0218bc6db3ab2...5ebe15c37a833d3345ad2e2df2c1916dd6cc3.service",
"source": "systemd",
"state": "inactive",
"status": "transient"
},
"68250d5fc0218bc6db3ab2ece605ebe15c37a833d3345ad2e2df2c1916dd6cc3.service": {
"name": "68250d5fc0218bc6db3ab2ece605ebe15c37a833d3345ad2e2df2c1916dd6cc3.service",
"source": "systemd",
"state": "stopped",
"status": "unknown"
},
"NetworkManager-dispatcher.service": {
"name": "NetworkManager-dispatcher.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"NetworkManager-wait-online.service": {
"name": "NetworkManager-wait-online.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"NetworkManager.service": {
"name": "NetworkManager.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"auditd.service": {
"name": "auditd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"auth-rpcgss-module.service": {
"name": "auth-rpcgss-module.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"autovt@.service": {
"name": "autovt@.service",
"source": "systemd",
"state": "unknown",
"status": "enabled"
},
"certmonger.service": {
"name": "certmonger.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"chrony-dnssrv@.service": {
"name": "chrony-dnssrv@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"chrony-wait.service": {
"name": "chrony-wait.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"chronyd.service": {
"name": "chronyd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"cloud-config.service": {
"name": "cloud-config.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"cloud-final.service": {
"name": "cloud-final.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"cloud-init-hotplugd.service": {
"name": "cloud-init-hotplugd.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"cloud-init-local.service": {
"name": "cloud-init-local.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"cloud-init.service": {
"name": "cloud-init.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"cni-dhcp.service": {
"name": "cni-dhcp.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"console-getty.service": {
"name": "console-getty.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"container-getty@.service": {
"name": "container-getty@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"cpupower.service": {
"name": "cpupower.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"crond.service": {
"name": "crond.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"dbus-org.fedoraproject.FirewallD1.service": {
"name": "dbus-org.fedoraproject.FirewallD1.service",
"source": "systemd",
"state": "active",
"status": "enabled"
},
"dbus-org.freedesktop.hostname1.service": {
"name": "dbus-org.freedesktop.hostname1.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dbus-org.freedesktop.locale1.service": {
"name": "dbus-org.freedesktop.locale1.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dbus-org.freedesktop.login1.service": {
"name": "dbus-org.freedesktop.login1.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dbus-org.freedesktop.nm-dispatcher.service": {
"name": "dbus-org.freedesktop.nm-dispatcher.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"dbus-org.freedesktop.portable1.service": {
"name": "dbus-org.freedesktop.portable1.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dbus-org.freedesktop.timedate1.service": {
"name": "dbus-org.freedesktop.timedate1.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"dbus.service": {
"name": "dbus.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"debug-shell.service": {
"name": "debug-shell.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"dnf-makecache.service": {
"name": "dnf-makecache.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dnf-system-upgrade-cleanup.service": {
"name": "dnf-system-upgrade-cleanup.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dnf-system-upgrade.service": {
"name": "dnf-system-upgrade.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"dnsmasq.service": {
"name": "dnsmasq.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"dracut-cmdline.service": {
"name": "dracut-cmdline.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-initqueue.service": {
"name": "dracut-initqueue.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-mount.service": {
"name": "dracut-mount.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-mount.service": {
"name": "dracut-pre-mount.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-pivot.service": {
"name": "dracut-pre-pivot.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-trigger.service": {
"name": "dracut-pre-trigger.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-udev.service": {
"name": "dracut-pre-udev.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-shutdown-onfailure.service": {
"name": "dracut-shutdown-onfailure.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-shutdown.service": {
"name": "dracut-shutdown.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"ebtables.service": {
"name": "ebtables.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"emergency.service": {
"name": "emergency.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"firewalld.service": {
"name": "firewalld.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"fstrim.service": {
"name": "fstrim.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"getty@.service": {
"name": "getty@.service",
"source": "systemd",
"state": "unknown",
"status": "enabled"
},
"getty@tty1.service": {
"name": "getty@tty1.service",
"source": "systemd",
"state": "running",
"status": "unknown"
},
"grub-boot-indeterminate.service": {
"name": "grub-boot-indeterminate.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"gssproxy.service": {
"name": "gssproxy.service",
"source": "systemd",
"state": "running",
"status": "disabled"
},
"halt-local.service": {
"name": "halt-local.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"import-state.service": {
"name": "import-state.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"initrd-cleanup.service": {
"name": "initrd-cleanup.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"initrd-parse-etc.service": {
"name": "initrd-parse-etc.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"initrd-switch-root.service": {
"name": "initrd-switch-root.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"initrd-udevadm-cleanup-db.service": {
"name": "initrd-udevadm-cleanup-db.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"iprdump.service": {
"name": "iprdump.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"iprinit.service": {
"name": "iprinit.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"iprupdate.service": {
"name": "iprupdate.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"irqbalance.service": {
"name": "irqbalance.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"kdump.service": {
"name": "kdump.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"kmod-static-nodes.service": {
"name": "kmod-static-nodes.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"kvm_stat.service": {
"name": "kvm_stat.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"ldconfig.service": {
"name": "ldconfig.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"loadmodules.service": {
"name": "loadmodules.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"man-db-cache-update.service": {
"name": "man-db-cache-update.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"man-db-restart-cache-update.service": {
"name": "man-db-restart-cache-update.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"messagebus.service": {
"name": "messagebus.service",
"source": "systemd",
"state": "active",
"status": "static"
},
"microcode.service": {
"name": "microcode.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"nfs-blkmap.service": {
"name": "nfs-blkmap.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"nfs-convert.service": {
"name": "nfs-convert.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"nfs-idmapd.service": {
"name": "nfs-idmapd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nfs-mountd.service": {
"name": "nfs-mountd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nfs-server.service": {
"name": "nfs-server.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"nfs-utils.service": {
"name": "nfs-utils.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nfsdcld.service": {
"name": "nfsdcld.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nftables.service": {
"name": "nftables.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"nis-domainname.service": {
"name": "nis-domainname.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"oddjobd.service": {
"name": "oddjobd.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"plymouth-halt.service": {
"name": "plymouth-halt.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-kexec.service": {
"name": "plymouth-kexec.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-poweroff.service": {
"name": "plymouth-poweroff.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-quit-wait.service": {
"name": "plymouth-quit-wait.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-quit.service": {
"name": "plymouth-quit.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-read-write.service": {
"name": "plymouth-read-write.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-reboot.service": {
"name": "plymouth-reboot.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-start.service": {
"name": "plymouth-start.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-switch-root-initramfs.service": {
"name": "plymouth-switch-root-initramfs.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-switch-root.service": {
"name": "plymouth-switch-root.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"podman-auto-update.service": {
"name": "podman-auto-update.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"podman-clean-transient.service": {
"name": "podman-clean-transient.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"podman-kube@.service": {
"name": "podman-kube@.service",
"source": "systemd",
"state": "unknown",
"status": "disabled"
},
"podman-restart.service": {
"name": "podman-restart.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"podman.service": {
"name": "podman.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"polkit.service": {
"name": "polkit.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"qemu-guest-agent.service": {
"name": "qemu-guest-agent.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"quadlet-demo-mysql-volume.service": {
"name": "quadlet-demo-mysql-volume.service",
"source": "systemd",
"state": "stopped",
"status": "generated"
},
"quadlet-demo-mysql.service": {
"name": "quadlet-demo-mysql.service",
"source": "systemd",
"state": "running",
"status": "generated"
},
"quadlet-demo-network.service": {
"name": "quadlet-demo-network.service",
"source": "systemd",
"state": "stopped",
"status": "generated"
},
"quotaon.service": {
"name": "quotaon.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"rc-local.service": {
"name": "rc-local.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rdisc.service": {
"name": "rdisc.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"rescue.service": {
"name": "rescue.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"restraintd.service": {
"name": "restraintd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"rngd.service": {
"name": "rngd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"rpc-gssd.service": {
"name": "rpc-gssd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rpc-statd-notify.service": {
"name": "rpc-statd-notify.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rpc-statd.service": {
"name": "rpc-statd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rpcbind.service": {
"name": "rpcbind.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"rsyslog.service": {
"name": "rsyslog.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"selinux-autorelabel-mark.service": {
"name": "selinux-autorelabel-mark.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"selinux-autorelabel.service": {
"name": "selinux-autorelabel.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"serial-getty@.service": {
"name": "serial-getty@.service",
"source": "systemd",
"state": "unknown",
"status": "disabled"
},
"sshd-keygen@.service": {
"name": "sshd-keygen@.service",
"source": "systemd",
"state": "unknown",
"status": "disabled"
},
"sshd-keygen@ecdsa.service": {
"name": "sshd-keygen@ecdsa.service",
"source": "systemd",
"state": "stopped",
"status": "unknown"
},
"sshd-keygen@ed25519.service": {
"name": "sshd-keygen@ed25519.service",
"source": "systemd",
"state": "stopped",
"status": "unknown"
},
"sshd-keygen@rsa.service": {
"name": "sshd-keygen@rsa.service",
"source": "systemd",
"state": "stopped",
"status": "unknown"
},
"sshd.service": {
"name": "sshd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"sshd@.service": {
"name": "sshd@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"sssd-autofs.service": {
"name": "sssd-autofs.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-kcm.service": {
"name": "sssd-kcm.service",
"source": "systemd",
"state": "stopped",
"status": "indirect"
},
"sssd-nss.service": {
"name": "sssd-nss.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-pac.service": {
"name": "sssd-pac.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-pam.service": {
"name": "sssd-pam.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-ssh.service": {
"name": "sssd-ssh.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-sudo.service": {
"name": "sssd-sudo.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd.service": {
"name": "sssd.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"syslog.service": {
"name": "syslog.service",
"source": "systemd",
"state": "active",
"status": "enabled"
},
"system-update-cleanup.service": {
"name": "system-update-cleanup.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-ask-password-console.service": {
"name": "systemd-ask-password-console.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-ask-password-plymouth.service": {
"name": "systemd-ask-password-plymouth.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-ask-password-wall.service": {
"name": "systemd-ask-password-wall.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-backlight@.service": {
"name": "systemd-backlight@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-binfmt.service": {
"name": "systemd-binfmt.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-coredump@.service": {
"name": "systemd-coredump@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-exit.service": {
"name": "systemd-exit.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-firstboot.service": {
"name": "systemd-firstboot.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-fsck-root.service": {
"name": "systemd-fsck-root.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-fsck@.service": {
"name": "systemd-fsck@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-halt.service": {
"name": "systemd-halt.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-hibernate-resume@.service": {
"name": "systemd-hibernate-resume@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-hibernate.service": {
"name": "systemd-hibernate.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-hostnamed.service": {
"name": "systemd-hostnamed.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-hwdb-update.service": {
"name": "systemd-hwdb-update.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-hybrid-sleep.service": {
"name": "systemd-hybrid-sleep.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-initctl.service": {
"name": "systemd-initctl.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-journal-catalog-update.service": {
"name": "systemd-journal-catalog-update.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-journal-flush.service": {
"name": "systemd-journal-flush.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-journald.service": {
"name": "systemd-journald.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"systemd-kexec.service": {
"name": "systemd-kexec.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-localed.service": {
"name": "systemd-localed.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-logind.service": {
"name": "systemd-logind.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-machine-id-commit.service": {
"name": "systemd-machine-id-commit.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-modules-load.service": {
"name": "systemd-modules-load.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-portabled.service": {
"name": "systemd-portabled.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-poweroff.service": {
"name": "systemd-poweroff.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-pstore.service": {
"name": "systemd-pstore.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"systemd-quotacheck.service": {
"name": "systemd-quotacheck.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-random-seed.service": {
"name": "systemd-random-seed.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-reboot.service": {
"name": "systemd-reboot.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-remount-fs.service": {
"name": "systemd-remount-fs.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-resolved.service": {
"name": "systemd-resolved.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"systemd-rfkill.service": {
"name": "systemd-rfkill.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-suspend-then-hibernate.service": {
"name": "systemd-suspend-then-hibernate.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-suspend.service": {
"name": "systemd-suspend.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-sysctl.service": {
"name": "systemd-sysctl.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-sysusers.service": {
"name": "systemd-sysusers.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-timedated.service": {
"name": "systemd-timedated.service",
"source": "systemd",
"state": "inactive",
"status": "masked"
},
"systemd-tmpfiles-clean.service": {
"name": "systemd-tmpfiles-clean.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-tmpfiles-setup-dev.service": {
"name": "systemd-tmpfiles-setup-dev.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-tmpfiles-setup.service": {
"name": "systemd-tmpfiles-setup.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-udev-settle.service": {
"name": "systemd-udev-settle.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-udev-trigger.service": {
"name": "systemd-udev-trigger.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-udevd.service": {
"name": "systemd-udevd.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"systemd-update-done.service": {
"name": "systemd-update-done.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-update-utmp-runlevel.service": {
"name": "systemd-update-utmp-runlevel.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-update-utmp.service": {
"name": "systemd-update-utmp.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-user-sessions.service": {
"name": "systemd-user-sessions.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-vconsole-setup.service": {
"name": "systemd-vconsole-setup.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-volatile-root.service": {
"name": "systemd-volatile-root.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"tcsd.service": {
"name": "tcsd.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"teamd@.service": {
"name": "teamd@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"timedatex.service": {
"name": "timedatex.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"tuned.service": {
"name": "tuned.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"unbound-anchor.service": {
"name": "unbound-anchor.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"user-runtime-dir@.service": {
"name": "user-runtime-dir@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"user-runtime-dir@0.service": {
"name": "user-runtime-dir@0.service",
"source": "systemd",
"state": "stopped",
"status": "unknown"
},
"user@.service": {
"name": "user@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"user@0.service": {
"name": "user@0.service",
"source": "systemd",
"state": "running",
"status": "unknown"
}
}
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Create and update quadlets] ***********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:114
Saturday 24 August 2024 12:35:55 -0400 (0:00:01.994) 0:02:04.439 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14
Saturday 24 August 2024 12:35:55 -0400 (0:00:00.033) 0:02:04.473 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_quadlet_file_src": "",
"__podman_quadlet_spec": {},
"__podman_quadlet_str": "---\napiVersion: v1\nkind: PersistentVolumeClaim\nmetadata:\n name: wp-pv-claim\n labels:\n app: wordpress\nspec:\n accessModes:\n - ReadWriteOnce\n resources:\n requests:\n storage: 20Gi\n---\napiVersion: v1\nkind: Pod\nmetadata:\n name: quadlet-demo\nspec:\n containers:\n - name: wordpress\n image: quay.io/linux-system-roles/wordpress:4.8-apache\n env:\n - name: WORDPRESS_DB_HOST\n value: quadlet-demo-mysql\n - name: WORDPRESS_DB_PASSWORD\n valueFrom:\n secretKeyRef:\n name: mysql-root-password-kube\n key: password\n volumeMounts:\n - name: wordpress-persistent-storage\n mountPath: /var/www/html\n resources:\n requests:\n memory: \"64Mi\"\n cpu: \"250m\"\n limits:\n memory: \"128Mi\"\n cpu: \"500m\"\n - name: envoy\n image: quay.io/linux-system-roles/envoyproxy:v1.25.0\n volumeMounts:\n - name: config-volume\n mountPath: /etc/envoy\n - name: certificates\n mountPath: /etc/envoy-certificates\n env:\n - name: ENVOY_UID\n value: \"0\"\n resources:\n requests:\n memory: \"64Mi\"\n cpu: \"250m\"\n limits:\n memory: \"128Mi\"\n cpu: \"500m\"\n volumes:\n - name: config-volume\n configMap:\n name: envoy-proxy-config\n - name: certificates\n secret:\n secretName: envoy-certificates\n - name: wordpress-persistent-storage\n persistentVolumeClaim:\n claimName: wp-pv-claim\n - name: www # not used - for testing hostpath\n hostPath:\n path: /tmp/httpd3\n - name: create # not used - for testing hostpath\n hostPath:\n path: /tmp/httpd3-create\n",
"__podman_quadlet_template_src": "quadlet-demo.yml.j2"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25
Saturday 24 August 2024 12:35:55 -0400 (0:00:00.111) 0:02:04.584 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_continue_if_pull_fails": false,
"__podman_pull_image": true,
"__podman_state": "absent",
"__podman_systemd_unit_scope": "",
"__podman_user": "root"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] *****
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35
Saturday 24 August 2024 12:35:56 -0400 (0:00:00.042) 0:02:04.626 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48
Saturday 24 August 2024 12:35:56 -0400 (0:00:00.036) 0:02:04.663 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_quadlet_name": "quadlet-demo",
"__podman_quadlet_type": "yml",
"__podman_rootless": false
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Check user and group information] *****
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57
Saturday 24 August 2024 12:35:56 -0400 (0:00:00.046) 0:02:04.709 *******
included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for managed_node1
TASK [fedora.linux_system_roles.podman : Get user information] *****************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2
Saturday 24 August 2024 12:35:56 -0400 (0:00:00.064) 0:02:04.774 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Fail if user does not exist] **********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9
Saturday 24 August 2024 12:35:56 -0400 (0:00:00.041) 0:02:04.816 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set group for podman user] ************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16
Saturday 24 August 2024 12:35:56 -0400 (0:00:00.044) 0:02:04.860 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_group": "0"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Get group information] ****************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:28
Saturday 24 August 2024 12:35:56 -0400 (0:00:00.046) 0:02:04.907 *******
ok: [managed_node1] => {
"ansible_facts": {
"getent_group": {
"root": [
"x",
"0",
""
]
}
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set group name] ***********************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:35
Saturday 24 August 2024 12:35:56 -0400 (0:00:00.369) 0:02:05.276 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_group_name": "root"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : See if getsubids exists] **************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:39
Saturday 24 August 2024 12:35:56 -0400 (0:00:00.043) 0:02:05.320 *******
ok: [managed_node1] => {
"changed": false,
"stat": {
"atime": 1724516929.7168462,
"attr_flags": "",
"attributes": [],
"block_size": 4096,
"blocks": 32,
"charset": "binary",
"checksum": "bb5b46ffbafcaa8c4021f3c8b3cb8594f48ef34b",
"ctime": 1724516896.725572,
"dev": 51713,
"device_type": 0,
"executable": true,
"exists": true,
"gid": 0,
"gr_name": "root",
"inode": 6986657,
"isblk": false,
"ischr": false,
"isdir": false,
"isfifo": false,
"isgid": false,
"islnk": false,
"isreg": true,
"issock": false,
"isuid": false,
"mimetype": "application/x-sharedlib",
"mode": "0755",
"mtime": 1700557386.0,
"nlink": 1,
"path": "/usr/bin/getsubids",
"pw_name": "root",
"readable": true,
"rgrp": true,
"roth": true,
"rusr": true,
"size": 12640,
"uid": 0,
"version": "1255679238",
"wgrp": false,
"woth": false,
"writeable": true,
"wusr": true,
"xgrp": true,
"xoth": true,
"xusr": true
}
}
TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:50
Saturday 24 August 2024 12:35:57 -0400 (0:00:00.460) 0:02:05.781 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Check group with getsubids] ***********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:55
Saturday 24 August 2024 12:35:57 -0400 (0:00:00.055) 0:02:05.837 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ******
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:60
Saturday 24 August 2024 12:35:57 -0400 (0:00:00.039) 0:02:05.877 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Get subuid file] **********************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:74
Saturday 24 August 2024 12:35:57 -0400 (0:00:00.033) 0:02:05.910 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Get subgid file] **********************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:79
Saturday 24 August 2024 12:35:57 -0400 (0:00:00.034) 0:02:05.945 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ******
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:84
Saturday 24 August 2024 12:35:57 -0400 (0:00:00.033) 0:02:05.978 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ******
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:94
Saturday 24 August 2024 12:35:57 -0400 (0:00:00.032) 0:02:06.011 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] *****
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:101
Saturday 24 August 2024 12:35:57 -0400 (0:00:00.036) 0:02:06.047 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62
Saturday 24 August 2024 12:35:57 -0400 (0:00:00.033) 0:02:06.080 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_activate_systemd_unit": true,
"__podman_images_found": [],
"__podman_kube_yamls_raw": "",
"__podman_service_name": "",
"__podman_systemd_scope": "system",
"__podman_user_home_dir": "/root",
"__podman_xdg_runtime_dir": "/run/user/0"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73
Saturday 24 August 2024 12:35:57 -0400 (0:00:00.057) 0:02:06.138 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_quadlet_path": "/etc/containers/systemd"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Get kube yaml contents] ***************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77
Saturday 24 August 2024 12:35:57 -0400 (0:00:00.033) 0:02:06.171 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87
Saturday 24 August 2024 12:35:57 -0400 (0:00:00.032) 0:02:06.204 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_images": [],
"__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo.yml",
"__podman_volumes": []
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 6] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:103
Saturday 24 August 2024 12:35:57 -0400 (0:00:00.071) 0:02:06.275 *******
ok: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : Cleanup quadlets] *********************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:110
Saturday 24 August 2024 12:35:57 -0400 (0:00:00.037) 0:02:06.313 *******
included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml for managed_node1
TASK [fedora.linux_system_roles.podman : Stat XDG_RUNTIME_DIR] *****************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:4
Saturday 24 August 2024 12:35:57 -0400 (0:00:00.074) 0:02:06.388 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Stop and disable service] *************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:12
Saturday 24 August 2024 12:35:57 -0400 (0:00:00.072) 0:02:06.460 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : See if quadlet file exists] ***********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:33
Saturday 24 August 2024 12:35:57 -0400 (0:00:00.035) 0:02:06.495 *******
ok: [managed_node1] => {
"changed": false,
"stat": {
"atime": 1724517300.8158789,
"attr_flags": "",
"attributes": [],
"block_size": 4096,
"blocks": 8,
"charset": "us-ascii",
"checksum": "998dccde0483b1654327a46ddd89cbaa47650370",
"ctime": 1724517297.6538534,
"dev": 51713,
"device_type": 0,
"executable": false,
"exists": true,
"gid": 0,
"gr_name": "root",
"inode": 473956741,
"isblk": false,
"ischr": false,
"isdir": false,
"isfifo": false,
"isgid": false,
"islnk": false,
"isreg": true,
"issock": false,
"isuid": false,
"mimetype": "text/plain",
"mode": "0644",
"mtime": 1724517297.3598511,
"nlink": 1,
"path": "/etc/containers/systemd/quadlet-demo.yml",
"pw_name": "root",
"readable": true,
"rgrp": true,
"roth": true,
"rusr": true,
"size": 1605,
"uid": 0,
"version": "336872421",
"wgrp": false,
"woth": false,
"writeable": true,
"wusr": true,
"xgrp": false,
"xoth": false,
"xusr": false
}
}
TASK [fedora.linux_system_roles.podman : Parse quadlet file] *******************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:38
Saturday 24 August 2024 12:35:58 -0400 (0:00:00.365) 0:02:06.861 *******
included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml for managed_node1
TASK [fedora.linux_system_roles.podman : Slurp quadlet file] *******************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml:6
Saturday 24 August 2024 12:35:58 -0400 (0:00:00.062) 0:02:06.923 *******
ok: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : Parse quadlet file] *******************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml:12
Saturday 24 August 2024 12:35:58 -0400 (0:00:00.357) 0:02:07.281 *******
skipping: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : Parse quadlet yaml file] **************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml:44
Saturday 24 August 2024 12:35:58 -0400 (0:00:00.033) 0:02:07.314 *******
ok: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : Reset raw variable] *******************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml:52
Saturday 24 August 2024 12:35:58 -0400 (0:00:00.039) 0:02:07.354 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_quadlet_raw": null
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Remove quadlet file] ******************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:42
Saturday 24 August 2024 12:35:58 -0400 (0:00:00.032) 0:02:07.387 *******
changed: [managed_node1] => {
"changed": true,
"path": "/etc/containers/systemd/quadlet-demo.yml",
"state": "absent"
}
TASK [fedora.linux_system_roles.podman : Refresh systemd] **********************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:48
Saturday 24 August 2024 12:35:59 -0400 (0:00:00.358) 0:02:07.745 *******
ok: [managed_node1] => {
"changed": false,
"name": null,
"status": {}
}
TASK [fedora.linux_system_roles.podman : Remove managed resource] **************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:58
Saturday 24 August 2024 12:35:59 -0400 (0:00:00.597) 0:02:08.342 *******
skipping: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : Remove volumes] ***********************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:95
Saturday 24 August 2024 12:35:59 -0400 (0:00:00.071) 0:02:08.413 *******
skipping: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : Clear parsed podman variable] *********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:112
Saturday 24 August 2024 12:35:59 -0400 (0:00:00.047) 0:02:08.461 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_quadlet_parsed": null
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Prune images no longer in use] ********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:116
Saturday 24 August 2024 12:35:59 -0400 (0:00:00.031) 0:02:08.493 *******
changed: [managed_node1] => {
"changed": true,
"cmd": [
"podman",
"image",
"prune",
"--all",
"-f"
],
"delta": "0:00:00.030382",
"end": "2024-08-24 12:36:00.203613",
"rc": 0,
"start": "2024-08-24 12:36:00.173231"
}
TASK [fedora.linux_system_roles.podman : Manage linger] ************************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:127
Saturday 24 August 2024 12:36:00 -0400 (0:00:00.393) 0:02:08.886 *******
included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for managed_node1
TASK [fedora.linux_system_roles.podman : Enable linger if needed] **************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12
Saturday 24 August 2024 12:36:00 -0400 (0:00:00.060) 0:02:08.947 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18
Saturday 24 August 2024 12:36:00 -0400 (0:00:00.033) 0:02:08.980 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22
Saturday 24 August 2024 12:36:00 -0400 (0:00:00.034) 0:02:09.015 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : For testing and debugging - images] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:137
Saturday 24 August 2024 12:36:00 -0400 (0:00:00.033) 0:02:09.049 *******
ok: [managed_node1] => {
"changed": false,
"cmd": [
"podman",
"images",
"-n"
],
"delta": "0:00:00.033089",
"end": "2024-08-24 12:36:00.767456",
"rc": 0,
"start": "2024-08-24 12:36:00.734367"
}
STDOUT:
quay.io/linux-system-roles/mysql 5.6 dd3b2a5dcb48 2 years ago 308 MB
TASK [fedora.linux_system_roles.podman : For testing and debugging - volumes] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:146
Saturday 24 August 2024 12:36:00 -0400 (0:00:00.401) 0:02:09.450 *******
ok: [managed_node1] => {
"changed": false,
"cmd": [
"podman",
"volume",
"ls",
"-n"
],
"delta": "0:00:00.030605",
"end": "2024-08-24 12:36:01.162074",
"rc": 0,
"start": "2024-08-24 12:36:01.131469"
}
STDOUT:
local systemd-quadlet-demo-mysql
local wp-pv-claim
local envoy-proxy-config
local envoy-certificates
TASK [fedora.linux_system_roles.podman : For testing and debugging - containers] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:155
Saturday 24 August 2024 12:36:01 -0400 (0:00:00.394) 0:02:09.845 *******
ok: [managed_node1] => {
"changed": false,
"cmd": [
"podman",
"ps",
"--noheading"
],
"delta": "0:00:00.035869",
"end": "2024-08-24 12:36:01.565909",
"rc": 0,
"start": "2024-08-24 12:36:01.530040"
}
STDOUT:
68250d5fc021 quay.io/linux-system-roles/mysql:5.6 mysqld About a minute ago Up About a minute (healthy) quadlet-demo-mysql
TASK [fedora.linux_system_roles.podman : For testing and debugging - networks] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:164
Saturday 24 August 2024 12:36:01 -0400 (0:00:00.403) 0:02:10.248 *******
ok: [managed_node1] => {
"changed": false,
"cmd": [
"podman",
"network",
"ls",
"-n",
"-q"
],
"delta": "0:00:00.056603",
"end": "2024-08-24 12:36:01.990366",
"rc": 0,
"start": "2024-08-24 12:36:01.933763"
}
STDOUT:
podman
podman-default-kube-network
systemd-quadlet-demo
TASK [fedora.linux_system_roles.podman : For testing and debugging - secrets] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:173
Saturday 24 August 2024 12:36:02 -0400 (0:00:00.424) 0:02:10.673 *******
ok: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : For testing and debugging - services] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:183
Saturday 24 August 2024 12:36:02 -0400 (0:00:00.393) 0:02:11.066 *******
ok: [managed_node1] => {
"ansible_facts": {
"services": {
"68250d5fc0218bc6db3ab2...5ebe15c37a833d3345ad2e2df2c1916dd6cc3.service": {
"name": "68250d5fc0218bc6db3ab2...5ebe15c37a833d3345ad2e2df2c1916dd6cc3.service",
"source": "systemd",
"state": "inactive",
"status": "transient"
},
"68250d5fc0218bc6db3ab2ece605ebe15c37a833d3345ad2e2df2c1916dd6cc3.service": {
"name": "68250d5fc0218bc6db3ab2ece605ebe15c37a833d3345ad2e2df2c1916dd6cc3.service",
"source": "systemd",
"state": "stopped",
"status": "unknown"
},
"NetworkManager-dispatcher.service": {
"name": "NetworkManager-dispatcher.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"NetworkManager-wait-online.service": {
"name": "NetworkManager-wait-online.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"NetworkManager.service": {
"name": "NetworkManager.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"auditd.service": {
"name": "auditd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"auth-rpcgss-module.service": {
"name": "auth-rpcgss-module.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"autovt@.service": {
"name": "autovt@.service",
"source": "systemd",
"state": "unknown",
"status": "enabled"
},
"certmonger.service": {
"name": "certmonger.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"chrony-dnssrv@.service": {
"name": "chrony-dnssrv@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"chrony-wait.service": {
"name": "chrony-wait.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"chronyd.service": {
"name": "chronyd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"cloud-config.service": {
"name": "cloud-config.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"cloud-final.service": {
"name": "cloud-final.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"cloud-init-hotplugd.service": {
"name": "cloud-init-hotplugd.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"cloud-init-local.service": {
"name": "cloud-init-local.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"cloud-init.service": {
"name": "cloud-init.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"cni-dhcp.service": {
"name": "cni-dhcp.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"console-getty.service": {
"name": "console-getty.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"container-getty@.service": {
"name": "container-getty@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"cpupower.service": {
"name": "cpupower.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"crond.service": {
"name": "crond.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"dbus-org.fedoraproject.FirewallD1.service": {
"name": "dbus-org.fedoraproject.FirewallD1.service",
"source": "systemd",
"state": "active",
"status": "enabled"
},
"dbus-org.freedesktop.hostname1.service": {
"name": "dbus-org.freedesktop.hostname1.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dbus-org.freedesktop.locale1.service": {
"name": "dbus-org.freedesktop.locale1.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dbus-org.freedesktop.login1.service": {
"name": "dbus-org.freedesktop.login1.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dbus-org.freedesktop.nm-dispatcher.service": {
"name": "dbus-org.freedesktop.nm-dispatcher.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"dbus-org.freedesktop.portable1.service": {
"name": "dbus-org.freedesktop.portable1.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dbus-org.freedesktop.timedate1.service": {
"name": "dbus-org.freedesktop.timedate1.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"dbus.service": {
"name": "dbus.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"debug-shell.service": {
"name": "debug-shell.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"dnf-makecache.service": {
"name": "dnf-makecache.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dnf-system-upgrade-cleanup.service": {
"name": "dnf-system-upgrade-cleanup.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dnf-system-upgrade.service": {
"name": "dnf-system-upgrade.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"dnsmasq.service": {
"name": "dnsmasq.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"dracut-cmdline.service": {
"name": "dracut-cmdline.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-initqueue.service": {
"name": "dracut-initqueue.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-mount.service": {
"name": "dracut-mount.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-mount.service": {
"name": "dracut-pre-mount.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-pivot.service": {
"name": "dracut-pre-pivot.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-trigger.service": {
"name": "dracut-pre-trigger.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-udev.service": {
"name": "dracut-pre-udev.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-shutdown-onfailure.service": {
"name": "dracut-shutdown-onfailure.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-shutdown.service": {
"name": "dracut-shutdown.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"ebtables.service": {
"name": "ebtables.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"emergency.service": {
"name": "emergency.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"firewalld.service": {
"name": "firewalld.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"fstrim.service": {
"name": "fstrim.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"getty@.service": {
"name": "getty@.service",
"source": "systemd",
"state": "unknown",
"status": "enabled"
},
"getty@tty1.service": {
"name": "getty@tty1.service",
"source": "systemd",
"state": "running",
"status": "unknown"
},
"grub-boot-indeterminate.service": {
"name": "grub-boot-indeterminate.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"gssproxy.service": {
"name": "gssproxy.service",
"source": "systemd",
"state": "running",
"status": "disabled"
},
"halt-local.service": {
"name": "halt-local.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"import-state.service": {
"name": "import-state.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"initrd-cleanup.service": {
"name": "initrd-cleanup.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"initrd-parse-etc.service": {
"name": "initrd-parse-etc.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"initrd-switch-root.service": {
"name": "initrd-switch-root.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"initrd-udevadm-cleanup-db.service": {
"name": "initrd-udevadm-cleanup-db.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"iprdump.service": {
"name": "iprdump.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"iprinit.service": {
"name": "iprinit.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"iprupdate.service": {
"name": "iprupdate.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"irqbalance.service": {
"name": "irqbalance.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"kdump.service": {
"name": "kdump.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"kmod-static-nodes.service": {
"name": "kmod-static-nodes.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"kvm_stat.service": {
"name": "kvm_stat.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"ldconfig.service": {
"name": "ldconfig.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"loadmodules.service": {
"name": "loadmodules.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"man-db-cache-update.service": {
"name": "man-db-cache-update.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"man-db-restart-cache-update.service": {
"name": "man-db-restart-cache-update.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"messagebus.service": {
"name": "messagebus.service",
"source": "systemd",
"state": "active",
"status": "static"
},
"microcode.service": {
"name": "microcode.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"nfs-blkmap.service": {
"name": "nfs-blkmap.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"nfs-convert.service": {
"name": "nfs-convert.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"nfs-idmapd.service": {
"name": "nfs-idmapd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nfs-mountd.service": {
"name": "nfs-mountd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nfs-server.service": {
"name": "nfs-server.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"nfs-utils.service": {
"name": "nfs-utils.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nfsdcld.service": {
"name": "nfsdcld.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nftables.service": {
"name": "nftables.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"nis-domainname.service": {
"name": "nis-domainname.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"oddjobd.service": {
"name": "oddjobd.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"plymouth-halt.service": {
"name": "plymouth-halt.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-kexec.service": {
"name": "plymouth-kexec.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-poweroff.service": {
"name": "plymouth-poweroff.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-quit-wait.service": {
"name": "plymouth-quit-wait.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-quit.service": {
"name": "plymouth-quit.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-read-write.service": {
"name": "plymouth-read-write.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-reboot.service": {
"name": "plymouth-reboot.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-start.service": {
"name": "plymouth-start.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-switch-root-initramfs.service": {
"name": "plymouth-switch-root-initramfs.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-switch-root.service": {
"name": "plymouth-switch-root.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"podman-auto-update.service": {
"name": "podman-auto-update.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"podman-clean-transient.service": {
"name": "podman-clean-transient.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"podman-kube@.service": {
"name": "podman-kube@.service",
"source": "systemd",
"state": "unknown",
"status": "disabled"
},
"podman-restart.service": {
"name": "podman-restart.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"podman.service": {
"name": "podman.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"polkit.service": {
"name": "polkit.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"qemu-guest-agent.service": {
"name": "qemu-guest-agent.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"quadlet-demo-mysql-volume.service": {
"name": "quadlet-demo-mysql-volume.service",
"source": "systemd",
"state": "stopped",
"status": "generated"
},
"quadlet-demo-mysql.service": {
"name": "quadlet-demo-mysql.service",
"source": "systemd",
"state": "running",
"status": "generated"
},
"quadlet-demo-network.service": {
"name": "quadlet-demo-network.service",
"source": "systemd",
"state": "stopped",
"status": "generated"
},
"quotaon.service": {
"name": "quotaon.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"rc-local.service": {
"name": "rc-local.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rdisc.service": {
"name": "rdisc.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"rescue.service": {
"name": "rescue.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"restraintd.service": {
"name": "restraintd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"rngd.service": {
"name": "rngd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"rpc-gssd.service": {
"name": "rpc-gssd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rpc-statd-notify.service": {
"name": "rpc-statd-notify.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rpc-statd.service": {
"name": "rpc-statd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rpcbind.service": {
"name": "rpcbind.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"rsyslog.service": {
"name": "rsyslog.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"selinux-autorelabel-mark.service": {
"name": "selinux-autorelabel-mark.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"selinux-autorelabel.service": {
"name": "selinux-autorelabel.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"serial-getty@.service": {
"name": "serial-getty@.service",
"source": "systemd",
"state": "unknown",
"status": "disabled"
},
"sshd-keygen@.service": {
"name": "sshd-keygen@.service",
"source": "systemd",
"state": "unknown",
"status": "disabled"
},
"sshd-keygen@ecdsa.service": {
"name": "sshd-keygen@ecdsa.service",
"source": "systemd",
"state": "stopped",
"status": "unknown"
},
"sshd-keygen@ed25519.service": {
"name": "sshd-keygen@ed25519.service",
"source": "systemd",
"state": "stopped",
"status": "unknown"
},
"sshd-keygen@rsa.service": {
"name": "sshd-keygen@rsa.service",
"source": "systemd",
"state": "stopped",
"status": "unknown"
},
"sshd.service": {
"name": "sshd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"sshd@.service": {
"name": "sshd@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"sssd-autofs.service": {
"name": "sssd-autofs.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-kcm.service": {
"name": "sssd-kcm.service",
"source": "systemd",
"state": "stopped",
"status": "indirect"
},
"sssd-nss.service": {
"name": "sssd-nss.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-pac.service": {
"name": "sssd-pac.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-pam.service": {
"name": "sssd-pam.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-ssh.service": {
"name": "sssd-ssh.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-sudo.service": {
"name": "sssd-sudo.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd.service": {
"name": "sssd.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"syslog.service": {
"name": "syslog.service",
"source": "systemd",
"state": "active",
"status": "enabled"
},
"system-update-cleanup.service": {
"name": "system-update-cleanup.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-ask-password-console.service": {
"name": "systemd-ask-password-console.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-ask-password-plymouth.service": {
"name": "systemd-ask-password-plymouth.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-ask-password-wall.service": {
"name": "systemd-ask-password-wall.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-backlight@.service": {
"name": "systemd-backlight@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-binfmt.service": {
"name": "systemd-binfmt.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-coredump@.service": {
"name": "systemd-coredump@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-exit.service": {
"name": "systemd-exit.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-firstboot.service": {
"name": "systemd-firstboot.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-fsck-root.service": {
"name": "systemd-fsck-root.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-fsck@.service": {
"name": "systemd-fsck@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-halt.service": {
"name": "systemd-halt.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-hibernate-resume@.service": {
"name": "systemd-hibernate-resume@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-hibernate.service": {
"name": "systemd-hibernate.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-hostnamed.service": {
"name": "systemd-hostnamed.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-hwdb-update.service": {
"name": "systemd-hwdb-update.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-hybrid-sleep.service": {
"name": "systemd-hybrid-sleep.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-initctl.service": {
"name": "systemd-initctl.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-journal-catalog-update.service": {
"name": "systemd-journal-catalog-update.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-journal-flush.service": {
"name": "systemd-journal-flush.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-journald.service": {
"name": "systemd-journald.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"systemd-kexec.service": {
"name": "systemd-kexec.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-localed.service": {
"name": "systemd-localed.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-logind.service": {
"name": "systemd-logind.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-machine-id-commit.service": {
"name": "systemd-machine-id-commit.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-modules-load.service": {
"name": "systemd-modules-load.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-portabled.service": {
"name": "systemd-portabled.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-poweroff.service": {
"name": "systemd-poweroff.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-pstore.service": {
"name": "systemd-pstore.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"systemd-quotacheck.service": {
"name": "systemd-quotacheck.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-random-seed.service": {
"name": "systemd-random-seed.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-reboot.service": {
"name": "systemd-reboot.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-remount-fs.service": {
"name": "systemd-remount-fs.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-resolved.service": {
"name": "systemd-resolved.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"systemd-rfkill.service": {
"name": "systemd-rfkill.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-suspend-then-hibernate.service": {
"name": "systemd-suspend-then-hibernate.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-suspend.service": {
"name": "systemd-suspend.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-sysctl.service": {
"name": "systemd-sysctl.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-sysusers.service": {
"name": "systemd-sysusers.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-timedated.service": {
"name": "systemd-timedated.service",
"source": "systemd",
"state": "inactive",
"status": "masked"
},
"systemd-tmpfiles-clean.service": {
"name": "systemd-tmpfiles-clean.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-tmpfiles-setup-dev.service": {
"name": "systemd-tmpfiles-setup-dev.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-tmpfiles-setup.service": {
"name": "systemd-tmpfiles-setup.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-udev-settle.service": {
"name": "systemd-udev-settle.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-udev-trigger.service": {
"name": "systemd-udev-trigger.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-udevd.service": {
"name": "systemd-udevd.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"systemd-update-done.service": {
"name": "systemd-update-done.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-update-utmp-runlevel.service": {
"name": "systemd-update-utmp-runlevel.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-update-utmp.service": {
"name": "systemd-update-utmp.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-user-sessions.service": {
"name": "systemd-user-sessions.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-vconsole-setup.service": {
"name": "systemd-vconsole-setup.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-volatile-root.service": {
"name": "systemd-volatile-root.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"tcsd.service": {
"name": "tcsd.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"teamd@.service": {
"name": "teamd@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"timedatex.service": {
"name": "timedatex.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"tuned.service": {
"name": "tuned.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"unbound-anchor.service": {
"name": "unbound-anchor.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"user-runtime-dir@.service": {
"name": "user-runtime-dir@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"user-runtime-dir@0.service": {
"name": "user-runtime-dir@0.service",
"source": "systemd",
"state": "stopped",
"status": "unknown"
},
"user@.service": {
"name": "user@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"user@0.service": {
"name": "user@0.service",
"source": "systemd",
"state": "running",
"status": "unknown"
}
}
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Create and update quadlets] ***********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:114
Saturday 24 August 2024 12:36:04 -0400 (0:00:01.589) 0:02:12.656 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14
Saturday 24 August 2024 12:36:04 -0400 (0:00:00.034) 0:02:12.691 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_quadlet_file_src": "envoy-proxy-configmap.yml",
"__podman_quadlet_spec": {},
"__podman_quadlet_str": "---\napiVersion: v1\nkind: ConfigMap\nmetadata:\n name: envoy-proxy-config\ndata:\n envoy.yaml: |\n admin:\n address:\n socket_address:\n address: 0.0.0.0\n port_value: 9901\n\n static_resources:\n listeners:\n - name: listener_0\n address:\n socket_address:\n address: 0.0.0.0\n port_value: 8080\n filter_chains:\n - filters:\n - name: envoy.filters.network.http_connection_manager\n typed_config:\n \"@type\": type.googleapis.com/envoy.extensions.filters.network.http_connection_manager.v3.HttpConnectionManager\n stat_prefix: ingress_http\n codec_type: AUTO\n route_config:\n name: local_route\n virtual_hosts:\n - name: local_service\n domains: [\"*\"]\n routes:\n - match:\n prefix: \"/\"\n route:\n cluster: backend\n http_filters:\n - name: envoy.filters.http.router\n typed_config:\n \"@type\": type.googleapis.com/envoy.extensions.filters.http.router.v3.Router\n transport_socket:\n name: envoy.transport_sockets.tls\n typed_config:\n \"@type\": type.googleapis.com/envoy.extensions.transport_sockets.tls.v3.DownstreamTlsContext\n common_tls_context:\n tls_certificates:\n - certificate_chain:\n filename: /etc/envoy-certificates/certificate.pem\n private_key:\n filename: /etc/envoy-certificates/certificate.key\n clusters:\n - name: backend\n connect_timeout: 5s\n type: STATIC\n dns_refresh_rate: 1800s\n lb_policy: ROUND_ROBIN\n load_assignment:\n cluster_name: backend\n endpoints:\n - lb_endpoints:\n - endpoint:\n address:\n socket_address:\n address: 127.0.0.1\n port_value: 80",
"__podman_quadlet_template_src": ""
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25
Saturday 24 August 2024 12:36:04 -0400 (0:00:00.087) 0:02:12.779 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_continue_if_pull_fails": false,
"__podman_pull_image": true,
"__podman_state": "absent",
"__podman_systemd_unit_scope": "",
"__podman_user": "root"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] *****
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35
Saturday 24 August 2024 12:36:04 -0400 (0:00:00.041) 0:02:12.821 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48
Saturday 24 August 2024 12:36:04 -0400 (0:00:00.037) 0:02:12.858 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_quadlet_name": "envoy-proxy-configmap",
"__podman_quadlet_type": "yml",
"__podman_rootless": false
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Check user and group information] *****
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57
Saturday 24 August 2024 12:36:04 -0400 (0:00:00.048) 0:02:12.906 *******
included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for managed_node1
TASK [fedora.linux_system_roles.podman : Get user information] *****************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2
Saturday 24 August 2024 12:36:04 -0400 (0:00:00.063) 0:02:12.969 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Fail if user does not exist] **********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9
Saturday 24 August 2024 12:36:04 -0400 (0:00:00.041) 0:02:13.011 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set group for podman user] ************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16
Saturday 24 August 2024 12:36:04 -0400 (0:00:00.041) 0:02:13.052 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_group": "0"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Get group information] ****************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:28
Saturday 24 August 2024 12:36:04 -0400 (0:00:00.048) 0:02:13.101 *******
ok: [managed_node1] => {
"ansible_facts": {
"getent_group": {
"root": [
"x",
"0",
""
]
}
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set group name] ***********************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:35
Saturday 24 August 2024 12:36:04 -0400 (0:00:00.370) 0:02:13.472 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_group_name": "root"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : See if getsubids exists] **************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:39
Saturday 24 August 2024 12:36:04 -0400 (0:00:00.042) 0:02:13.514 *******
ok: [managed_node1] => {
"changed": false,
"stat": {
"atime": 1724516929.7168462,
"attr_flags": "",
"attributes": [],
"block_size": 4096,
"blocks": 32,
"charset": "binary",
"checksum": "bb5b46ffbafcaa8c4021f3c8b3cb8594f48ef34b",
"ctime": 1724516896.725572,
"dev": 51713,
"device_type": 0,
"executable": true,
"exists": true,
"gid": 0,
"gr_name": "root",
"inode": 6986657,
"isblk": false,
"ischr": false,
"isdir": false,
"isfifo": false,
"isgid": false,
"islnk": false,
"isreg": true,
"issock": false,
"isuid": false,
"mimetype": "application/x-sharedlib",
"mode": "0755",
"mtime": 1700557386.0,
"nlink": 1,
"path": "/usr/bin/getsubids",
"pw_name": "root",
"readable": true,
"rgrp": true,
"roth": true,
"rusr": true,
"size": 12640,
"uid": 0,
"version": "1255679238",
"wgrp": false,
"woth": false,
"writeable": true,
"wusr": true,
"xgrp": true,
"xoth": true,
"xusr": true
}
}
TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:50
Saturday 24 August 2024 12:36:05 -0400 (0:00:00.357) 0:02:13.871 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Check group with getsubids] ***********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:55
Saturday 24 August 2024 12:36:05 -0400 (0:00:00.033) 0:02:13.905 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ******
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:60
Saturday 24 August 2024 12:36:05 -0400 (0:00:00.032) 0:02:13.937 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Get subuid file] **********************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:74
Saturday 24 August 2024 12:36:05 -0400 (0:00:00.071) 0:02:14.008 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Get subgid file] **********************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:79
Saturday 24 August 2024 12:36:05 -0400 (0:00:00.034) 0:02:14.042 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ******
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:84
Saturday 24 August 2024 12:36:05 -0400 (0:00:00.033) 0:02:14.076 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ******
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:94
Saturday 24 August 2024 12:36:05 -0400 (0:00:00.039) 0:02:14.115 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] *****
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:101
Saturday 24 August 2024 12:36:05 -0400 (0:00:00.034) 0:02:14.149 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62
Saturday 24 August 2024 12:36:05 -0400 (0:00:00.032) 0:02:14.182 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_activate_systemd_unit": true,
"__podman_images_found": [],
"__podman_kube_yamls_raw": "",
"__podman_service_name": "",
"__podman_systemd_scope": "system",
"__podman_user_home_dir": "/root",
"__podman_xdg_runtime_dir": "/run/user/0"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73
Saturday 24 August 2024 12:36:05 -0400 (0:00:00.060) 0:02:14.243 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_quadlet_path": "/etc/containers/systemd"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Get kube yaml contents] ***************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77
Saturday 24 August 2024 12:36:05 -0400 (0:00:00.036) 0:02:14.279 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87
Saturday 24 August 2024 12:36:05 -0400 (0:00:00.032) 0:02:14.311 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_images": [],
"__podman_quadlet_file": "/etc/containers/systemd/envoy-proxy-configmap.yml",
"__podman_volumes": []
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 6] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:103
Saturday 24 August 2024 12:36:05 -0400 (0:00:00.074) 0:02:14.386 *******
ok: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : Cleanup quadlets] *********************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:110
Saturday 24 August 2024 12:36:05 -0400 (0:00:00.039) 0:02:14.425 *******
included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml for managed_node1
TASK [fedora.linux_system_roles.podman : Stat XDG_RUNTIME_DIR] *****************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:4
Saturday 24 August 2024 12:36:05 -0400 (0:00:00.076) 0:02:14.501 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Stop and disable service] *************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:12
Saturday 24 August 2024 12:36:05 -0400 (0:00:00.033) 0:02:14.535 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : See if quadlet file exists] ***********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:33
Saturday 24 August 2024 12:36:05 -0400 (0:00:00.035) 0:02:14.570 *******
ok: [managed_node1] => {
"changed": false,
"stat": {
"atime": 1724517322.898056,
"attr_flags": "",
"attributes": [],
"block_size": 4096,
"blocks": 8,
"charset": "us-ascii",
"checksum": "d681c7d56f912150d041873e880818b22a90c188",
"ctime": 1724517292.8068147,
"dev": 51713,
"device_type": 0,
"executable": false,
"exists": true,
"gid": 0,
"gr_name": "root",
"inode": 448790699,
"isblk": false,
"ischr": false,
"isdir": false,
"isfifo": false,
"isgid": false,
"islnk": false,
"isreg": true,
"issock": false,
"isuid": false,
"mimetype": "text/plain",
"mode": "0644",
"mtime": 1724517292.5298123,
"nlink": 1,
"path": "/etc/containers/systemd/envoy-proxy-configmap.yml",
"pw_name": "root",
"readable": true,
"rgrp": true,
"roth": true,
"rusr": true,
"size": 2102,
"uid": 0,
"version": "2763400557",
"wgrp": false,
"woth": false,
"writeable": true,
"wusr": true,
"xgrp": false,
"xoth": false,
"xusr": false
}
}
TASK [fedora.linux_system_roles.podman : Parse quadlet file] *******************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:38
Saturday 24 August 2024 12:36:06 -0400 (0:00:00.384) 0:02:14.955 *******
included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml for managed_node1
TASK [fedora.linux_system_roles.podman : Slurp quadlet file] *******************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml:6
Saturday 24 August 2024 12:36:06 -0400 (0:00:00.061) 0:02:15.016 *******
ok: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : Parse quadlet file] *******************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml:12
Saturday 24 August 2024 12:36:06 -0400 (0:00:00.348) 0:02:15.365 *******
skipping: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : Parse quadlet yaml file] **************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml:44
Saturday 24 August 2024 12:36:06 -0400 (0:00:00.033) 0:02:15.399 *******
ok: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : Reset raw variable] *******************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml:52
Saturday 24 August 2024 12:36:06 -0400 (0:00:00.038) 0:02:15.438 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_quadlet_raw": null
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Remove quadlet file] ******************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:42
Saturday 24 August 2024 12:36:06 -0400 (0:00:00.033) 0:02:15.471 *******
changed: [managed_node1] => {
"changed": true,
"path": "/etc/containers/systemd/envoy-proxy-configmap.yml",
"state": "absent"
}
TASK [fedora.linux_system_roles.podman : Refresh systemd] **********************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:48
Saturday 24 August 2024 12:36:07 -0400 (0:00:00.360) 0:02:15.832 *******
ok: [managed_node1] => {
"changed": false,
"name": null,
"status": {}
}
TASK [fedora.linux_system_roles.podman : Remove managed resource] **************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:58
Saturday 24 August 2024 12:36:07 -0400 (0:00:00.597) 0:02:16.429 *******
skipping: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : Remove volumes] ***********************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:95
Saturday 24 August 2024 12:36:07 -0400 (0:00:00.036) 0:02:16.465 *******
skipping: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : Clear parsed podman variable] *********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:112
Saturday 24 August 2024 12:36:07 -0400 (0:00:00.046) 0:02:16.511 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_quadlet_parsed": null
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Prune images no longer in use] ********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:116
Saturday 24 August 2024 12:36:07 -0400 (0:00:00.069) 0:02:16.581 *******
changed: [managed_node1] => {
"changed": true,
"cmd": [
"podman",
"image",
"prune",
"--all",
"-f"
],
"delta": "0:00:00.030343",
"end": "2024-08-24 12:36:08.291158",
"rc": 0,
"start": "2024-08-24 12:36:08.260815"
}
TASK [fedora.linux_system_roles.podman : Manage linger] ************************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:127
Saturday 24 August 2024 12:36:08 -0400 (0:00:00.392) 0:02:16.973 *******
included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for managed_node1
TASK [fedora.linux_system_roles.podman : Enable linger if needed] **************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12
Saturday 24 August 2024 12:36:08 -0400 (0:00:00.058) 0:02:17.032 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18
Saturday 24 August 2024 12:36:08 -0400 (0:00:00.033) 0:02:17.066 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22
Saturday 24 August 2024 12:36:08 -0400 (0:00:00.032) 0:02:17.098 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : For testing and debugging - images] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:137
Saturday 24 August 2024 12:36:08 -0400 (0:00:00.038) 0:02:17.136 *******
ok: [managed_node1] => {
"changed": false,
"cmd": [
"podman",
"images",
"-n"
],
"delta": "0:00:00.031808",
"end": "2024-08-24 12:36:08.852358",
"rc": 0,
"start": "2024-08-24 12:36:08.820550"
}
STDOUT:
quay.io/linux-system-roles/mysql 5.6 dd3b2a5dcb48 2 years ago 308 MB
TASK [fedora.linux_system_roles.podman : For testing and debugging - volumes] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:146
Saturday 24 August 2024 12:36:08 -0400 (0:00:00.399) 0:02:17.536 *******
ok: [managed_node1] => {
"changed": false,
"cmd": [
"podman",
"volume",
"ls",
"-n"
],
"delta": "0:00:00.029144",
"end": "2024-08-24 12:36:09.242928",
"rc": 0,
"start": "2024-08-24 12:36:09.213784"
}
STDOUT:
local systemd-quadlet-demo-mysql
local wp-pv-claim
local envoy-proxy-config
local envoy-certificates
TASK [fedora.linux_system_roles.podman : For testing and debugging - containers] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:155
Saturday 24 August 2024 12:36:09 -0400 (0:00:00.390) 0:02:17.926 *******
ok: [managed_node1] => {
"changed": false,
"cmd": [
"podman",
"ps",
"--noheading"
],
"delta": "0:00:00.035472",
"end": "2024-08-24 12:36:09.639561",
"rc": 0,
"start": "2024-08-24 12:36:09.604089"
}
STDOUT:
68250d5fc021 quay.io/linux-system-roles/mysql:5.6 mysqld About a minute ago Up About a minute (healthy) quadlet-demo-mysql
TASK [fedora.linux_system_roles.podman : For testing and debugging - networks] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:164
Saturday 24 August 2024 12:36:09 -0400 (0:00:00.396) 0:02:18.323 *******
ok: [managed_node1] => {
"changed": false,
"cmd": [
"podman",
"network",
"ls",
"-n",
"-q"
],
"delta": "0:00:00.059485",
"end": "2024-08-24 12:36:10.065687",
"rc": 0,
"start": "2024-08-24 12:36:10.006202"
}
STDOUT:
podman
podman-default-kube-network
systemd-quadlet-demo
TASK [fedora.linux_system_roles.podman : For testing and debugging - secrets] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:173
Saturday 24 August 2024 12:36:10 -0400 (0:00:00.426) 0:02:18.749 *******
ok: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : For testing and debugging - services] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:183
Saturday 24 August 2024 12:36:10 -0400 (0:00:00.396) 0:02:19.145 *******
ok: [managed_node1] => {
"ansible_facts": {
"services": {
"68250d5fc0218bc6db3ab2...5ebe15c37a833d3345ad2e2df2c1916dd6cc3.service": {
"name": "68250d5fc0218bc6db3ab2...5ebe15c37a833d3345ad2e2df2c1916dd6cc3.service",
"source": "systemd",
"state": "inactive",
"status": "transient"
},
"68250d5fc0218bc6db3ab2ece605ebe15c37a833d3345ad2e2df2c1916dd6cc3.service": {
"name": "68250d5fc0218bc6db3ab2ece605ebe15c37a833d3345ad2e2df2c1916dd6cc3.service",
"source": "systemd",
"state": "stopped",
"status": "unknown"
},
"NetworkManager-dispatcher.service": {
"name": "NetworkManager-dispatcher.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"NetworkManager-wait-online.service": {
"name": "NetworkManager-wait-online.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"NetworkManager.service": {
"name": "NetworkManager.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"auditd.service": {
"name": "auditd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"auth-rpcgss-module.service": {
"name": "auth-rpcgss-module.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"autovt@.service": {
"name": "autovt@.service",
"source": "systemd",
"state": "unknown",
"status": "enabled"
},
"certmonger.service": {
"name": "certmonger.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"chrony-dnssrv@.service": {
"name": "chrony-dnssrv@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"chrony-wait.service": {
"name": "chrony-wait.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"chronyd.service": {
"name": "chronyd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"cloud-config.service": {
"name": "cloud-config.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"cloud-final.service": {
"name": "cloud-final.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"cloud-init-hotplugd.service": {
"name": "cloud-init-hotplugd.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"cloud-init-local.service": {
"name": "cloud-init-local.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"cloud-init.service": {
"name": "cloud-init.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"cni-dhcp.service": {
"name": "cni-dhcp.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"console-getty.service": {
"name": "console-getty.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"container-getty@.service": {
"name": "container-getty@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"cpupower.service": {
"name": "cpupower.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"crond.service": {
"name": "crond.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"dbus-org.fedoraproject.FirewallD1.service": {
"name": "dbus-org.fedoraproject.FirewallD1.service",
"source": "systemd",
"state": "active",
"status": "enabled"
},
"dbus-org.freedesktop.hostname1.service": {
"name": "dbus-org.freedesktop.hostname1.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dbus-org.freedesktop.locale1.service": {
"name": "dbus-org.freedesktop.locale1.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dbus-org.freedesktop.login1.service": {
"name": "dbus-org.freedesktop.login1.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dbus-org.freedesktop.nm-dispatcher.service": {
"name": "dbus-org.freedesktop.nm-dispatcher.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"dbus-org.freedesktop.portable1.service": {
"name": "dbus-org.freedesktop.portable1.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dbus-org.freedesktop.timedate1.service": {
"name": "dbus-org.freedesktop.timedate1.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"dbus.service": {
"name": "dbus.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"debug-shell.service": {
"name": "debug-shell.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"dnf-makecache.service": {
"name": "dnf-makecache.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dnf-system-upgrade-cleanup.service": {
"name": "dnf-system-upgrade-cleanup.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dnf-system-upgrade.service": {
"name": "dnf-system-upgrade.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"dnsmasq.service": {
"name": "dnsmasq.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"dracut-cmdline.service": {
"name": "dracut-cmdline.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-initqueue.service": {
"name": "dracut-initqueue.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-mount.service": {
"name": "dracut-mount.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-mount.service": {
"name": "dracut-pre-mount.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-pivot.service": {
"name": "dracut-pre-pivot.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-trigger.service": {
"name": "dracut-pre-trigger.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-udev.service": {
"name": "dracut-pre-udev.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-shutdown-onfailure.service": {
"name": "dracut-shutdown-onfailure.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-shutdown.service": {
"name": "dracut-shutdown.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"ebtables.service": {
"name": "ebtables.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"emergency.service": {
"name": "emergency.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"firewalld.service": {
"name": "firewalld.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"fstrim.service": {
"name": "fstrim.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"getty@.service": {
"name": "getty@.service",
"source": "systemd",
"state": "unknown",
"status": "enabled"
},
"getty@tty1.service": {
"name": "getty@tty1.service",
"source": "systemd",
"state": "running",
"status": "unknown"
},
"grub-boot-indeterminate.service": {
"name": "grub-boot-indeterminate.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"gssproxy.service": {
"name": "gssproxy.service",
"source": "systemd",
"state": "running",
"status": "disabled"
},
"halt-local.service": {
"name": "halt-local.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"import-state.service": {
"name": "import-state.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"initrd-cleanup.service": {
"name": "initrd-cleanup.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"initrd-parse-etc.service": {
"name": "initrd-parse-etc.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"initrd-switch-root.service": {
"name": "initrd-switch-root.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"initrd-udevadm-cleanup-db.service": {
"name": "initrd-udevadm-cleanup-db.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"iprdump.service": {
"name": "iprdump.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"iprinit.service": {
"name": "iprinit.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"iprupdate.service": {
"name": "iprupdate.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"irqbalance.service": {
"name": "irqbalance.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"kdump.service": {
"name": "kdump.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"kmod-static-nodes.service": {
"name": "kmod-static-nodes.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"kvm_stat.service": {
"name": "kvm_stat.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"ldconfig.service": {
"name": "ldconfig.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"loadmodules.service": {
"name": "loadmodules.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"man-db-cache-update.service": {
"name": "man-db-cache-update.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"man-db-restart-cache-update.service": {
"name": "man-db-restart-cache-update.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"messagebus.service": {
"name": "messagebus.service",
"source": "systemd",
"state": "active",
"status": "static"
},
"microcode.service": {
"name": "microcode.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"nfs-blkmap.service": {
"name": "nfs-blkmap.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"nfs-convert.service": {
"name": "nfs-convert.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"nfs-idmapd.service": {
"name": "nfs-idmapd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nfs-mountd.service": {
"name": "nfs-mountd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nfs-server.service": {
"name": "nfs-server.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"nfs-utils.service": {
"name": "nfs-utils.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nfsdcld.service": {
"name": "nfsdcld.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nftables.service": {
"name": "nftables.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"nis-domainname.service": {
"name": "nis-domainname.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"oddjobd.service": {
"name": "oddjobd.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"plymouth-halt.service": {
"name": "plymouth-halt.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-kexec.service": {
"name": "plymouth-kexec.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-poweroff.service": {
"name": "plymouth-poweroff.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-quit-wait.service": {
"name": "plymouth-quit-wait.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-quit.service": {
"name": "plymouth-quit.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-read-write.service": {
"name": "plymouth-read-write.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-reboot.service": {
"name": "plymouth-reboot.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-start.service": {
"name": "plymouth-start.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-switch-root-initramfs.service": {
"name": "plymouth-switch-root-initramfs.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-switch-root.service": {
"name": "plymouth-switch-root.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"podman-auto-update.service": {
"name": "podman-auto-update.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"podman-clean-transient.service": {
"name": "podman-clean-transient.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"podman-kube@.service": {
"name": "podman-kube@.service",
"source": "systemd",
"state": "unknown",
"status": "disabled"
},
"podman-restart.service": {
"name": "podman-restart.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"podman.service": {
"name": "podman.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"polkit.service": {
"name": "polkit.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"qemu-guest-agent.service": {
"name": "qemu-guest-agent.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"quadlet-demo-mysql-volume.service": {
"name": "quadlet-demo-mysql-volume.service",
"source": "systemd",
"state": "stopped",
"status": "generated"
},
"quadlet-demo-mysql.service": {
"name": "quadlet-demo-mysql.service",
"source": "systemd",
"state": "running",
"status": "generated"
},
"quadlet-demo-network.service": {
"name": "quadlet-demo-network.service",
"source": "systemd",
"state": "stopped",
"status": "generated"
},
"quotaon.service": {
"name": "quotaon.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"rc-local.service": {
"name": "rc-local.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rdisc.service": {
"name": "rdisc.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"rescue.service": {
"name": "rescue.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"restraintd.service": {
"name": "restraintd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"rngd.service": {
"name": "rngd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"rpc-gssd.service": {
"name": "rpc-gssd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rpc-statd-notify.service": {
"name": "rpc-statd-notify.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rpc-statd.service": {
"name": "rpc-statd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rpcbind.service": {
"name": "rpcbind.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"rsyslog.service": {
"name": "rsyslog.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"selinux-autorelabel-mark.service": {
"name": "selinux-autorelabel-mark.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"selinux-autorelabel.service": {
"name": "selinux-autorelabel.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"serial-getty@.service": {
"name": "serial-getty@.service",
"source": "systemd",
"state": "unknown",
"status": "disabled"
},
"sshd-keygen@.service": {
"name": "sshd-keygen@.service",
"source": "systemd",
"state": "unknown",
"status": "disabled"
},
"sshd-keygen@ecdsa.service": {
"name": "sshd-keygen@ecdsa.service",
"source": "systemd",
"state": "stopped",
"status": "unknown"
},
"sshd-keygen@ed25519.service": {
"name": "sshd-keygen@ed25519.service",
"source": "systemd",
"state": "stopped",
"status": "unknown"
},
"sshd-keygen@rsa.service": {
"name": "sshd-keygen@rsa.service",
"source": "systemd",
"state": "stopped",
"status": "unknown"
},
"sshd.service": {
"name": "sshd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"sshd@.service": {
"name": "sshd@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"sssd-autofs.service": {
"name": "sssd-autofs.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-kcm.service": {
"name": "sssd-kcm.service",
"source": "systemd",
"state": "stopped",
"status": "indirect"
},
"sssd-nss.service": {
"name": "sssd-nss.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-pac.service": {
"name": "sssd-pac.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-pam.service": {
"name": "sssd-pam.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-ssh.service": {
"name": "sssd-ssh.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-sudo.service": {
"name": "sssd-sudo.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd.service": {
"name": "sssd.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"syslog.service": {
"name": "syslog.service",
"source": "systemd",
"state": "active",
"status": "enabled"
},
"system-update-cleanup.service": {
"name": "system-update-cleanup.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-ask-password-console.service": {
"name": "systemd-ask-password-console.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-ask-password-plymouth.service": {
"name": "systemd-ask-password-plymouth.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-ask-password-wall.service": {
"name": "systemd-ask-password-wall.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-backlight@.service": {
"name": "systemd-backlight@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-binfmt.service": {
"name": "systemd-binfmt.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-coredump@.service": {
"name": "systemd-coredump@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-exit.service": {
"name": "systemd-exit.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-firstboot.service": {
"name": "systemd-firstboot.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-fsck-root.service": {
"name": "systemd-fsck-root.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-fsck@.service": {
"name": "systemd-fsck@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-halt.service": {
"name": "systemd-halt.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-hibernate-resume@.service": {
"name": "systemd-hibernate-resume@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-hibernate.service": {
"name": "systemd-hibernate.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-hostnamed.service": {
"name": "systemd-hostnamed.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-hwdb-update.service": {
"name": "systemd-hwdb-update.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-hybrid-sleep.service": {
"name": "systemd-hybrid-sleep.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-initctl.service": {
"name": "systemd-initctl.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-journal-catalog-update.service": {
"name": "systemd-journal-catalog-update.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-journal-flush.service": {
"name": "systemd-journal-flush.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-journald.service": {
"name": "systemd-journald.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"systemd-kexec.service": {
"name": "systemd-kexec.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-localed.service": {
"name": "systemd-localed.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-logind.service": {
"name": "systemd-logind.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-machine-id-commit.service": {
"name": "systemd-machine-id-commit.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-modules-load.service": {
"name": "systemd-modules-load.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-portabled.service": {
"name": "systemd-portabled.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-poweroff.service": {
"name": "systemd-poweroff.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-pstore.service": {
"name": "systemd-pstore.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"systemd-quotacheck.service": {
"name": "systemd-quotacheck.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-random-seed.service": {
"name": "systemd-random-seed.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-reboot.service": {
"name": "systemd-reboot.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-remount-fs.service": {
"name": "systemd-remount-fs.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-resolved.service": {
"name": "systemd-resolved.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"systemd-rfkill.service": {
"name": "systemd-rfkill.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-suspend-then-hibernate.service": {
"name": "systemd-suspend-then-hibernate.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-suspend.service": {
"name": "systemd-suspend.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-sysctl.service": {
"name": "systemd-sysctl.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-sysusers.service": {
"name": "systemd-sysusers.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-timedated.service": {
"name": "systemd-timedated.service",
"source": "systemd",
"state": "inactive",
"status": "masked"
},
"systemd-tmpfiles-clean.service": {
"name": "systemd-tmpfiles-clean.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-tmpfiles-setup-dev.service": {
"name": "systemd-tmpfiles-setup-dev.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-tmpfiles-setup.service": {
"name": "systemd-tmpfiles-setup.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-udev-settle.service": {
"name": "systemd-udev-settle.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-udev-trigger.service": {
"name": "systemd-udev-trigger.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-udevd.service": {
"name": "systemd-udevd.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"systemd-update-done.service": {
"name": "systemd-update-done.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-update-utmp-runlevel.service": {
"name": "systemd-update-utmp-runlevel.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-update-utmp.service": {
"name": "systemd-update-utmp.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-user-sessions.service": {
"name": "systemd-user-sessions.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-vconsole-setup.service": {
"name": "systemd-vconsole-setup.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-volatile-root.service": {
"name": "systemd-volatile-root.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"tcsd.service": {
"name": "tcsd.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"teamd@.service": {
"name": "teamd@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"timedatex.service": {
"name": "timedatex.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"tuned.service": {
"name": "tuned.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"unbound-anchor.service": {
"name": "unbound-anchor.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"user-runtime-dir@.service": {
"name": "user-runtime-dir@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"user-runtime-dir@0.service": {
"name": "user-runtime-dir@0.service",
"source": "systemd",
"state": "stopped",
"status": "unknown"
},
"user@.service": {
"name": "user@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"user@0.service": {
"name": "user@0.service",
"source": "systemd",
"state": "running",
"status": "unknown"
}
}
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Create and update quadlets] ***********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:114
Saturday 24 August 2024 12:36:12 -0400 (0:00:01.601) 0:02:20.747 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14
Saturday 24 August 2024 12:36:12 -0400 (0:00:00.033) 0:02:20.780 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_quadlet_file_src": "",
"__podman_quadlet_spec": {},
"__podman_quadlet_str": "[Install]\nWantedBy=default.target\n\n[Container]\nImage=quay.io/linux-system-roles/mysql:5.6\nContainerName=quadlet-demo-mysql\nVolume=quadlet-demo-mysql.volume:/var/lib/mysql\nVolume=/tmp/quadlet_demo:/var/lib/quadlet_demo:Z\nNetwork=quadlet-demo.network\nSecret=mysql-root-password-container,type=env,target=MYSQL_ROOT_PASSWORD\nHealthCmd=/bin/true\nHealthOnFailure=kill\n",
"__podman_quadlet_template_src": "quadlet-demo-mysql.container.j2"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25
Saturday 24 August 2024 12:36:12 -0400 (0:00:00.107) 0:02:20.888 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_continue_if_pull_fails": false,
"__podman_pull_image": true,
"__podman_state": "absent",
"__podman_systemd_unit_scope": "",
"__podman_user": "root"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] *****
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35
Saturday 24 August 2024 12:36:12 -0400 (0:00:00.042) 0:02:20.930 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48
Saturday 24 August 2024 12:36:12 -0400 (0:00:00.035) 0:02:20.965 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_quadlet_name": "quadlet-demo-mysql",
"__podman_quadlet_type": "container",
"__podman_rootless": false
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Check user and group information] *****
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57
Saturday 24 August 2024 12:36:12 -0400 (0:00:00.081) 0:02:21.047 *******
included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for managed_node1
TASK [fedora.linux_system_roles.podman : Get user information] *****************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2
Saturday 24 August 2024 12:36:12 -0400 (0:00:00.063) 0:02:21.110 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Fail if user does not exist] **********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9
Saturday 24 August 2024 12:36:12 -0400 (0:00:00.040) 0:02:21.151 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set group for podman user] ************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16
Saturday 24 August 2024 12:36:12 -0400 (0:00:00.045) 0:02:21.197 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_group": "0"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Get group information] ****************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:28
Saturday 24 August 2024 12:36:12 -0400 (0:00:00.045) 0:02:21.242 *******
ok: [managed_node1] => {
"ansible_facts": {
"getent_group": {
"root": [
"x",
"0",
""
]
}
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set group name] ***********************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:35
Saturday 24 August 2024 12:36:12 -0400 (0:00:00.369) 0:02:21.611 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_group_name": "root"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : See if getsubids exists] **************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:39
Saturday 24 August 2024 12:36:13 -0400 (0:00:00.043) 0:02:21.655 *******
ok: [managed_node1] => {
"changed": false,
"stat": {
"atime": 1724516929.7168462,
"attr_flags": "",
"attributes": [],
"block_size": 4096,
"blocks": 32,
"charset": "binary",
"checksum": "bb5b46ffbafcaa8c4021f3c8b3cb8594f48ef34b",
"ctime": 1724516896.725572,
"dev": 51713,
"device_type": 0,
"executable": true,
"exists": true,
"gid": 0,
"gr_name": "root",
"inode": 6986657,
"isblk": false,
"ischr": false,
"isdir": false,
"isfifo": false,
"isgid": false,
"islnk": false,
"isreg": true,
"issock": false,
"isuid": false,
"mimetype": "application/x-sharedlib",
"mode": "0755",
"mtime": 1700557386.0,
"nlink": 1,
"path": "/usr/bin/getsubids",
"pw_name": "root",
"readable": true,
"rgrp": true,
"roth": true,
"rusr": true,
"size": 12640,
"uid": 0,
"version": "1255679238",
"wgrp": false,
"woth": false,
"writeable": true,
"wusr": true,
"xgrp": true,
"xoth": true,
"xusr": true
}
}
TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:50
Saturday 24 August 2024 12:36:13 -0400 (0:00:00.363) 0:02:22.018 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Check group with getsubids] ***********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:55
Saturday 24 August 2024 12:36:13 -0400 (0:00:00.032) 0:02:22.051 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ******
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:60
Saturday 24 August 2024 12:36:13 -0400 (0:00:00.033) 0:02:22.084 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Get subuid file] **********************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:74
Saturday 24 August 2024 12:36:13 -0400 (0:00:00.033) 0:02:22.117 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Get subgid file] **********************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:79
Saturday 24 August 2024 12:36:13 -0400 (0:00:00.032) 0:02:22.150 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ******
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:84
Saturday 24 August 2024 12:36:13 -0400 (0:00:00.077) 0:02:22.227 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ******
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:94
Saturday 24 August 2024 12:36:13 -0400 (0:00:00.033) 0:02:22.261 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] *****
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:101
Saturday 24 August 2024 12:36:13 -0400 (0:00:00.035) 0:02:22.296 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62
Saturday 24 August 2024 12:36:13 -0400 (0:00:00.034) 0:02:22.330 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_activate_systemd_unit": true,
"__podman_images_found": [
"quay.io/linux-system-roles/mysql:5.6"
],
"__podman_kube_yamls_raw": "",
"__podman_service_name": "quadlet-demo-mysql.service",
"__podman_systemd_scope": "system",
"__podman_user_home_dir": "/root",
"__podman_xdg_runtime_dir": "/run/user/0"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73
Saturday 24 August 2024 12:36:13 -0400 (0:00:00.060) 0:02:22.391 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_quadlet_path": "/etc/containers/systemd"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Get kube yaml contents] ***************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77
Saturday 24 August 2024 12:36:13 -0400 (0:00:00.035) 0:02:22.426 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87
Saturday 24 August 2024 12:36:13 -0400 (0:00:00.034) 0:02:22.461 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_images": [
"quay.io/linux-system-roles/mysql:5.6"
],
"__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo-mysql.container",
"__podman_volumes": [
"/tmp/quadlet_demo"
]
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 6] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:103
Saturday 24 August 2024 12:36:13 -0400 (0:00:00.074) 0:02:22.535 *******
ok: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : Cleanup quadlets] *********************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:110
Saturday 24 August 2024 12:36:13 -0400 (0:00:00.039) 0:02:22.574 *******
included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml for managed_node1
TASK [fedora.linux_system_roles.podman : Stat XDG_RUNTIME_DIR] *****************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:4
Saturday 24 August 2024 12:36:14 -0400 (0:00:00.078) 0:02:22.653 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Stop and disable service] *************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:12
Saturday 24 August 2024 12:36:14 -0400 (0:00:00.034) 0:02:22.687 *******
changed: [managed_node1] => {
"changed": true,
"enabled": false,
"failed_when_result": false,
"name": "quadlet-demo-mysql.service",
"state": "stopped",
"status": {
"ActiveEnterTimestamp": "Sat 2024-08-24 12:34:48 EDT",
"ActiveEnterTimestampMonotonic": "691670672",
"ActiveExitTimestampMonotonic": "0",
"ActiveState": "active",
"After": "basic.target system.slice quadlet-demo-network.service quadlet-demo-mysql-volume.service systemd-journald.socket tmp.mount -.mount sysinit.target",
"AllowIsolate": "no",
"AllowedCPUs": "",
"AllowedMemoryNodes": "",
"AmbientCapabilities": "",
"AssertResult": "yes",
"AssertTimestamp": "Sat 2024-08-24 12:34:48 EDT",
"AssertTimestampMonotonic": "691284549",
"Before": "shutdown.target",
"BlockIOAccounting": "no",
"BlockIOWeight": "[not set]",
"CPUAccounting": "no",
"CPUAffinity": "",
"CPUAffinityFromNUMA": "no",
"CPUQuotaPerSecUSec": "infinity",
"CPUQuotaPeriodUSec": "infinity",
"CPUSchedulingPolicy": "0",
"CPUSchedulingPriority": "0",
"CPUSchedulingResetOnFork": "no",
"CPUShares": "[not set]",
"CPUUsageNSec": "[not set]",
"CPUWeight": "[not set]",
"CacheDirectoryMode": "0755",
"CanFreeze": "yes",
"CanIsolate": "no",
"CanReload": "no",
"CanStart": "yes",
"CanStop": "yes",
"CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf",
"CollectMode": "inactive",
"ConditionResult": "yes",
"ConditionTimestamp": "Sat 2024-08-24 12:34:48 EDT",
"ConditionTimestampMonotonic": "691284547",
"ConfigurationDirectoryMode": "0755",
"Conflicts": "shutdown.target",
"ControlGroup": "/system.slice/quadlet-demo-mysql.service",
"ControlPID": "0",
"DefaultDependencies": "yes",
"DefaultMemoryLow": "0",
"DefaultMemoryMin": "0",
"Delegate": "yes",
"DelegateControllers": "cpu cpuacct cpuset io blkio memory devices pids",
"Description": "quadlet-demo-mysql.service",
"DevicePolicy": "auto",
"DynamicUser": "no",
"EffectiveCPUs": "",
"EffectiveMemoryNodes": "",
"Environment": "PODMAN_SYSTEMD_UNIT=quadlet-demo-mysql.service",
"ExecMainCode": "0",
"ExecMainExitTimestampMonotonic": "0",
"ExecMainPID": "84084",
"ExecMainStartTimestamp": "Sat 2024-08-24 12:34:48 EDT",
"ExecMainStartTimestampMonotonic": "691670652",
"ExecMainStatus": "0",
"ExecStart": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman run --name=quadlet-demo-mysql --cidfile=/run/quadlet-demo-mysql.cid --replace --rm --cgroups=split --network=systemd-quadlet-demo --sdnotify=conmon -d -v systemd-quadlet-demo-mysql:/var/lib/mysql -v /tmp/quadlet_demo:/var/lib/quadlet_demo:Z --secret mysql-root-password-container,type=env,target=MYSQL_ROOT_PASSWORD --health-cmd /bin/true --health-on-failure kill quay.io/linux-system-roles/mysql:5.6 ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"ExecStop": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman rm -v -f -i --cidfile=/run/quadlet-demo-mysql.cid ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"ExecStopPost": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman rm -v -f -i --cidfile=/run/quadlet-demo-mysql.cid ; ignore_errors=yes ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"FailureAction": "none",
"FileDescriptorStoreMax": "0",
"FragmentPath": "/run/systemd/generator/quadlet-demo-mysql.service",
"FreezerState": "running",
"GID": "[not set]",
"GuessMainPID": "yes",
"IOAccounting": "no",
"IOSchedulingClass": "0",
"IOSchedulingPriority": "0",
"IOWeight": "[not set]",
"IPAccounting": "no",
"IPEgressBytes": "18446744073709551615",
"IPEgressPackets": "18446744073709551615",
"IPIngressBytes": "18446744073709551615",
"IPIngressPackets": "18446744073709551615",
"Id": "quadlet-demo-mysql.service",
"IgnoreOnIsolate": "no",
"IgnoreSIGPIPE": "yes",
"InactiveEnterTimestampMonotonic": "0",
"InactiveExitTimestamp": "Sat 2024-08-24 12:34:48 EDT",
"InactiveExitTimestampMonotonic": "691285998",
"InvocationID": "f636d1c329514adfa01cbd36661564f5",
"JobRunningTimeoutUSec": "infinity",
"JobTimeoutAction": "none",
"JobTimeoutUSec": "infinity",
"KeyringMode": "private",
"KillMode": "mixed",
"KillSignal": "15",
"LimitAS": "infinity",
"LimitASSoft": "infinity",
"LimitCORE": "infinity",
"LimitCORESoft": "0",
"LimitCPU": "infinity",
"LimitCPUSoft": "infinity",
"LimitDATA": "infinity",
"LimitDATASoft": "infinity",
"LimitFSIZE": "infinity",
"LimitFSIZESoft": "infinity",
"LimitLOCKS": "infinity",
"LimitLOCKSSoft": "infinity",
"LimitMEMLOCK": "65536",
"LimitMEMLOCKSoft": "65536",
"LimitMSGQUEUE": "819200",
"LimitMSGQUEUESoft": "819200",
"LimitNICE": "0",
"LimitNICESoft": "0",
"LimitNOFILE": "262144",
"LimitNOFILESoft": "1024",
"LimitNPROC": "14003",
"LimitNPROCSoft": "14003",
"LimitRSS": "infinity",
"LimitRSSSoft": "infinity",
"LimitRTPRIO": "0",
"LimitRTPRIOSoft": "0",
"LimitRTTIME": "infinity",
"LimitRTTIMESoft": "infinity",
"LimitSIGPENDING": "14003",
"LimitSIGPENDINGSoft": "14003",
"LimitSTACK": "infinity",
"LimitSTACKSoft": "8388608",
"LoadState": "loaded",
"LockPersonality": "no",
"LogLevelMax": "-1",
"LogRateLimitBurst": "0",
"LogRateLimitIntervalUSec": "0",
"LogsDirectoryMode": "0755",
"MainPID": "84084",
"MemoryAccounting": "yes",
"MemoryCurrent": "604258304",
"MemoryDenyWriteExecute": "no",
"MemoryHigh": "infinity",
"MemoryLimit": "infinity",
"MemoryLow": "0",
"MemoryMax": "infinity",
"MemoryMin": "0",
"MemorySwapMax": "infinity",
"MountAPIVFS": "no",
"MountFlags": "",
"NFileDescriptorStore": "0",
"NRestarts": "0",
"NUMAMask": "",
"NUMAPolicy": "n/a",
"Names": "quadlet-demo-mysql.service",
"NeedDaemonReload": "no",
"Nice": "0",
"NoNewPrivileges": "no",
"NonBlocking": "no",
"NotifyAccess": "all",
"OOMScoreAdjust": "0",
"OnFailureJobMode": "replace",
"PermissionsStartOnly": "no",
"Perpetual": "no",
"PrivateDevices": "no",
"PrivateMounts": "no",
"PrivateNetwork": "no",
"PrivateTmp": "no",
"PrivateUsers": "no",
"ProtectControlGroups": "no",
"ProtectHome": "no",
"ProtectKernelModules": "no",
"ProtectKernelTunables": "no",
"ProtectSystem": "no",
"RefuseManualStart": "no",
"RefuseManualStop": "no",
"RemainAfterExit": "no",
"RemoveIPC": "no",
"Requires": "quadlet-demo-mysql-volume.service sysinit.target system.slice -.mount quadlet-demo-network.service",
"RequiresMountsFor": "/tmp/quadlet_demo /run/containers",
"Restart": "no",
"RestartUSec": "100ms",
"RestrictNamespaces": "no",
"RestrictRealtime": "no",
"RestrictSUIDSGID": "no",
"Result": "success",
"RootDirectoryStartOnly": "no",
"RuntimeDirectoryMode": "0755",
"RuntimeDirectoryPreserve": "no",
"RuntimeMaxUSec": "infinity",
"SameProcessGroup": "no",
"SecureBits": "0",
"SendSIGHUP": "no",
"SendSIGKILL": "yes",
"Slice": "system.slice",
"SourcePath": "/etc/containers/systemd/quadlet-demo-mysql.container",
"StandardError": "inherit",
"StandardInput": "null",
"StandardInputData": "",
"StandardOutput": "journal",
"StartLimitAction": "none",
"StartLimitBurst": "5",
"StartLimitIntervalUSec": "10s",
"StartupBlockIOWeight": "[not set]",
"StartupCPUShares": "[not set]",
"StartupCPUWeight": "[not set]",
"StartupIOWeight": "[not set]",
"StateChangeTimestamp": "Sat 2024-08-24 12:34:48 EDT",
"StateChangeTimestampMonotonic": "691670672",
"StateDirectoryMode": "0755",
"StatusErrno": "0",
"StopWhenUnneeded": "no",
"SubState": "running",
"SuccessAction": "none",
"SyslogFacility": "3",
"SyslogIdentifier": "quadlet-demo-mysql",
"SyslogLevel": "6",
"SyslogLevelPrefix": "yes",
"SyslogPriority": "30",
"SystemCallErrorNumber": "0",
"TTYReset": "no",
"TTYVHangup": "no",
"TTYVTDisallocate": "no",
"TasksAccounting": "yes",
"TasksCurrent": "24",
"TasksMax": "22405",
"TimeoutStartUSec": "1min 30s",
"TimeoutStopUSec": "1min 30s",
"TimerSlackNSec": "50000",
"Transient": "no",
"Type": "notify",
"UID": "[not set]",
"UMask": "0022",
"UnitFilePreset": "disabled",
"UnitFileState": "generated",
"UtmpMode": "init",
"WatchdogTimestamp": "Sat 2024-08-24 12:34:48 EDT",
"WatchdogTimestampMonotonic": "691670669",
"WatchdogUSec": "0"
}
}
TASK [fedora.linux_system_roles.podman : See if quadlet file exists] ***********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:33
Saturday 24 August 2024 12:36:16 -0400 (0:00:01.942) 0:02:24.630 *******
ok: [managed_node1] => {
"changed": false,
"stat": {
"atime": 1724517287.8987753,
"attr_flags": "",
"attributes": [],
"block_size": 4096,
"blocks": 8,
"charset": "us-ascii",
"checksum": "ca62b2ad3cc9afb5b5371ebbf797b9bc4fd7edd4",
"ctime": 1724517287.3017704,
"dev": 51713,
"device_type": 0,
"executable": false,
"exists": true,
"gid": 0,
"gr_name": "root",
"inode": 369098918,
"isblk": false,
"ischr": false,
"isdir": false,
"isfifo": false,
"isgid": false,
"islnk": false,
"isreg": true,
"issock": false,
"isuid": false,
"mimetype": "text/plain",
"mode": "0644",
"mtime": 1724517287.0177681,
"nlink": 1,
"path": "/etc/containers/systemd/quadlet-demo-mysql.container",
"pw_name": "root",
"readable": true,
"rgrp": true,
"roth": true,
"rusr": true,
"size": 363,
"uid": 0,
"version": "4287155550",
"wgrp": false,
"woth": false,
"writeable": true,
"wusr": true,
"xgrp": false,
"xoth": false,
"xusr": false
}
}
TASK [fedora.linux_system_roles.podman : Parse quadlet file] *******************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:38
Saturday 24 August 2024 12:36:16 -0400 (0:00:00.365) 0:02:24.996 *******
included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml for managed_node1
TASK [fedora.linux_system_roles.podman : Slurp quadlet file] *******************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml:6
Saturday 24 August 2024 12:36:16 -0400 (0:00:00.061) 0:02:25.057 *******
ok: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : Parse quadlet file] *******************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml:12
Saturday 24 August 2024 12:36:16 -0400 (0:00:00.352) 0:02:25.410 *******
ok: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : Parse quadlet yaml file] **************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml:44
Saturday 24 August 2024 12:36:16 -0400 (0:00:00.047) 0:02:25.457 *******
skipping: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : Reset raw variable] *******************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml:52
Saturday 24 August 2024 12:36:16 -0400 (0:00:00.031) 0:02:25.489 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_quadlet_raw": null
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Remove quadlet file] ******************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:42
Saturday 24 August 2024 12:36:16 -0400 (0:00:00.032) 0:02:25.521 *******
changed: [managed_node1] => {
"changed": true,
"path": "/etc/containers/systemd/quadlet-demo-mysql.container",
"state": "absent"
}
TASK [fedora.linux_system_roles.podman : Refresh systemd] **********************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:48
Saturday 24 August 2024 12:36:17 -0400 (0:00:00.360) 0:02:25.882 *******
ok: [managed_node1] => {
"changed": false,
"name": null,
"status": {}
}
TASK [fedora.linux_system_roles.podman : Remove managed resource] **************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:58
Saturday 24 August 2024 12:36:17 -0400 (0:00:00.656) 0:02:26.538 *******
ok: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : Remove volumes] ***********************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:95
Saturday 24 August 2024 12:36:18 -0400 (0:00:00.415) 0:02:26.954 *******
skipping: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : Clear parsed podman variable] *********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:112
Saturday 24 August 2024 12:36:18 -0400 (0:00:00.048) 0:02:27.002 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_quadlet_parsed": null
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Prune images no longer in use] ********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:116
Saturday 24 August 2024 12:36:18 -0400 (0:00:00.032) 0:02:27.034 *******
changed: [managed_node1] => {
"changed": true,
"cmd": [
"podman",
"image",
"prune",
"--all",
"-f"
],
"delta": "0:00:00.234770",
"end": "2024-08-24 12:36:18.949782",
"rc": 0,
"start": "2024-08-24 12:36:18.715012"
}
STDOUT:
dd3b2a5dcb48ff61113592ed5ddd762581be4387c7bc552375a2159422aa6bf5
TASK [fedora.linux_system_roles.podman : Manage linger] ************************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:127
Saturday 24 August 2024 12:36:19 -0400 (0:00:00.598) 0:02:27.632 *******
included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for managed_node1
TASK [fedora.linux_system_roles.podman : Enable linger if needed] **************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12
Saturday 24 August 2024 12:36:19 -0400 (0:00:00.060) 0:02:27.692 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18
Saturday 24 August 2024 12:36:19 -0400 (0:00:00.032) 0:02:27.724 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22
Saturday 24 August 2024 12:36:19 -0400 (0:00:00.032) 0:02:27.757 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : For testing and debugging - images] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:137
Saturday 24 August 2024 12:36:19 -0400 (0:00:00.032) 0:02:27.789 *******
ok: [managed_node1] => {
"changed": false,
"cmd": [
"podman",
"images",
"-n"
],
"delta": "0:00:00.036357",
"end": "2024-08-24 12:36:19.506720",
"rc": 0,
"start": "2024-08-24 12:36:19.470363"
}
TASK [fedora.linux_system_roles.podman : For testing and debugging - volumes] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:146
Saturday 24 August 2024 12:36:19 -0400 (0:00:00.447) 0:02:28.237 *******
ok: [managed_node1] => {
"changed": false,
"cmd": [
"podman",
"volume",
"ls",
"-n"
],
"delta": "0:00:00.036451",
"end": "2024-08-24 12:36:19.956747",
"rc": 0,
"start": "2024-08-24 12:36:19.920296"
}
STDOUT:
local systemd-quadlet-demo-mysql
local wp-pv-claim
local envoy-proxy-config
local envoy-certificates
TASK [fedora.linux_system_roles.podman : For testing and debugging - containers] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:155
Saturday 24 August 2024 12:36:20 -0400 (0:00:00.401) 0:02:28.639 *******
ok: [managed_node1] => {
"changed": false,
"cmd": [
"podman",
"ps",
"--noheading"
],
"delta": "0:00:00.031885",
"end": "2024-08-24 12:36:20.352641",
"rc": 0,
"start": "2024-08-24 12:36:20.320756"
}
TASK [fedora.linux_system_roles.podman : For testing and debugging - networks] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:164
Saturday 24 August 2024 12:36:20 -0400 (0:00:00.396) 0:02:29.035 *******
ok: [managed_node1] => {
"changed": false,
"cmd": [
"podman",
"network",
"ls",
"-n",
"-q"
],
"delta": "0:00:00.057821",
"end": "2024-08-24 12:36:20.775500",
"rc": 0,
"start": "2024-08-24 12:36:20.717679"
}
STDOUT:
podman
podman-default-kube-network
systemd-quadlet-demo
TASK [fedora.linux_system_roles.podman : For testing and debugging - secrets] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:173
Saturday 24 August 2024 12:36:20 -0400 (0:00:00.422) 0:02:29.458 *******
ok: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : For testing and debugging - services] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:183
Saturday 24 August 2024 12:36:21 -0400 (0:00:00.399) 0:02:29.858 *******
ok: [managed_node1] => {
"ansible_facts": {
"services": {
"NetworkManager-dispatcher.service": {
"name": "NetworkManager-dispatcher.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"NetworkManager-wait-online.service": {
"name": "NetworkManager-wait-online.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"NetworkManager.service": {
"name": "NetworkManager.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"auditd.service": {
"name": "auditd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"auth-rpcgss-module.service": {
"name": "auth-rpcgss-module.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"autovt@.service": {
"name": "autovt@.service",
"source": "systemd",
"state": "unknown",
"status": "enabled"
},
"certmonger.service": {
"name": "certmonger.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"chrony-dnssrv@.service": {
"name": "chrony-dnssrv@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"chrony-wait.service": {
"name": "chrony-wait.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"chronyd.service": {
"name": "chronyd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"cloud-config.service": {
"name": "cloud-config.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"cloud-final.service": {
"name": "cloud-final.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"cloud-init-hotplugd.service": {
"name": "cloud-init-hotplugd.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"cloud-init-local.service": {
"name": "cloud-init-local.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"cloud-init.service": {
"name": "cloud-init.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"cni-dhcp.service": {
"name": "cni-dhcp.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"console-getty.service": {
"name": "console-getty.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"container-getty@.service": {
"name": "container-getty@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"cpupower.service": {
"name": "cpupower.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"crond.service": {
"name": "crond.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"dbus-org.fedoraproject.FirewallD1.service": {
"name": "dbus-org.fedoraproject.FirewallD1.service",
"source": "systemd",
"state": "active",
"status": "enabled"
},
"dbus-org.freedesktop.hostname1.service": {
"name": "dbus-org.freedesktop.hostname1.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dbus-org.freedesktop.locale1.service": {
"name": "dbus-org.freedesktop.locale1.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dbus-org.freedesktop.login1.service": {
"name": "dbus-org.freedesktop.login1.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dbus-org.freedesktop.nm-dispatcher.service": {
"name": "dbus-org.freedesktop.nm-dispatcher.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"dbus-org.freedesktop.portable1.service": {
"name": "dbus-org.freedesktop.portable1.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dbus-org.freedesktop.timedate1.service": {
"name": "dbus-org.freedesktop.timedate1.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"dbus.service": {
"name": "dbus.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"debug-shell.service": {
"name": "debug-shell.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"dnf-makecache.service": {
"name": "dnf-makecache.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dnf-system-upgrade-cleanup.service": {
"name": "dnf-system-upgrade-cleanup.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dnf-system-upgrade.service": {
"name": "dnf-system-upgrade.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"dnsmasq.service": {
"name": "dnsmasq.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"dracut-cmdline.service": {
"name": "dracut-cmdline.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-initqueue.service": {
"name": "dracut-initqueue.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-mount.service": {
"name": "dracut-mount.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-mount.service": {
"name": "dracut-pre-mount.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-pivot.service": {
"name": "dracut-pre-pivot.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-trigger.service": {
"name": "dracut-pre-trigger.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-udev.service": {
"name": "dracut-pre-udev.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-shutdown-onfailure.service": {
"name": "dracut-shutdown-onfailure.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-shutdown.service": {
"name": "dracut-shutdown.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"ebtables.service": {
"name": "ebtables.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"emergency.service": {
"name": "emergency.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"firewalld.service": {
"name": "firewalld.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"fstrim.service": {
"name": "fstrim.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"getty@.service": {
"name": "getty@.service",
"source": "systemd",
"state": "unknown",
"status": "enabled"
},
"getty@tty1.service": {
"name": "getty@tty1.service",
"source": "systemd",
"state": "running",
"status": "unknown"
},
"grub-boot-indeterminate.service": {
"name": "grub-boot-indeterminate.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"gssproxy.service": {
"name": "gssproxy.service",
"source": "systemd",
"state": "running",
"status": "disabled"
},
"halt-local.service": {
"name": "halt-local.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"import-state.service": {
"name": "import-state.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"initrd-cleanup.service": {
"name": "initrd-cleanup.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"initrd-parse-etc.service": {
"name": "initrd-parse-etc.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"initrd-switch-root.service": {
"name": "initrd-switch-root.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"initrd-udevadm-cleanup-db.service": {
"name": "initrd-udevadm-cleanup-db.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"iprdump.service": {
"name": "iprdump.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"iprinit.service": {
"name": "iprinit.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"iprupdate.service": {
"name": "iprupdate.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"irqbalance.service": {
"name": "irqbalance.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"kdump.service": {
"name": "kdump.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"kmod-static-nodes.service": {
"name": "kmod-static-nodes.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"kvm_stat.service": {
"name": "kvm_stat.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"ldconfig.service": {
"name": "ldconfig.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"loadmodules.service": {
"name": "loadmodules.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"man-db-cache-update.service": {
"name": "man-db-cache-update.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"man-db-restart-cache-update.service": {
"name": "man-db-restart-cache-update.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"messagebus.service": {
"name": "messagebus.service",
"source": "systemd",
"state": "active",
"status": "static"
},
"microcode.service": {
"name": "microcode.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"nfs-blkmap.service": {
"name": "nfs-blkmap.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"nfs-convert.service": {
"name": "nfs-convert.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"nfs-idmapd.service": {
"name": "nfs-idmapd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nfs-mountd.service": {
"name": "nfs-mountd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nfs-server.service": {
"name": "nfs-server.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"nfs-utils.service": {
"name": "nfs-utils.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nfsdcld.service": {
"name": "nfsdcld.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nftables.service": {
"name": "nftables.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"nis-domainname.service": {
"name": "nis-domainname.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"oddjobd.service": {
"name": "oddjobd.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"plymouth-halt.service": {
"name": "plymouth-halt.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-kexec.service": {
"name": "plymouth-kexec.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-poweroff.service": {
"name": "plymouth-poweroff.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-quit-wait.service": {
"name": "plymouth-quit-wait.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-quit.service": {
"name": "plymouth-quit.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-read-write.service": {
"name": "plymouth-read-write.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-reboot.service": {
"name": "plymouth-reboot.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-start.service": {
"name": "plymouth-start.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-switch-root-initramfs.service": {
"name": "plymouth-switch-root-initramfs.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-switch-root.service": {
"name": "plymouth-switch-root.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"podman-auto-update.service": {
"name": "podman-auto-update.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"podman-clean-transient.service": {
"name": "podman-clean-transient.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"podman-kube@.service": {
"name": "podman-kube@.service",
"source": "systemd",
"state": "unknown",
"status": "disabled"
},
"podman-restart.service": {
"name": "podman-restart.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"podman.service": {
"name": "podman.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"polkit.service": {
"name": "polkit.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"qemu-guest-agent.service": {
"name": "qemu-guest-agent.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"quadlet-demo-mysql-volume.service": {
"name": "quadlet-demo-mysql-volume.service",
"source": "systemd",
"state": "stopped",
"status": "generated"
},
"quadlet-demo-network.service": {
"name": "quadlet-demo-network.service",
"source": "systemd",
"state": "stopped",
"status": "generated"
},
"quotaon.service": {
"name": "quotaon.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"rc-local.service": {
"name": "rc-local.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rdisc.service": {
"name": "rdisc.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"rescue.service": {
"name": "rescue.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"restraintd.service": {
"name": "restraintd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"rngd.service": {
"name": "rngd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"rpc-gssd.service": {
"name": "rpc-gssd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rpc-statd-notify.service": {
"name": "rpc-statd-notify.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rpc-statd.service": {
"name": "rpc-statd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rpcbind.service": {
"name": "rpcbind.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"rsyslog.service": {
"name": "rsyslog.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"selinux-autorelabel-mark.service": {
"name": "selinux-autorelabel-mark.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"selinux-autorelabel.service": {
"name": "selinux-autorelabel.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"serial-getty@.service": {
"name": "serial-getty@.service",
"source": "systemd",
"state": "unknown",
"status": "disabled"
},
"sshd-keygen@.service": {
"name": "sshd-keygen@.service",
"source": "systemd",
"state": "unknown",
"status": "disabled"
},
"sshd-keygen@ecdsa.service": {
"name": "sshd-keygen@ecdsa.service",
"source": "systemd",
"state": "stopped",
"status": "unknown"
},
"sshd-keygen@ed25519.service": {
"name": "sshd-keygen@ed25519.service",
"source": "systemd",
"state": "stopped",
"status": "unknown"
},
"sshd-keygen@rsa.service": {
"name": "sshd-keygen@rsa.service",
"source": "systemd",
"state": "stopped",
"status": "unknown"
},
"sshd.service": {
"name": "sshd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"sshd@.service": {
"name": "sshd@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"sssd-autofs.service": {
"name": "sssd-autofs.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-kcm.service": {
"name": "sssd-kcm.service",
"source": "systemd",
"state": "stopped",
"status": "indirect"
},
"sssd-nss.service": {
"name": "sssd-nss.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-pac.service": {
"name": "sssd-pac.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-pam.service": {
"name": "sssd-pam.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-ssh.service": {
"name": "sssd-ssh.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-sudo.service": {
"name": "sssd-sudo.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd.service": {
"name": "sssd.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"syslog.service": {
"name": "syslog.service",
"source": "systemd",
"state": "active",
"status": "enabled"
},
"system-update-cleanup.service": {
"name": "system-update-cleanup.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-ask-password-console.service": {
"name": "systemd-ask-password-console.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-ask-password-plymouth.service": {
"name": "systemd-ask-password-plymouth.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-ask-password-wall.service": {
"name": "systemd-ask-password-wall.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-backlight@.service": {
"name": "systemd-backlight@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-binfmt.service": {
"name": "systemd-binfmt.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-coredump@.service": {
"name": "systemd-coredump@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-exit.service": {
"name": "systemd-exit.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-firstboot.service": {
"name": "systemd-firstboot.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-fsck-root.service": {
"name": "systemd-fsck-root.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-fsck@.service": {
"name": "systemd-fsck@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-halt.service": {
"name": "systemd-halt.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-hibernate-resume@.service": {
"name": "systemd-hibernate-resume@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-hibernate.service": {
"name": "systemd-hibernate.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-hostnamed.service": {
"name": "systemd-hostnamed.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-hwdb-update.service": {
"name": "systemd-hwdb-update.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-hybrid-sleep.service": {
"name": "systemd-hybrid-sleep.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-initctl.service": {
"name": "systemd-initctl.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-journal-catalog-update.service": {
"name": "systemd-journal-catalog-update.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-journal-flush.service": {
"name": "systemd-journal-flush.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-journald.service": {
"name": "systemd-journald.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"systemd-kexec.service": {
"name": "systemd-kexec.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-localed.service": {
"name": "systemd-localed.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-logind.service": {
"name": "systemd-logind.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-machine-id-commit.service": {
"name": "systemd-machine-id-commit.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-modules-load.service": {
"name": "systemd-modules-load.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-portabled.service": {
"name": "systemd-portabled.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-poweroff.service": {
"name": "systemd-poweroff.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-pstore.service": {
"name": "systemd-pstore.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"systemd-quotacheck.service": {
"name": "systemd-quotacheck.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-random-seed.service": {
"name": "systemd-random-seed.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-reboot.service": {
"name": "systemd-reboot.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-remount-fs.service": {
"name": "systemd-remount-fs.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-resolved.service": {
"name": "systemd-resolved.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"systemd-rfkill.service": {
"name": "systemd-rfkill.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-suspend-then-hibernate.service": {
"name": "systemd-suspend-then-hibernate.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-suspend.service": {
"name": "systemd-suspend.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-sysctl.service": {
"name": "systemd-sysctl.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-sysusers.service": {
"name": "systemd-sysusers.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-timedated.service": {
"name": "systemd-timedated.service",
"source": "systemd",
"state": "inactive",
"status": "masked"
},
"systemd-tmpfiles-clean.service": {
"name": "systemd-tmpfiles-clean.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-tmpfiles-setup-dev.service": {
"name": "systemd-tmpfiles-setup-dev.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-tmpfiles-setup.service": {
"name": "systemd-tmpfiles-setup.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-udev-settle.service": {
"name": "systemd-udev-settle.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-udev-trigger.service": {
"name": "systemd-udev-trigger.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-udevd.service": {
"name": "systemd-udevd.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"systemd-update-done.service": {
"name": "systemd-update-done.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-update-utmp-runlevel.service": {
"name": "systemd-update-utmp-runlevel.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-update-utmp.service": {
"name": "systemd-update-utmp.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-user-sessions.service": {
"name": "systemd-user-sessions.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-vconsole-setup.service": {
"name": "systemd-vconsole-setup.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-volatile-root.service": {
"name": "systemd-volatile-root.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"tcsd.service": {
"name": "tcsd.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"teamd@.service": {
"name": "teamd@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"timedatex.service": {
"name": "timedatex.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"tuned.service": {
"name": "tuned.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"unbound-anchor.service": {
"name": "unbound-anchor.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"user-runtime-dir@.service": {
"name": "user-runtime-dir@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"user-runtime-dir@0.service": {
"name": "user-runtime-dir@0.service",
"source": "systemd",
"state": "stopped",
"status": "unknown"
},
"user@.service": {
"name": "user@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"user@0.service": {
"name": "user@0.service",
"source": "systemd",
"state": "running",
"status": "unknown"
}
}
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Create and update quadlets] ***********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:114
Saturday 24 August 2024 12:36:22 -0400 (0:00:01.575) 0:02:31.433 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14
Saturday 24 August 2024 12:36:22 -0400 (0:00:00.067) 0:02:31.500 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_quadlet_file_src": "quadlet-demo-mysql.volume",
"__podman_quadlet_spec": {},
"__podman_quadlet_str": "[Volume]",
"__podman_quadlet_template_src": ""
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25
Saturday 24 August 2024 12:36:22 -0400 (0:00:00.045) 0:02:31.545 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_continue_if_pull_fails": false,
"__podman_pull_image": true,
"__podman_state": "absent",
"__podman_systemd_unit_scope": "",
"__podman_user": "root"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] *****
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35
Saturday 24 August 2024 12:36:22 -0400 (0:00:00.040) 0:02:31.586 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48
Saturday 24 August 2024 12:36:23 -0400 (0:00:00.035) 0:02:31.621 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_quadlet_name": "quadlet-demo-mysql",
"__podman_quadlet_type": "volume",
"__podman_rootless": false
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Check user and group information] *****
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57
Saturday 24 August 2024 12:36:23 -0400 (0:00:00.046) 0:02:31.667 *******
included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for managed_node1
TASK [fedora.linux_system_roles.podman : Get user information] *****************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2
Saturday 24 August 2024 12:36:23 -0400 (0:00:00.061) 0:02:31.728 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Fail if user does not exist] **********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9
Saturday 24 August 2024 12:36:23 -0400 (0:00:00.041) 0:02:31.770 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set group for podman user] ************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16
Saturday 24 August 2024 12:36:23 -0400 (0:00:00.040) 0:02:31.810 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_group": "0"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Get group information] ****************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:28
Saturday 24 August 2024 12:36:23 -0400 (0:00:00.047) 0:02:31.857 *******
ok: [managed_node1] => {
"ansible_facts": {
"getent_group": {
"root": [
"x",
"0",
""
]
}
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set group name] ***********************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:35
Saturday 24 August 2024 12:36:23 -0400 (0:00:00.366) 0:02:32.223 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_group_name": "root"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : See if getsubids exists] **************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:39
Saturday 24 August 2024 12:36:23 -0400 (0:00:00.046) 0:02:32.270 *******
ok: [managed_node1] => {
"changed": false,
"stat": {
"atime": 1724516929.7168462,
"attr_flags": "",
"attributes": [],
"block_size": 4096,
"blocks": 32,
"charset": "binary",
"checksum": "bb5b46ffbafcaa8c4021f3c8b3cb8594f48ef34b",
"ctime": 1724516896.725572,
"dev": 51713,
"device_type": 0,
"executable": true,
"exists": true,
"gid": 0,
"gr_name": "root",
"inode": 6986657,
"isblk": false,
"ischr": false,
"isdir": false,
"isfifo": false,
"isgid": false,
"islnk": false,
"isreg": true,
"issock": false,
"isuid": false,
"mimetype": "application/x-sharedlib",
"mode": "0755",
"mtime": 1700557386.0,
"nlink": 1,
"path": "/usr/bin/getsubids",
"pw_name": "root",
"readable": true,
"rgrp": true,
"roth": true,
"rusr": true,
"size": 12640,
"uid": 0,
"version": "1255679238",
"wgrp": false,
"woth": false,
"writeable": true,
"wusr": true,
"xgrp": true,
"xoth": true,
"xusr": true
}
}
TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:50
Saturday 24 August 2024 12:36:24 -0400 (0:00:00.363) 0:02:32.634 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Check group with getsubids] ***********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:55
Saturday 24 August 2024 12:36:24 -0400 (0:00:00.081) 0:02:32.715 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ******
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:60
Saturday 24 August 2024 12:36:24 -0400 (0:00:00.034) 0:02:32.750 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Get subuid file] **********************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:74
Saturday 24 August 2024 12:36:24 -0400 (0:00:00.035) 0:02:32.786 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Get subgid file] **********************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:79
Saturday 24 August 2024 12:36:24 -0400 (0:00:00.035) 0:02:32.822 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ******
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:84
Saturday 24 August 2024 12:36:24 -0400 (0:00:00.034) 0:02:32.856 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ******
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:94
Saturday 24 August 2024 12:36:24 -0400 (0:00:00.035) 0:02:32.892 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] *****
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:101
Saturday 24 August 2024 12:36:24 -0400 (0:00:00.035) 0:02:32.928 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62
Saturday 24 August 2024 12:36:24 -0400 (0:00:00.034) 0:02:32.962 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_activate_systemd_unit": true,
"__podman_images_found": [],
"__podman_kube_yamls_raw": "",
"__podman_service_name": "quadlet-demo-mysql-volume.service",
"__podman_systemd_scope": "system",
"__podman_user_home_dir": "/root",
"__podman_xdg_runtime_dir": "/run/user/0"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73
Saturday 24 August 2024 12:36:24 -0400 (0:00:00.062) 0:02:33.025 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_quadlet_path": "/etc/containers/systemd"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Get kube yaml contents] ***************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77
Saturday 24 August 2024 12:36:24 -0400 (0:00:00.037) 0:02:33.062 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87
Saturday 24 August 2024 12:36:24 -0400 (0:00:00.033) 0:02:33.095 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_images": [],
"__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo-mysql.volume",
"__podman_volumes": []
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 6] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:103
Saturday 24 August 2024 12:36:24 -0400 (0:00:00.076) 0:02:33.172 *******
ok: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : Cleanup quadlets] *********************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:110
Saturday 24 August 2024 12:36:24 -0400 (0:00:00.039) 0:02:33.211 *******
included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml for managed_node1
TASK [fedora.linux_system_roles.podman : Stat XDG_RUNTIME_DIR] *****************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:4
Saturday 24 August 2024 12:36:24 -0400 (0:00:00.081) 0:02:33.293 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Stop and disable service] *************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:12
Saturday 24 August 2024 12:36:24 -0400 (0:00:00.034) 0:02:33.327 *******
changed: [managed_node1] => {
"changed": true,
"enabled": false,
"failed_when_result": false,
"name": "quadlet-demo-mysql-volume.service",
"state": "stopped",
"status": {
"ActiveEnterTimestamp": "Sat 2024-08-24 12:34:35 EDT",
"ActiveEnterTimestampMonotonic": "678725800",
"ActiveExitTimestampMonotonic": "0",
"ActiveState": "active",
"After": "basic.target -.mount sysinit.target systemd-journald.socket system.slice",
"AllowIsolate": "no",
"AllowedCPUs": "",
"AllowedMemoryNodes": "",
"AmbientCapabilities": "",
"AssertResult": "yes",
"AssertTimestamp": "Sat 2024-08-24 12:34:35 EDT",
"AssertTimestampMonotonic": "678677120",
"Before": "shutdown.target",
"BlockIOAccounting": "no",
"BlockIOWeight": "[not set]",
"CPUAccounting": "no",
"CPUAffinity": "",
"CPUAffinityFromNUMA": "no",
"CPUQuotaPerSecUSec": "infinity",
"CPUQuotaPeriodUSec": "infinity",
"CPUSchedulingPolicy": "0",
"CPUSchedulingPriority": "0",
"CPUSchedulingResetOnFork": "no",
"CPUShares": "[not set]",
"CPUUsageNSec": "[not set]",
"CPUWeight": "[not set]",
"CacheDirectoryMode": "0755",
"CanFreeze": "yes",
"CanIsolate": "no",
"CanReload": "no",
"CanStart": "yes",
"CanStop": "yes",
"CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf",
"CollectMode": "inactive",
"ConditionResult": "yes",
"ConditionTimestamp": "Sat 2024-08-24 12:34:35 EDT",
"ConditionTimestampMonotonic": "678677118",
"ConfigurationDirectoryMode": "0755",
"Conflicts": "shutdown.target",
"ControlGroup": "/system.slice/quadlet-demo-mysql-volume.service",
"ControlPID": "0",
"DefaultDependencies": "yes",
"DefaultMemoryLow": "0",
"DefaultMemoryMin": "0",
"Delegate": "no",
"Description": "quadlet-demo-mysql-volume.service",
"DevicePolicy": "auto",
"DynamicUser": "no",
"EffectiveCPUs": "",
"EffectiveMemoryNodes": "",
"ExecMainCode": "1",
"ExecMainExitTimestamp": "Sat 2024-08-24 12:34:35 EDT",
"ExecMainExitTimestampMonotonic": "678725428",
"ExecMainPID": "82689",
"ExecMainStartTimestamp": "Sat 2024-08-24 12:34:35 EDT",
"ExecMainStartTimestampMonotonic": "678678139",
"ExecMainStatus": "0",
"ExecStart": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman volume create --ignore systemd-quadlet-demo-mysql ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"FailureAction": "none",
"FileDescriptorStoreMax": "0",
"FragmentPath": "/run/systemd/generator/quadlet-demo-mysql-volume.service",
"FreezerState": "running",
"GID": "[not set]",
"GuessMainPID": "yes",
"IOAccounting": "no",
"IOSchedulingClass": "0",
"IOSchedulingPriority": "0",
"IOWeight": "[not set]",
"IPAccounting": "no",
"IPEgressBytes": "18446744073709551615",
"IPEgressPackets": "18446744073709551615",
"IPIngressBytes": "18446744073709551615",
"IPIngressPackets": "18446744073709551615",
"Id": "quadlet-demo-mysql-volume.service",
"IgnoreOnIsolate": "no",
"IgnoreSIGPIPE": "yes",
"InactiveEnterTimestampMonotonic": "0",
"InactiveExitTimestamp": "Sat 2024-08-24 12:34:35 EDT",
"InactiveExitTimestampMonotonic": "678678193",
"InvocationID": "4947f49535a04e7d84bd4de87f0ddea2",
"JobRunningTimeoutUSec": "infinity",
"JobTimeoutAction": "none",
"JobTimeoutUSec": "infinity",
"KeyringMode": "private",
"KillMode": "control-group",
"KillSignal": "15",
"LimitAS": "infinity",
"LimitASSoft": "infinity",
"LimitCORE": "infinity",
"LimitCORESoft": "0",
"LimitCPU": "infinity",
"LimitCPUSoft": "infinity",
"LimitDATA": "infinity",
"LimitDATASoft": "infinity",
"LimitFSIZE": "infinity",
"LimitFSIZESoft": "infinity",
"LimitLOCKS": "infinity",
"LimitLOCKSSoft": "infinity",
"LimitMEMLOCK": "65536",
"LimitMEMLOCKSoft": "65536",
"LimitMSGQUEUE": "819200",
"LimitMSGQUEUESoft": "819200",
"LimitNICE": "0",
"LimitNICESoft": "0",
"LimitNOFILE": "262144",
"LimitNOFILESoft": "1024",
"LimitNPROC": "14003",
"LimitNPROCSoft": "14003",
"LimitRSS": "infinity",
"LimitRSSSoft": "infinity",
"LimitRTPRIO": "0",
"LimitRTPRIOSoft": "0",
"LimitRTTIME": "infinity",
"LimitRTTIMESoft": "infinity",
"LimitSIGPENDING": "14003",
"LimitSIGPENDINGSoft": "14003",
"LimitSTACK": "infinity",
"LimitSTACKSoft": "8388608",
"LoadState": "loaded",
"LockPersonality": "no",
"LogLevelMax": "-1",
"LogRateLimitBurst": "0",
"LogRateLimitIntervalUSec": "0",
"LogsDirectoryMode": "0755",
"MainPID": "0",
"MemoryAccounting": "yes",
"MemoryCurrent": "0",
"MemoryDenyWriteExecute": "no",
"MemoryHigh": "infinity",
"MemoryLimit": "infinity",
"MemoryLow": "0",
"MemoryMax": "infinity",
"MemoryMin": "0",
"MemorySwapMax": "infinity",
"MountAPIVFS": "no",
"MountFlags": "",
"NFileDescriptorStore": "0",
"NRestarts": "0",
"NUMAMask": "",
"NUMAPolicy": "n/a",
"Names": "quadlet-demo-mysql-volume.service",
"NeedDaemonReload": "no",
"Nice": "0",
"NoNewPrivileges": "no",
"NonBlocking": "no",
"NotifyAccess": "none",
"OOMScoreAdjust": "0",
"OnFailureJobMode": "replace",
"PermissionsStartOnly": "no",
"Perpetual": "no",
"PrivateDevices": "no",
"PrivateMounts": "no",
"PrivateNetwork": "no",
"PrivateTmp": "no",
"PrivateUsers": "no",
"ProtectControlGroups": "no",
"ProtectHome": "no",
"ProtectKernelModules": "no",
"ProtectKernelTunables": "no",
"ProtectSystem": "no",
"RefuseManualStart": "no",
"RefuseManualStop": "no",
"RemainAfterExit": "yes",
"RemoveIPC": "no",
"Requires": "-.mount sysinit.target system.slice",
"RequiresMountsFor": "/run/containers",
"Restart": "no",
"RestartUSec": "100ms",
"RestrictNamespaces": "no",
"RestrictRealtime": "no",
"RestrictSUIDSGID": "no",
"Result": "success",
"RootDirectoryStartOnly": "no",
"RuntimeDirectoryMode": "0755",
"RuntimeDirectoryPreserve": "no",
"RuntimeMaxUSec": "infinity",
"SameProcessGroup": "no",
"SecureBits": "0",
"SendSIGHUP": "no",
"SendSIGKILL": "yes",
"Slice": "system.slice",
"StandardError": "inherit",
"StandardInput": "null",
"StandardInputData": "",
"StandardOutput": "journal",
"StartLimitAction": "none",
"StartLimitBurst": "5",
"StartLimitIntervalUSec": "10s",
"StartupBlockIOWeight": "[not set]",
"StartupCPUShares": "[not set]",
"StartupCPUWeight": "[not set]",
"StartupIOWeight": "[not set]",
"StateChangeTimestamp": "Sat 2024-08-24 12:34:35 EDT",
"StateChangeTimestampMonotonic": "678725800",
"StateDirectoryMode": "0755",
"StatusErrno": "0",
"StopWhenUnneeded": "no",
"SubState": "exited",
"SuccessAction": "none",
"SyslogFacility": "3",
"SyslogIdentifier": "quadlet-demo-mysql-volume",
"SyslogLevel": "6",
"SyslogLevelPrefix": "yes",
"SyslogPriority": "30",
"SystemCallErrorNumber": "0",
"TTYReset": "no",
"TTYVHangup": "no",
"TTYVTDisallocate": "no",
"TasksAccounting": "yes",
"TasksCurrent": "0",
"TasksMax": "22405",
"TimeoutStartUSec": "infinity",
"TimeoutStopUSec": "1min 30s",
"TimerSlackNSec": "50000",
"Transient": "no",
"Type": "oneshot",
"UID": "[not set]",
"UMask": "0022",
"UnitFilePreset": "disabled",
"UnitFileState": "generated",
"UtmpMode": "init",
"WatchdogTimestampMonotonic": "0",
"WatchdogUSec": "0"
}
}
TASK [fedora.linux_system_roles.podman : See if quadlet file exists] ***********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:33
Saturday 24 August 2024 12:36:25 -0400 (0:00:00.643) 0:02:33.970 *******
ok: [managed_node1] => {
"changed": false,
"stat": {
"atime": 1724517275.3136742,
"attr_flags": "",
"attributes": [],
"block_size": 4096,
"blocks": 8,
"charset": "us-ascii",
"checksum": "585f8cbdf0ec73000f9227dcffbef71e9552ea4a",
"ctime": 1724517274.5916681,
"dev": 51713,
"device_type": 0,
"executable": false,
"exists": true,
"gid": 0,
"gr_name": "root",
"inode": 213909769,
"isblk": false,
"ischr": false,
"isdir": false,
"isfifo": false,
"isgid": false,
"islnk": false,
"isreg": true,
"issock": false,
"isuid": false,
"mimetype": "text/plain",
"mode": "0644",
"mtime": 1724517274.2876654,
"nlink": 1,
"path": "/etc/containers/systemd/quadlet-demo-mysql.volume",
"pw_name": "root",
"readable": true,
"rgrp": true,
"roth": true,
"rusr": true,
"size": 9,
"uid": 0,
"version": "147345147",
"wgrp": false,
"woth": false,
"writeable": true,
"wusr": true,
"xgrp": false,
"xoth": false,
"xusr": false
}
}
TASK [fedora.linux_system_roles.podman : Parse quadlet file] *******************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:38
Saturday 24 August 2024 12:36:25 -0400 (0:00:00.374) 0:02:34.344 *******
included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml for managed_node1
TASK [fedora.linux_system_roles.podman : Slurp quadlet file] *******************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml:6
Saturday 24 August 2024 12:36:25 -0400 (0:00:00.061) 0:02:34.406 *******
ok: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : Parse quadlet file] *******************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml:12
Saturday 24 August 2024 12:36:26 -0400 (0:00:00.351) 0:02:34.758 *******
ok: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : Parse quadlet yaml file] **************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml:44
Saturday 24 August 2024 12:36:26 -0400 (0:00:00.050) 0:02:34.808 *******
skipping: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : Reset raw variable] *******************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml:52
Saturday 24 August 2024 12:36:26 -0400 (0:00:00.033) 0:02:34.841 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_quadlet_raw": null
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Remove quadlet file] ******************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:42
Saturday 24 August 2024 12:36:26 -0400 (0:00:00.033) 0:02:34.875 *******
changed: [managed_node1] => {
"changed": true,
"path": "/etc/containers/systemd/quadlet-demo-mysql.volume",
"state": "absent"
}
TASK [fedora.linux_system_roles.podman : Refresh systemd] **********************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:48
Saturday 24 August 2024 12:36:26 -0400 (0:00:00.363) 0:02:35.238 *******
ok: [managed_node1] => {
"changed": false,
"name": null,
"status": {}
}
TASK [fedora.linux_system_roles.podman : Remove managed resource] **************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:58
Saturday 24 August 2024 12:36:27 -0400 (0:00:00.592) 0:02:35.831 *******
changed: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": true
}
TASK [fedora.linux_system_roles.podman : Remove volumes] ***********************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:95
Saturday 24 August 2024 12:36:27 -0400 (0:00:00.430) 0:02:36.261 *******
skipping: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : Clear parsed podman variable] *********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:112
Saturday 24 August 2024 12:36:27 -0400 (0:00:00.050) 0:02:36.312 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_quadlet_parsed": null
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Prune images no longer in use] ********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:116
Saturday 24 August 2024 12:36:27 -0400 (0:00:00.033) 0:02:36.345 *******
changed: [managed_node1] => {
"changed": true,
"cmd": [
"podman",
"image",
"prune",
"--all",
"-f"
],
"delta": "0:00:00.033667",
"end": "2024-08-24 12:36:28.057630",
"rc": 0,
"start": "2024-08-24 12:36:28.023963"
}
TASK [fedora.linux_system_roles.podman : Manage linger] ************************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:127
Saturday 24 August 2024 12:36:28 -0400 (0:00:00.394) 0:02:36.739 *******
included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for managed_node1
TASK [fedora.linux_system_roles.podman : Enable linger if needed] **************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12
Saturday 24 August 2024 12:36:28 -0400 (0:00:00.059) 0:02:36.799 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18
Saturday 24 August 2024 12:36:28 -0400 (0:00:00.032) 0:02:36.831 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22
Saturday 24 August 2024 12:36:28 -0400 (0:00:00.033) 0:02:36.864 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : For testing and debugging - images] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:137
Saturday 24 August 2024 12:36:28 -0400 (0:00:00.032) 0:02:36.897 *******
ok: [managed_node1] => {
"changed": false,
"cmd": [
"podman",
"images",
"-n"
],
"delta": "0:00:00.033438",
"end": "2024-08-24 12:36:28.608598",
"rc": 0,
"start": "2024-08-24 12:36:28.575160"
}
TASK [fedora.linux_system_roles.podman : For testing and debugging - volumes] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:146
Saturday 24 August 2024 12:36:28 -0400 (0:00:00.394) 0:02:37.292 *******
ok: [managed_node1] => {
"changed": false,
"cmd": [
"podman",
"volume",
"ls",
"-n"
],
"delta": "0:00:00.035375",
"end": "2024-08-24 12:36:29.020418",
"rc": 0,
"start": "2024-08-24 12:36:28.985043"
}
STDOUT:
local wp-pv-claim
local envoy-proxy-config
local envoy-certificates
TASK [fedora.linux_system_roles.podman : For testing and debugging - containers] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:155
Saturday 24 August 2024 12:36:29 -0400 (0:00:00.412) 0:02:37.704 *******
ok: [managed_node1] => {
"changed": false,
"cmd": [
"podman",
"ps",
"--noheading"
],
"delta": "0:00:00.031622",
"end": "2024-08-24 12:36:29.419551",
"rc": 0,
"start": "2024-08-24 12:36:29.387929"
}
TASK [fedora.linux_system_roles.podman : For testing and debugging - networks] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:164
Saturday 24 August 2024 12:36:29 -0400 (0:00:00.398) 0:02:38.103 *******
ok: [managed_node1] => {
"changed": false,
"cmd": [
"podman",
"network",
"ls",
"-n",
"-q"
],
"delta": "0:00:00.058888",
"end": "2024-08-24 12:36:29.844427",
"rc": 0,
"start": "2024-08-24 12:36:29.785539"
}
STDOUT:
podman
podman-default-kube-network
systemd-quadlet-demo
TASK [fedora.linux_system_roles.podman : For testing and debugging - secrets] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:173
Saturday 24 August 2024 12:36:29 -0400 (0:00:00.424) 0:02:38.527 *******
ok: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : For testing and debugging - services] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:183
Saturday 24 August 2024 12:36:30 -0400 (0:00:00.397) 0:02:38.925 *******
ok: [managed_node1] => {
"ansible_facts": {
"services": {
"NetworkManager-dispatcher.service": {
"name": "NetworkManager-dispatcher.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"NetworkManager-wait-online.service": {
"name": "NetworkManager-wait-online.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"NetworkManager.service": {
"name": "NetworkManager.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"auditd.service": {
"name": "auditd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"auth-rpcgss-module.service": {
"name": "auth-rpcgss-module.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"autovt@.service": {
"name": "autovt@.service",
"source": "systemd",
"state": "unknown",
"status": "enabled"
},
"certmonger.service": {
"name": "certmonger.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"chrony-dnssrv@.service": {
"name": "chrony-dnssrv@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"chrony-wait.service": {
"name": "chrony-wait.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"chronyd.service": {
"name": "chronyd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"cloud-config.service": {
"name": "cloud-config.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"cloud-final.service": {
"name": "cloud-final.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"cloud-init-hotplugd.service": {
"name": "cloud-init-hotplugd.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"cloud-init-local.service": {
"name": "cloud-init-local.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"cloud-init.service": {
"name": "cloud-init.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"cni-dhcp.service": {
"name": "cni-dhcp.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"console-getty.service": {
"name": "console-getty.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"container-getty@.service": {
"name": "container-getty@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"cpupower.service": {
"name": "cpupower.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"crond.service": {
"name": "crond.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"dbus-org.fedoraproject.FirewallD1.service": {
"name": "dbus-org.fedoraproject.FirewallD1.service",
"source": "systemd",
"state": "active",
"status": "enabled"
},
"dbus-org.freedesktop.hostname1.service": {
"name": "dbus-org.freedesktop.hostname1.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dbus-org.freedesktop.locale1.service": {
"name": "dbus-org.freedesktop.locale1.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dbus-org.freedesktop.login1.service": {
"name": "dbus-org.freedesktop.login1.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dbus-org.freedesktop.nm-dispatcher.service": {
"name": "dbus-org.freedesktop.nm-dispatcher.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"dbus-org.freedesktop.portable1.service": {
"name": "dbus-org.freedesktop.portable1.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dbus-org.freedesktop.timedate1.service": {
"name": "dbus-org.freedesktop.timedate1.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"dbus.service": {
"name": "dbus.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"debug-shell.service": {
"name": "debug-shell.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"dnf-makecache.service": {
"name": "dnf-makecache.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dnf-system-upgrade-cleanup.service": {
"name": "dnf-system-upgrade-cleanup.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dnf-system-upgrade.service": {
"name": "dnf-system-upgrade.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"dnsmasq.service": {
"name": "dnsmasq.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"dracut-cmdline.service": {
"name": "dracut-cmdline.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-initqueue.service": {
"name": "dracut-initqueue.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-mount.service": {
"name": "dracut-mount.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-mount.service": {
"name": "dracut-pre-mount.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-pivot.service": {
"name": "dracut-pre-pivot.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-trigger.service": {
"name": "dracut-pre-trigger.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-udev.service": {
"name": "dracut-pre-udev.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-shutdown-onfailure.service": {
"name": "dracut-shutdown-onfailure.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-shutdown.service": {
"name": "dracut-shutdown.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"ebtables.service": {
"name": "ebtables.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"emergency.service": {
"name": "emergency.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"firewalld.service": {
"name": "firewalld.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"fstrim.service": {
"name": "fstrim.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"getty@.service": {
"name": "getty@.service",
"source": "systemd",
"state": "unknown",
"status": "enabled"
},
"getty@tty1.service": {
"name": "getty@tty1.service",
"source": "systemd",
"state": "running",
"status": "unknown"
},
"grub-boot-indeterminate.service": {
"name": "grub-boot-indeterminate.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"gssproxy.service": {
"name": "gssproxy.service",
"source": "systemd",
"state": "running",
"status": "disabled"
},
"halt-local.service": {
"name": "halt-local.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"import-state.service": {
"name": "import-state.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"initrd-cleanup.service": {
"name": "initrd-cleanup.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"initrd-parse-etc.service": {
"name": "initrd-parse-etc.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"initrd-switch-root.service": {
"name": "initrd-switch-root.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"initrd-udevadm-cleanup-db.service": {
"name": "initrd-udevadm-cleanup-db.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"iprdump.service": {
"name": "iprdump.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"iprinit.service": {
"name": "iprinit.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"iprupdate.service": {
"name": "iprupdate.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"irqbalance.service": {
"name": "irqbalance.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"kdump.service": {
"name": "kdump.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"kmod-static-nodes.service": {
"name": "kmod-static-nodes.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"kvm_stat.service": {
"name": "kvm_stat.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"ldconfig.service": {
"name": "ldconfig.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"loadmodules.service": {
"name": "loadmodules.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"man-db-cache-update.service": {
"name": "man-db-cache-update.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"man-db-restart-cache-update.service": {
"name": "man-db-restart-cache-update.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"messagebus.service": {
"name": "messagebus.service",
"source": "systemd",
"state": "active",
"status": "static"
},
"microcode.service": {
"name": "microcode.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"nfs-blkmap.service": {
"name": "nfs-blkmap.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"nfs-convert.service": {
"name": "nfs-convert.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"nfs-idmapd.service": {
"name": "nfs-idmapd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nfs-mountd.service": {
"name": "nfs-mountd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nfs-server.service": {
"name": "nfs-server.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"nfs-utils.service": {
"name": "nfs-utils.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nfsdcld.service": {
"name": "nfsdcld.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nftables.service": {
"name": "nftables.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"nis-domainname.service": {
"name": "nis-domainname.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"oddjobd.service": {
"name": "oddjobd.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"plymouth-halt.service": {
"name": "plymouth-halt.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-kexec.service": {
"name": "plymouth-kexec.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-poweroff.service": {
"name": "plymouth-poweroff.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-quit-wait.service": {
"name": "plymouth-quit-wait.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-quit.service": {
"name": "plymouth-quit.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-read-write.service": {
"name": "plymouth-read-write.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-reboot.service": {
"name": "plymouth-reboot.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-start.service": {
"name": "plymouth-start.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-switch-root-initramfs.service": {
"name": "plymouth-switch-root-initramfs.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-switch-root.service": {
"name": "plymouth-switch-root.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"podman-auto-update.service": {
"name": "podman-auto-update.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"podman-clean-transient.service": {
"name": "podman-clean-transient.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"podman-kube@.service": {
"name": "podman-kube@.service",
"source": "systemd",
"state": "unknown",
"status": "disabled"
},
"podman-restart.service": {
"name": "podman-restart.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"podman.service": {
"name": "podman.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"polkit.service": {
"name": "polkit.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"qemu-guest-agent.service": {
"name": "qemu-guest-agent.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"quadlet-demo-network.service": {
"name": "quadlet-demo-network.service",
"source": "systemd",
"state": "stopped",
"status": "generated"
},
"quotaon.service": {
"name": "quotaon.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"rc-local.service": {
"name": "rc-local.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rdisc.service": {
"name": "rdisc.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"rescue.service": {
"name": "rescue.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"restraintd.service": {
"name": "restraintd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"rngd.service": {
"name": "rngd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"rpc-gssd.service": {
"name": "rpc-gssd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rpc-statd-notify.service": {
"name": "rpc-statd-notify.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rpc-statd.service": {
"name": "rpc-statd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rpcbind.service": {
"name": "rpcbind.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"rsyslog.service": {
"name": "rsyslog.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"selinux-autorelabel-mark.service": {
"name": "selinux-autorelabel-mark.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"selinux-autorelabel.service": {
"name": "selinux-autorelabel.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"serial-getty@.service": {
"name": "serial-getty@.service",
"source": "systemd",
"state": "unknown",
"status": "disabled"
},
"sshd-keygen@.service": {
"name": "sshd-keygen@.service",
"source": "systemd",
"state": "unknown",
"status": "disabled"
},
"sshd-keygen@ecdsa.service": {
"name": "sshd-keygen@ecdsa.service",
"source": "systemd",
"state": "stopped",
"status": "unknown"
},
"sshd-keygen@ed25519.service": {
"name": "sshd-keygen@ed25519.service",
"source": "systemd",
"state": "stopped",
"status": "unknown"
},
"sshd-keygen@rsa.service": {
"name": "sshd-keygen@rsa.service",
"source": "systemd",
"state": "stopped",
"status": "unknown"
},
"sshd.service": {
"name": "sshd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"sshd@.service": {
"name": "sshd@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"sssd-autofs.service": {
"name": "sssd-autofs.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-kcm.service": {
"name": "sssd-kcm.service",
"source": "systemd",
"state": "stopped",
"status": "indirect"
},
"sssd-nss.service": {
"name": "sssd-nss.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-pac.service": {
"name": "sssd-pac.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-pam.service": {
"name": "sssd-pam.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-ssh.service": {
"name": "sssd-ssh.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-sudo.service": {
"name": "sssd-sudo.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd.service": {
"name": "sssd.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"syslog.service": {
"name": "syslog.service",
"source": "systemd",
"state": "active",
"status": "enabled"
},
"system-update-cleanup.service": {
"name": "system-update-cleanup.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-ask-password-console.service": {
"name": "systemd-ask-password-console.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-ask-password-plymouth.service": {
"name": "systemd-ask-password-plymouth.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-ask-password-wall.service": {
"name": "systemd-ask-password-wall.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-backlight@.service": {
"name": "systemd-backlight@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-binfmt.service": {
"name": "systemd-binfmt.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-coredump@.service": {
"name": "systemd-coredump@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-exit.service": {
"name": "systemd-exit.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-firstboot.service": {
"name": "systemd-firstboot.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-fsck-root.service": {
"name": "systemd-fsck-root.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-fsck@.service": {
"name": "systemd-fsck@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-halt.service": {
"name": "systemd-halt.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-hibernate-resume@.service": {
"name": "systemd-hibernate-resume@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-hibernate.service": {
"name": "systemd-hibernate.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-hostnamed.service": {
"name": "systemd-hostnamed.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-hwdb-update.service": {
"name": "systemd-hwdb-update.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-hybrid-sleep.service": {
"name": "systemd-hybrid-sleep.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-initctl.service": {
"name": "systemd-initctl.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-journal-catalog-update.service": {
"name": "systemd-journal-catalog-update.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-journal-flush.service": {
"name": "systemd-journal-flush.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-journald.service": {
"name": "systemd-journald.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"systemd-kexec.service": {
"name": "systemd-kexec.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-localed.service": {
"name": "systemd-localed.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-logind.service": {
"name": "systemd-logind.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-machine-id-commit.service": {
"name": "systemd-machine-id-commit.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-modules-load.service": {
"name": "systemd-modules-load.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-portabled.service": {
"name": "systemd-portabled.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-poweroff.service": {
"name": "systemd-poweroff.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-pstore.service": {
"name": "systemd-pstore.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"systemd-quotacheck.service": {
"name": "systemd-quotacheck.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-random-seed.service": {
"name": "systemd-random-seed.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-reboot.service": {
"name": "systemd-reboot.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-remount-fs.service": {
"name": "systemd-remount-fs.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-resolved.service": {
"name": "systemd-resolved.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"systemd-rfkill.service": {
"name": "systemd-rfkill.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-suspend-then-hibernate.service": {
"name": "systemd-suspend-then-hibernate.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-suspend.service": {
"name": "systemd-suspend.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-sysctl.service": {
"name": "systemd-sysctl.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-sysusers.service": {
"name": "systemd-sysusers.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-timedated.service": {
"name": "systemd-timedated.service",
"source": "systemd",
"state": "inactive",
"status": "masked"
},
"systemd-tmpfiles-clean.service": {
"name": "systemd-tmpfiles-clean.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-tmpfiles-setup-dev.service": {
"name": "systemd-tmpfiles-setup-dev.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-tmpfiles-setup.service": {
"name": "systemd-tmpfiles-setup.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-udev-settle.service": {
"name": "systemd-udev-settle.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-udev-trigger.service": {
"name": "systemd-udev-trigger.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-udevd.service": {
"name": "systemd-udevd.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"systemd-update-done.service": {
"name": "systemd-update-done.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-update-utmp-runlevel.service": {
"name": "systemd-update-utmp-runlevel.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-update-utmp.service": {
"name": "systemd-update-utmp.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-user-sessions.service": {
"name": "systemd-user-sessions.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-vconsole-setup.service": {
"name": "systemd-vconsole-setup.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-volatile-root.service": {
"name": "systemd-volatile-root.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"tcsd.service": {
"name": "tcsd.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"teamd@.service": {
"name": "teamd@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"timedatex.service": {
"name": "timedatex.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"tuned.service": {
"name": "tuned.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"unbound-anchor.service": {
"name": "unbound-anchor.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"user-runtime-dir@.service": {
"name": "user-runtime-dir@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"user-runtime-dir@0.service": {
"name": "user-runtime-dir@0.service",
"source": "systemd",
"state": "stopped",
"status": "unknown"
},
"user@.service": {
"name": "user@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"user@0.service": {
"name": "user@0.service",
"source": "systemd",
"state": "running",
"status": "unknown"
}
}
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Create and update quadlets] ***********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:114
Saturday 24 August 2024 12:36:31 -0400 (0:00:01.613) 0:02:40.538 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14
Saturday 24 August 2024 12:36:31 -0400 (0:00:00.065) 0:02:40.603 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_quadlet_file_src": "quadlet-demo.network",
"__podman_quadlet_spec": {},
"__podman_quadlet_str": "[Network]\nSubnet=192.168.30.0/24\nGateway=192.168.30.1\nLabel=app=wordpress",
"__podman_quadlet_template_src": ""
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25
Saturday 24 August 2024 12:36:32 -0400 (0:00:00.045) 0:02:40.649 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_continue_if_pull_fails": false,
"__podman_pull_image": true,
"__podman_state": "absent",
"__podman_systemd_unit_scope": "",
"__podman_user": "root"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] *****
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35
Saturday 24 August 2024 12:36:32 -0400 (0:00:00.039) 0:02:40.688 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48
Saturday 24 August 2024 12:36:32 -0400 (0:00:00.034) 0:02:40.723 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_quadlet_name": "quadlet-demo",
"__podman_quadlet_type": "network",
"__podman_rootless": false
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Check user and group information] *****
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57
Saturday 24 August 2024 12:36:32 -0400 (0:00:00.046) 0:02:40.769 *******
included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for managed_node1
TASK [fedora.linux_system_roles.podman : Get user information] *****************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2
Saturday 24 August 2024 12:36:32 -0400 (0:00:00.062) 0:02:40.832 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Fail if user does not exist] **********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9
Saturday 24 August 2024 12:36:32 -0400 (0:00:00.041) 0:02:40.874 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set group for podman user] ************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16
Saturday 24 August 2024 12:36:32 -0400 (0:00:00.040) 0:02:40.914 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_group": "0"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Get group information] ****************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:28
Saturday 24 August 2024 12:36:32 -0400 (0:00:00.043) 0:02:40.958 *******
ok: [managed_node1] => {
"ansible_facts": {
"getent_group": {
"root": [
"x",
"0",
""
]
}
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set group name] ***********************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:35
Saturday 24 August 2024 12:36:32 -0400 (0:00:00.370) 0:02:41.328 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_group_name": "root"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : See if getsubids exists] **************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:39
Saturday 24 August 2024 12:36:32 -0400 (0:00:00.044) 0:02:41.373 *******
ok: [managed_node1] => {
"changed": false,
"stat": {
"atime": 1724516929.7168462,
"attr_flags": "",
"attributes": [],
"block_size": 4096,
"blocks": 32,
"charset": "binary",
"checksum": "bb5b46ffbafcaa8c4021f3c8b3cb8594f48ef34b",
"ctime": 1724516896.725572,
"dev": 51713,
"device_type": 0,
"executable": true,
"exists": true,
"gid": 0,
"gr_name": "root",
"inode": 6986657,
"isblk": false,
"ischr": false,
"isdir": false,
"isfifo": false,
"isgid": false,
"islnk": false,
"isreg": true,
"issock": false,
"isuid": false,
"mimetype": "application/x-sharedlib",
"mode": "0755",
"mtime": 1700557386.0,
"nlink": 1,
"path": "/usr/bin/getsubids",
"pw_name": "root",
"readable": true,
"rgrp": true,
"roth": true,
"rusr": true,
"size": 12640,
"uid": 0,
"version": "1255679238",
"wgrp": false,
"woth": false,
"writeable": true,
"wusr": true,
"xgrp": true,
"xoth": true,
"xusr": true
}
}
TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:50
Saturday 24 August 2024 12:36:33 -0400 (0:00:00.358) 0:02:41.731 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Check group with getsubids] ***********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:55
Saturday 24 August 2024 12:36:33 -0400 (0:00:00.074) 0:02:41.806 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ******
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:60
Saturday 24 August 2024 12:36:33 -0400 (0:00:00.033) 0:02:41.839 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Get subuid file] **********************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:74
Saturday 24 August 2024 12:36:33 -0400 (0:00:00.034) 0:02:41.873 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Get subgid file] **********************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:79
Saturday 24 August 2024 12:36:33 -0400 (0:00:00.033) 0:02:41.907 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ******
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:84
Saturday 24 August 2024 12:36:33 -0400 (0:00:00.033) 0:02:41.940 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ******
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:94
Saturday 24 August 2024 12:36:33 -0400 (0:00:00.033) 0:02:41.974 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] *****
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:101
Saturday 24 August 2024 12:36:33 -0400 (0:00:00.032) 0:02:42.007 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62
Saturday 24 August 2024 12:36:33 -0400 (0:00:00.031) 0:02:42.039 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_activate_systemd_unit": true,
"__podman_images_found": [],
"__podman_kube_yamls_raw": "",
"__podman_service_name": "quadlet-demo-network.service",
"__podman_systemd_scope": "system",
"__podman_user_home_dir": "/root",
"__podman_xdg_runtime_dir": "/run/user/0"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73
Saturday 24 August 2024 12:36:33 -0400 (0:00:00.057) 0:02:42.096 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_quadlet_path": "/etc/containers/systemd"
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Get kube yaml contents] ***************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77
Saturday 24 August 2024 12:36:33 -0400 (0:00:00.034) 0:02:42.130 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87
Saturday 24 August 2024 12:36:33 -0400 (0:00:00.033) 0:02:42.163 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_images": [],
"__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo.network",
"__podman_volumes": []
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Set per-container variables part 6] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:103
Saturday 24 August 2024 12:36:33 -0400 (0:00:00.070) 0:02:42.234 *******
ok: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : Cleanup quadlets] *********************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:110
Saturday 24 August 2024 12:36:33 -0400 (0:00:00.038) 0:02:42.273 *******
included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml for managed_node1
TASK [fedora.linux_system_roles.podman : Stat XDG_RUNTIME_DIR] *****************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:4
Saturday 24 August 2024 12:36:33 -0400 (0:00:00.080) 0:02:42.353 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Stop and disable service] *************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:12
Saturday 24 August 2024 12:36:33 -0400 (0:00:00.032) 0:02:42.386 *******
changed: [managed_node1] => {
"changed": true,
"enabled": false,
"failed_when_result": false,
"name": "quadlet-demo-network.service",
"state": "stopped",
"status": {
"ActiveEnterTimestamp": "Sat 2024-08-24 12:34:30 EDT",
"ActiveEnterTimestampMonotonic": "673375366",
"ActiveExitTimestampMonotonic": "0",
"ActiveState": "active",
"After": "sysinit.target systemd-journald.socket system.slice basic.target -.mount",
"AllowIsolate": "no",
"AllowedCPUs": "",
"AllowedMemoryNodes": "",
"AmbientCapabilities": "",
"AssertResult": "yes",
"AssertTimestamp": "Sat 2024-08-24 12:34:30 EDT",
"AssertTimestampMonotonic": "673320204",
"Before": "shutdown.target",
"BlockIOAccounting": "no",
"BlockIOWeight": "[not set]",
"CPUAccounting": "no",
"CPUAffinity": "",
"CPUAffinityFromNUMA": "no",
"CPUQuotaPerSecUSec": "infinity",
"CPUQuotaPeriodUSec": "infinity",
"CPUSchedulingPolicy": "0",
"CPUSchedulingPriority": "0",
"CPUSchedulingResetOnFork": "no",
"CPUShares": "[not set]",
"CPUUsageNSec": "[not set]",
"CPUWeight": "[not set]",
"CacheDirectoryMode": "0755",
"CanFreeze": "yes",
"CanIsolate": "no",
"CanReload": "no",
"CanStart": "yes",
"CanStop": "yes",
"CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf",
"CollectMode": "inactive",
"ConditionResult": "yes",
"ConditionTimestamp": "Sat 2024-08-24 12:34:30 EDT",
"ConditionTimestampMonotonic": "673320203",
"ConfigurationDirectoryMode": "0755",
"Conflicts": "shutdown.target",
"ControlGroup": "/system.slice/quadlet-demo-network.service",
"ControlPID": "0",
"DefaultDependencies": "yes",
"DefaultMemoryLow": "0",
"DefaultMemoryMin": "0",
"Delegate": "no",
"Description": "quadlet-demo-network.service",
"DevicePolicy": "auto",
"DynamicUser": "no",
"EffectiveCPUs": "",
"EffectiveMemoryNodes": "",
"ExecMainCode": "1",
"ExecMainExitTimestamp": "Sat 2024-08-24 12:34:30 EDT",
"ExecMainExitTimestampMonotonic": "673374927",
"ExecMainPID": "81779",
"ExecMainStartTimestamp": "Sat 2024-08-24 12:34:30 EDT",
"ExecMainStartTimestampMonotonic": "673321150",
"ExecMainStatus": "0",
"ExecStart": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman network create --ignore --subnet=192.168.30.0/24 --gateway=192.168.30.1 --label app=wordpress systemd-quadlet-demo ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"FailureAction": "none",
"FileDescriptorStoreMax": "0",
"FragmentPath": "/run/systemd/generator/quadlet-demo-network.service",
"FreezerState": "running",
"GID": "[not set]",
"GuessMainPID": "yes",
"IOAccounting": "no",
"IOSchedulingClass": "0",
"IOSchedulingPriority": "0",
"IOWeight": "[not set]",
"IPAccounting": "no",
"IPEgressBytes": "18446744073709551615",
"IPEgressPackets": "18446744073709551615",
"IPIngressBytes": "18446744073709551615",
"IPIngressPackets": "18446744073709551615",
"Id": "quadlet-demo-network.service",
"IgnoreOnIsolate": "no",
"IgnoreSIGPIPE": "yes",
"InactiveEnterTimestampMonotonic": "0",
"InactiveExitTimestamp": "Sat 2024-08-24 12:34:30 EDT",
"InactiveExitTimestampMonotonic": "673321190",
"InvocationID": "124aa2ffb79e4600a823e33d270d8a42",
"JobRunningTimeoutUSec": "infinity",
"JobTimeoutAction": "none",
"JobTimeoutUSec": "infinity",
"KeyringMode": "private",
"KillMode": "control-group",
"KillSignal": "15",
"LimitAS": "infinity",
"LimitASSoft": "infinity",
"LimitCORE": "infinity",
"LimitCORESoft": "0",
"LimitCPU": "infinity",
"LimitCPUSoft": "infinity",
"LimitDATA": "infinity",
"LimitDATASoft": "infinity",
"LimitFSIZE": "infinity",
"LimitFSIZESoft": "infinity",
"LimitLOCKS": "infinity",
"LimitLOCKSSoft": "infinity",
"LimitMEMLOCK": "65536",
"LimitMEMLOCKSoft": "65536",
"LimitMSGQUEUE": "819200",
"LimitMSGQUEUESoft": "819200",
"LimitNICE": "0",
"LimitNICESoft": "0",
"LimitNOFILE": "262144",
"LimitNOFILESoft": "1024",
"LimitNPROC": "14003",
"LimitNPROCSoft": "14003",
"LimitRSS": "infinity",
"LimitRSSSoft": "infinity",
"LimitRTPRIO": "0",
"LimitRTPRIOSoft": "0",
"LimitRTTIME": "infinity",
"LimitRTTIMESoft": "infinity",
"LimitSIGPENDING": "14003",
"LimitSIGPENDINGSoft": "14003",
"LimitSTACK": "infinity",
"LimitSTACKSoft": "8388608",
"LoadState": "loaded",
"LockPersonality": "no",
"LogLevelMax": "-1",
"LogRateLimitBurst": "0",
"LogRateLimitIntervalUSec": "0",
"LogsDirectoryMode": "0755",
"MainPID": "0",
"MemoryAccounting": "yes",
"MemoryCurrent": "0",
"MemoryDenyWriteExecute": "no",
"MemoryHigh": "infinity",
"MemoryLimit": "infinity",
"MemoryLow": "0",
"MemoryMax": "infinity",
"MemoryMin": "0",
"MemorySwapMax": "infinity",
"MountAPIVFS": "no",
"MountFlags": "",
"NFileDescriptorStore": "0",
"NRestarts": "0",
"NUMAMask": "",
"NUMAPolicy": "n/a",
"Names": "quadlet-demo-network.service",
"NeedDaemonReload": "no",
"Nice": "0",
"NoNewPrivileges": "no",
"NonBlocking": "no",
"NotifyAccess": "none",
"OOMScoreAdjust": "0",
"OnFailureJobMode": "replace",
"PermissionsStartOnly": "no",
"Perpetual": "no",
"PrivateDevices": "no",
"PrivateMounts": "no",
"PrivateNetwork": "no",
"PrivateTmp": "no",
"PrivateUsers": "no",
"ProtectControlGroups": "no",
"ProtectHome": "no",
"ProtectKernelModules": "no",
"ProtectKernelTunables": "no",
"ProtectSystem": "no",
"RefuseManualStart": "no",
"RefuseManualStop": "no",
"RemainAfterExit": "yes",
"RemoveIPC": "no",
"Requires": "sysinit.target system.slice -.mount",
"RequiresMountsFor": "/run/containers",
"Restart": "no",
"RestartUSec": "100ms",
"RestrictNamespaces": "no",
"RestrictRealtime": "no",
"RestrictSUIDSGID": "no",
"Result": "success",
"RootDirectoryStartOnly": "no",
"RuntimeDirectoryMode": "0755",
"RuntimeDirectoryPreserve": "no",
"RuntimeMaxUSec": "infinity",
"SameProcessGroup": "no",
"SecureBits": "0",
"SendSIGHUP": "no",
"SendSIGKILL": "yes",
"Slice": "system.slice",
"StandardError": "inherit",
"StandardInput": "null",
"StandardInputData": "",
"StandardOutput": "journal",
"StartLimitAction": "none",
"StartLimitBurst": "5",
"StartLimitIntervalUSec": "10s",
"StartupBlockIOWeight": "[not set]",
"StartupCPUShares": "[not set]",
"StartupCPUWeight": "[not set]",
"StartupIOWeight": "[not set]",
"StateChangeTimestamp": "Sat 2024-08-24 12:34:30 EDT",
"StateChangeTimestampMonotonic": "673375366",
"StateDirectoryMode": "0755",
"StatusErrno": "0",
"StopWhenUnneeded": "no",
"SubState": "exited",
"SuccessAction": "none",
"SyslogFacility": "3",
"SyslogIdentifier": "quadlet-demo-network",
"SyslogLevel": "6",
"SyslogLevelPrefix": "yes",
"SyslogPriority": "30",
"SystemCallErrorNumber": "0",
"TTYReset": "no",
"TTYVHangup": "no",
"TTYVTDisallocate": "no",
"TasksAccounting": "yes",
"TasksCurrent": "0",
"TasksMax": "22405",
"TimeoutStartUSec": "infinity",
"TimeoutStopUSec": "1min 30s",
"TimerSlackNSec": "50000",
"Transient": "no",
"Type": "oneshot",
"UID": "[not set]",
"UMask": "0022",
"UnitFilePreset": "disabled",
"UnitFileState": "generated",
"UtmpMode": "init",
"WatchdogTimestampMonotonic": "0",
"WatchdogUSec": "0"
}
}
TASK [fedora.linux_system_roles.podman : See if quadlet file exists] ***********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:33
Saturday 24 August 2024 12:36:34 -0400 (0:00:00.629) 0:02:43.015 *******
ok: [managed_node1] => {
"changed": false,
"stat": {
"atime": 1724517269.990629,
"attr_flags": "",
"attributes": [],
"block_size": 4096,
"blocks": 8,
"charset": "us-ascii",
"checksum": "e57c08d49aff4bae8daab138d913aeddaa8682a0",
"ctime": 1724517269.3686237,
"dev": 51713,
"device_type": 0,
"executable": false,
"exists": true,
"gid": 0,
"gr_name": "root",
"inode": 186646789,
"isblk": false,
"ischr": false,
"isdir": false,
"isfifo": false,
"isgid": false,
"islnk": false,
"isreg": true,
"issock": false,
"isuid": false,
"mimetype": "text/plain",
"mode": "0644",
"mtime": 1724517269.029621,
"nlink": 1,
"path": "/etc/containers/systemd/quadlet-demo.network",
"pw_name": "root",
"readable": true,
"rgrp": true,
"roth": true,
"rusr": true,
"size": 74,
"uid": 0,
"version": "2153314366",
"wgrp": false,
"woth": false,
"writeable": true,
"wusr": true,
"xgrp": false,
"xoth": false,
"xusr": false
}
}
TASK [fedora.linux_system_roles.podman : Parse quadlet file] *******************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:38
Saturday 24 August 2024 12:36:34 -0400 (0:00:00.369) 0:02:43.385 *******
included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml for managed_node1
TASK [fedora.linux_system_roles.podman : Slurp quadlet file] *******************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml:6
Saturday 24 August 2024 12:36:34 -0400 (0:00:00.062) 0:02:43.447 *******
ok: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : Parse quadlet file] *******************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml:12
Saturday 24 August 2024 12:36:35 -0400 (0:00:00.353) 0:02:43.801 *******
ok: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : Parse quadlet yaml file] **************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml:44
Saturday 24 August 2024 12:36:35 -0400 (0:00:00.047) 0:02:43.849 *******
skipping: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : Reset raw variable] *******************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml:52
Saturday 24 August 2024 12:36:35 -0400 (0:00:00.032) 0:02:43.881 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_quadlet_raw": null
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Remove quadlet file] ******************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:42
Saturday 24 August 2024 12:36:35 -0400 (0:00:00.031) 0:02:43.913 *******
changed: [managed_node1] => {
"changed": true,
"path": "/etc/containers/systemd/quadlet-demo.network",
"state": "absent"
}
TASK [fedora.linux_system_roles.podman : Refresh systemd] **********************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:48
Saturday 24 August 2024 12:36:35 -0400 (0:00:00.357) 0:02:44.270 *******
ok: [managed_node1] => {
"changed": false,
"name": null,
"status": {}
}
TASK [fedora.linux_system_roles.podman : Remove managed resource] **************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:58
Saturday 24 August 2024 12:36:36 -0400 (0:00:00.588) 0:02:44.858 *******
changed: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": true
}
TASK [fedora.linux_system_roles.podman : Remove volumes] ***********************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:95
Saturday 24 August 2024 12:36:36 -0400 (0:00:00.458) 0:02:45.316 *******
skipping: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : Clear parsed podman variable] *********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:112
Saturday 24 August 2024 12:36:36 -0400 (0:00:00.050) 0:02:45.367 *******
ok: [managed_node1] => {
"ansible_facts": {
"__podman_quadlet_parsed": null
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Prune images no longer in use] ********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:116
Saturday 24 August 2024 12:36:36 -0400 (0:00:00.032) 0:02:45.400 *******
changed: [managed_node1] => {
"changed": true,
"cmd": [
"podman",
"image",
"prune",
"--all",
"-f"
],
"delta": "0:00:00.037303",
"end": "2024-08-24 12:36:37.138650",
"rc": 0,
"start": "2024-08-24 12:36:37.101347"
}
TASK [fedora.linux_system_roles.podman : Manage linger] ************************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:127
Saturday 24 August 2024 12:36:37 -0400 (0:00:00.421) 0:02:45.822 *******
included: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for managed_node1
TASK [fedora.linux_system_roles.podman : Enable linger if needed] **************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12
Saturday 24 August 2024 12:36:37 -0400 (0:00:00.060) 0:02:45.883 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18
Saturday 24 August 2024 12:36:37 -0400 (0:00:00.034) 0:02:45.917 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22
Saturday 24 August 2024 12:36:37 -0400 (0:00:00.033) 0:02:45.951 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : For testing and debugging - images] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:137
Saturday 24 August 2024 12:36:37 -0400 (0:00:00.033) 0:02:45.985 *******
ok: [managed_node1] => {
"changed": false,
"cmd": [
"podman",
"images",
"-n"
],
"delta": "0:00:00.035636",
"end": "2024-08-24 12:36:37.703710",
"rc": 0,
"start": "2024-08-24 12:36:37.668074"
}
TASK [fedora.linux_system_roles.podman : For testing and debugging - volumes] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:146
Saturday 24 August 2024 12:36:37 -0400 (0:00:00.448) 0:02:46.433 *******
ok: [managed_node1] => {
"changed": false,
"cmd": [
"podman",
"volume",
"ls",
"-n"
],
"delta": "0:00:00.034941",
"end": "2024-08-24 12:36:38.153593",
"rc": 0,
"start": "2024-08-24 12:36:38.118652"
}
STDOUT:
local wp-pv-claim
local envoy-proxy-config
local envoy-certificates
TASK [fedora.linux_system_roles.podman : For testing and debugging - containers] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:155
Saturday 24 August 2024 12:36:38 -0400 (0:00:00.403) 0:02:46.836 *******
ok: [managed_node1] => {
"changed": false,
"cmd": [
"podman",
"ps",
"--noheading"
],
"delta": "0:00:00.032042",
"end": "2024-08-24 12:36:38.553555",
"rc": 0,
"start": "2024-08-24 12:36:38.521513"
}
TASK [fedora.linux_system_roles.podman : For testing and debugging - networks] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:164
Saturday 24 August 2024 12:36:38 -0400 (0:00:00.399) 0:02:47.236 *******
ok: [managed_node1] => {
"changed": false,
"cmd": [
"podman",
"network",
"ls",
"-n",
"-q"
],
"delta": "0:00:00.051693",
"end": "2024-08-24 12:36:38.969610",
"rc": 0,
"start": "2024-08-24 12:36:38.917917"
}
STDOUT:
podman
podman-default-kube-network
TASK [fedora.linux_system_roles.podman : For testing and debugging - secrets] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:173
Saturday 24 August 2024 12:36:39 -0400 (0:00:00.417) 0:02:47.653 *******
ok: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : For testing and debugging - services] ***
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:183
Saturday 24 August 2024 12:36:39 -0400 (0:00:00.396) 0:02:48.049 *******
ok: [managed_node1] => {
"ansible_facts": {
"services": {
"NetworkManager-dispatcher.service": {
"name": "NetworkManager-dispatcher.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"NetworkManager-wait-online.service": {
"name": "NetworkManager-wait-online.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"NetworkManager.service": {
"name": "NetworkManager.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"auditd.service": {
"name": "auditd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"auth-rpcgss-module.service": {
"name": "auth-rpcgss-module.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"autovt@.service": {
"name": "autovt@.service",
"source": "systemd",
"state": "unknown",
"status": "enabled"
},
"certmonger.service": {
"name": "certmonger.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"chrony-dnssrv@.service": {
"name": "chrony-dnssrv@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"chrony-wait.service": {
"name": "chrony-wait.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"chronyd.service": {
"name": "chronyd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"cloud-config.service": {
"name": "cloud-config.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"cloud-final.service": {
"name": "cloud-final.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"cloud-init-hotplugd.service": {
"name": "cloud-init-hotplugd.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"cloud-init-local.service": {
"name": "cloud-init-local.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"cloud-init.service": {
"name": "cloud-init.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"cni-dhcp.service": {
"name": "cni-dhcp.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"console-getty.service": {
"name": "console-getty.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"container-getty@.service": {
"name": "container-getty@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"cpupower.service": {
"name": "cpupower.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"crond.service": {
"name": "crond.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"dbus-org.fedoraproject.FirewallD1.service": {
"name": "dbus-org.fedoraproject.FirewallD1.service",
"source": "systemd",
"state": "active",
"status": "enabled"
},
"dbus-org.freedesktop.hostname1.service": {
"name": "dbus-org.freedesktop.hostname1.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dbus-org.freedesktop.locale1.service": {
"name": "dbus-org.freedesktop.locale1.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dbus-org.freedesktop.login1.service": {
"name": "dbus-org.freedesktop.login1.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dbus-org.freedesktop.nm-dispatcher.service": {
"name": "dbus-org.freedesktop.nm-dispatcher.service",
"source": "systemd",
"state": "active",
"status": "enabled"
},
"dbus-org.freedesktop.portable1.service": {
"name": "dbus-org.freedesktop.portable1.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dbus-org.freedesktop.timedate1.service": {
"name": "dbus-org.freedesktop.timedate1.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"dbus.service": {
"name": "dbus.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"debug-shell.service": {
"name": "debug-shell.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"dnf-makecache.service": {
"name": "dnf-makecache.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dnf-system-upgrade-cleanup.service": {
"name": "dnf-system-upgrade-cleanup.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dnf-system-upgrade.service": {
"name": "dnf-system-upgrade.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"dnsmasq.service": {
"name": "dnsmasq.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"dracut-cmdline.service": {
"name": "dracut-cmdline.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-initqueue.service": {
"name": "dracut-initqueue.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-mount.service": {
"name": "dracut-mount.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-mount.service": {
"name": "dracut-pre-mount.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-pivot.service": {
"name": "dracut-pre-pivot.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-trigger.service": {
"name": "dracut-pre-trigger.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-udev.service": {
"name": "dracut-pre-udev.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-shutdown-onfailure.service": {
"name": "dracut-shutdown-onfailure.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-shutdown.service": {
"name": "dracut-shutdown.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"ebtables.service": {
"name": "ebtables.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"emergency.service": {
"name": "emergency.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"firewalld.service": {
"name": "firewalld.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"fstrim.service": {
"name": "fstrim.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"getty@.service": {
"name": "getty@.service",
"source": "systemd",
"state": "unknown",
"status": "enabled"
},
"getty@tty1.service": {
"name": "getty@tty1.service",
"source": "systemd",
"state": "running",
"status": "unknown"
},
"grub-boot-indeterminate.service": {
"name": "grub-boot-indeterminate.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"gssproxy.service": {
"name": "gssproxy.service",
"source": "systemd",
"state": "running",
"status": "disabled"
},
"halt-local.service": {
"name": "halt-local.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"import-state.service": {
"name": "import-state.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"initrd-cleanup.service": {
"name": "initrd-cleanup.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"initrd-parse-etc.service": {
"name": "initrd-parse-etc.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"initrd-switch-root.service": {
"name": "initrd-switch-root.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"initrd-udevadm-cleanup-db.service": {
"name": "initrd-udevadm-cleanup-db.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"iprdump.service": {
"name": "iprdump.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"iprinit.service": {
"name": "iprinit.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"iprupdate.service": {
"name": "iprupdate.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"irqbalance.service": {
"name": "irqbalance.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"kdump.service": {
"name": "kdump.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"kmod-static-nodes.service": {
"name": "kmod-static-nodes.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"kvm_stat.service": {
"name": "kvm_stat.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"ldconfig.service": {
"name": "ldconfig.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"loadmodules.service": {
"name": "loadmodules.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"man-db-cache-update.service": {
"name": "man-db-cache-update.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"man-db-restart-cache-update.service": {
"name": "man-db-restart-cache-update.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"messagebus.service": {
"name": "messagebus.service",
"source": "systemd",
"state": "active",
"status": "static"
},
"microcode.service": {
"name": "microcode.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"nfs-blkmap.service": {
"name": "nfs-blkmap.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"nfs-convert.service": {
"name": "nfs-convert.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"nfs-idmapd.service": {
"name": "nfs-idmapd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nfs-mountd.service": {
"name": "nfs-mountd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nfs-server.service": {
"name": "nfs-server.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"nfs-utils.service": {
"name": "nfs-utils.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nfsdcld.service": {
"name": "nfsdcld.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nftables.service": {
"name": "nftables.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"nis-domainname.service": {
"name": "nis-domainname.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"oddjobd.service": {
"name": "oddjobd.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"plymouth-halt.service": {
"name": "plymouth-halt.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-kexec.service": {
"name": "plymouth-kexec.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-poweroff.service": {
"name": "plymouth-poweroff.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-quit-wait.service": {
"name": "plymouth-quit-wait.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-quit.service": {
"name": "plymouth-quit.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-read-write.service": {
"name": "plymouth-read-write.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-reboot.service": {
"name": "plymouth-reboot.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-start.service": {
"name": "plymouth-start.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-switch-root-initramfs.service": {
"name": "plymouth-switch-root-initramfs.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-switch-root.service": {
"name": "plymouth-switch-root.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"podman-auto-update.service": {
"name": "podman-auto-update.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"podman-clean-transient.service": {
"name": "podman-clean-transient.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"podman-kube@.service": {
"name": "podman-kube@.service",
"source": "systemd",
"state": "unknown",
"status": "disabled"
},
"podman-restart.service": {
"name": "podman-restart.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"podman.service": {
"name": "podman.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"polkit.service": {
"name": "polkit.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"qemu-guest-agent.service": {
"name": "qemu-guest-agent.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"quotaon.service": {
"name": "quotaon.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"rc-local.service": {
"name": "rc-local.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rdisc.service": {
"name": "rdisc.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"rescue.service": {
"name": "rescue.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"restraintd.service": {
"name": "restraintd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"rngd.service": {
"name": "rngd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"rpc-gssd.service": {
"name": "rpc-gssd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rpc-statd-notify.service": {
"name": "rpc-statd-notify.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rpc-statd.service": {
"name": "rpc-statd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rpcbind.service": {
"name": "rpcbind.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"rsyslog.service": {
"name": "rsyslog.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"selinux-autorelabel-mark.service": {
"name": "selinux-autorelabel-mark.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"selinux-autorelabel.service": {
"name": "selinux-autorelabel.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"serial-getty@.service": {
"name": "serial-getty@.service",
"source": "systemd",
"state": "unknown",
"status": "disabled"
},
"sshd-keygen@.service": {
"name": "sshd-keygen@.service",
"source": "systemd",
"state": "unknown",
"status": "disabled"
},
"sshd-keygen@ecdsa.service": {
"name": "sshd-keygen@ecdsa.service",
"source": "systemd",
"state": "stopped",
"status": "unknown"
},
"sshd-keygen@ed25519.service": {
"name": "sshd-keygen@ed25519.service",
"source": "systemd",
"state": "stopped",
"status": "unknown"
},
"sshd-keygen@rsa.service": {
"name": "sshd-keygen@rsa.service",
"source": "systemd",
"state": "stopped",
"status": "unknown"
},
"sshd.service": {
"name": "sshd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"sshd@.service": {
"name": "sshd@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"sssd-autofs.service": {
"name": "sssd-autofs.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-kcm.service": {
"name": "sssd-kcm.service",
"source": "systemd",
"state": "stopped",
"status": "indirect"
},
"sssd-nss.service": {
"name": "sssd-nss.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-pac.service": {
"name": "sssd-pac.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-pam.service": {
"name": "sssd-pam.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-ssh.service": {
"name": "sssd-ssh.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-sudo.service": {
"name": "sssd-sudo.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd.service": {
"name": "sssd.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"syslog.service": {
"name": "syslog.service",
"source": "systemd",
"state": "active",
"status": "enabled"
},
"system-update-cleanup.service": {
"name": "system-update-cleanup.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-ask-password-console.service": {
"name": "systemd-ask-password-console.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-ask-password-plymouth.service": {
"name": "systemd-ask-password-plymouth.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-ask-password-wall.service": {
"name": "systemd-ask-password-wall.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-backlight@.service": {
"name": "systemd-backlight@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-binfmt.service": {
"name": "systemd-binfmt.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-coredump@.service": {
"name": "systemd-coredump@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-exit.service": {
"name": "systemd-exit.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-firstboot.service": {
"name": "systemd-firstboot.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-fsck-root.service": {
"name": "systemd-fsck-root.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-fsck@.service": {
"name": "systemd-fsck@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-halt.service": {
"name": "systemd-halt.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-hibernate-resume@.service": {
"name": "systemd-hibernate-resume@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-hibernate.service": {
"name": "systemd-hibernate.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-hostnamed.service": {
"name": "systemd-hostnamed.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-hwdb-update.service": {
"name": "systemd-hwdb-update.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-hybrid-sleep.service": {
"name": "systemd-hybrid-sleep.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-initctl.service": {
"name": "systemd-initctl.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-journal-catalog-update.service": {
"name": "systemd-journal-catalog-update.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-journal-flush.service": {
"name": "systemd-journal-flush.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-journald.service": {
"name": "systemd-journald.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"systemd-kexec.service": {
"name": "systemd-kexec.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-localed.service": {
"name": "systemd-localed.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-logind.service": {
"name": "systemd-logind.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-machine-id-commit.service": {
"name": "systemd-machine-id-commit.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-modules-load.service": {
"name": "systemd-modules-load.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-portabled.service": {
"name": "systemd-portabled.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-poweroff.service": {
"name": "systemd-poweroff.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-pstore.service": {
"name": "systemd-pstore.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"systemd-quotacheck.service": {
"name": "systemd-quotacheck.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-random-seed.service": {
"name": "systemd-random-seed.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-reboot.service": {
"name": "systemd-reboot.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-remount-fs.service": {
"name": "systemd-remount-fs.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-resolved.service": {
"name": "systemd-resolved.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"systemd-rfkill.service": {
"name": "systemd-rfkill.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-suspend-then-hibernate.service": {
"name": "systemd-suspend-then-hibernate.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-suspend.service": {
"name": "systemd-suspend.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-sysctl.service": {
"name": "systemd-sysctl.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-sysusers.service": {
"name": "systemd-sysusers.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-timedated.service": {
"name": "systemd-timedated.service",
"source": "systemd",
"state": "inactive",
"status": "masked"
},
"systemd-tmpfiles-clean.service": {
"name": "systemd-tmpfiles-clean.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-tmpfiles-setup-dev.service": {
"name": "systemd-tmpfiles-setup-dev.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-tmpfiles-setup.service": {
"name": "systemd-tmpfiles-setup.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-udev-settle.service": {
"name": "systemd-udev-settle.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-udev-trigger.service": {
"name": "systemd-udev-trigger.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-udevd.service": {
"name": "systemd-udevd.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"systemd-update-done.service": {
"name": "systemd-update-done.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-update-utmp-runlevel.service": {
"name": "systemd-update-utmp-runlevel.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-update-utmp.service": {
"name": "systemd-update-utmp.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-user-sessions.service": {
"name": "systemd-user-sessions.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-vconsole-setup.service": {
"name": "systemd-vconsole-setup.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-volatile-root.service": {
"name": "systemd-volatile-root.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"tcsd.service": {
"name": "tcsd.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"teamd@.service": {
"name": "teamd@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"timedatex.service": {
"name": "timedatex.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"tuned.service": {
"name": "tuned.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"unbound-anchor.service": {
"name": "unbound-anchor.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"user-runtime-dir@.service": {
"name": "user-runtime-dir@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"user-runtime-dir@0.service": {
"name": "user-runtime-dir@0.service",
"source": "systemd",
"state": "stopped",
"status": "unknown"
},
"user@.service": {
"name": "user@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"user@0.service": {
"name": "user@0.service",
"source": "systemd",
"state": "running",
"status": "unknown"
}
}
},
"changed": false
}
TASK [fedora.linux_system_roles.podman : Create and update quadlets] ***********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:114
Saturday 24 August 2024 12:36:41 -0400 (0:00:01.615) 0:02:49.665 *******
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.podman : Cancel linger] ************************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:167
Saturday 24 August 2024 12:36:41 -0400 (0:00:00.032) 0:02:49.698 *******
TASK [fedora.linux_system_roles.podman : Handle credential files - absent] *****
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:173
Saturday 24 August 2024 12:36:41 -0400 (0:00:00.065) 0:02:49.763 *******
skipping: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.podman : Handle certs.d files - absent] ********
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:182
Saturday 24 August 2024 12:36:41 -0400 (0:00:00.032) 0:02:49.796 *******
skipping: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [Ensure no resources] *****************************************************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_demo.yml:188
Saturday 24 August 2024 12:36:41 -0400 (0:00:00.030) 0:02:49.827 *******
fatal: [managed_node1]: FAILED! => {
"assertion": "__podman_test_debug_volumes.stdout == \"\"",
"changed": false,
"evaluated_to": false
}
MSG:
Assertion failed
TASK [Debug] *******************************************************************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_demo.yml:199
Saturday 24 August 2024 12:36:41 -0400 (0:00:00.043) 0:02:49.871 *******
ok: [managed_node1] => {
"changed": false,
"cmd": "exec 1>&2\nset -x\nset -o pipefail\nsystemctl list-units --plain -l --all | grep quadlet || :\nsystemctl list-unit-files --all | grep quadlet || :\nsystemctl list-units --plain --failed -l --all | grep quadlet || :\n",
"delta": "0:00:00.322673",
"end": "2024-08-24 12:36:41.863123",
"rc": 0,
"start": "2024-08-24 12:36:41.540450"
}
STDERR:
+ set -o pipefail
+ systemctl list-units --plain -l --all
+ grep quadlet
+ :
+ systemctl list-unit-files --all
+ grep quadlet
+ :
+ systemctl list-units --plain --failed -l --all
+ grep quadlet
+ :
TASK [Get journald] ************************************************************
task path: /tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_demo.yml:209
Saturday 24 August 2024 12:36:41 -0400 (0:00:00.674) 0:02:50.545 *******
fatal: [managed_node1]: FAILED! => {
"changed": false,
"cmd": [
"journalctl",
"-ex"
],
"delta": "0:00:00.106022",
"end": "2024-08-24 12:36:42.321270",
"failed_when_result": true,
"rc": 0,
"start": "2024-08-24 12:36:42.215248"
}
STDOUT:
-- Logs begin at Sat 2024-08-24 12:23:16 EDT, end at Sat 2024-08-24 12:36:42 EDT. --
Aug 24 12:31:03 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state.
Aug 24 12:31:03 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 32ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources.
Aug 24 12:31:03 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.
-- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up.
--
-- The start-up result is done.
Aug 24 12:31:03 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state.
Aug 24 12:31:03 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 33ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources.
Aug 24 12:31:03 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.
-- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up.
--
-- The start-up result is done.
Aug 24 12:31:03 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[37282]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None
Aug 24 12:31:03 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state.
Aug 24 12:31:03 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 38ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources.
Aug 24 12:31:03 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.
-- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up.
--
-- The start-up result is done.
Aug 24 12:31:04 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state.
Aug 24 12:31:04 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 32ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources.
Aug 24 12:31:04 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.
-- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up.
--
-- The start-up result is done.
Aug 24 12:31:04 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state.
Aug 24 12:31:04 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 33ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources.
Aug 24 12:31:04 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.
-- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up.
--
-- The start-up result is done.
Aug 24 12:31:04 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[37501]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1
Aug 24 12:31:04 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state.
Aug 24 12:31:04 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 40ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources.
Aug 24 12:31:04 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.
-- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up.
--
-- The start-up result is done.
Aug 24 12:31:04 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state.
Aug 24 12:31:04 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 37ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources.
Aug 24 12:31:04 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.
-- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up.
--
-- The start-up result is done.
Aug 24 12:31:04 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state.
Aug 24 12:31:04 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 31ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources.
Aug 24 12:31:05 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.
-- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up.
--
-- The start-up result is done.
Aug 24 12:31:05 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state.
Aug 24 12:31:05 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 31ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources.
Aug 24 12:31:05 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.
-- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up.
--
-- The start-up result is done.
Aug 24 12:31:05 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state.
Aug 24 12:31:05 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 30ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources.
Aug 24 12:31:05 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.
-- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up.
--
-- The start-up result is done.
Aug 24 12:31:05 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state.
Aug 24 12:31:05 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 30ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources.
Aug 24 12:31:05 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.
-- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up.
--
-- The start-up result is done.
Aug 24 12:31:05 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state.
Aug 24 12:31:05 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 31ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources.
Aug 24 12:31:05 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.
-- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up.
--
-- The start-up result is done.
Aug 24 12:31:05 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state.
Aug 24 12:31:05 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 30ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources.
Aug 24 12:31:05 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.
-- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up.
--
-- The start-up result is done.
Aug 24 12:31:06 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state.
Aug 24 12:31:06 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 32ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources.
Aug 24 12:31:06 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.
-- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up.
--
-- The start-up result is done.
Aug 24 12:31:06 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state.
Aug 24 12:31:06 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 33ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources.
Aug 24 12:31:06 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.
-- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up.
--
-- The start-up result is done.
Aug 24 12:31:06 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state.
Aug 24 12:31:06 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 39ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources.
Aug 24 12:31:06 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.
-- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up.
--
-- The start-up result is done.
Aug 24 12:31:06 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state.
Aug 24 12:31:06 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 37ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources.
Aug 24 12:31:06 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.
-- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up.
--
-- The start-up result is done.
Aug 24 12:31:07 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state.
Aug 24 12:31:07 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 38ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources.
Aug 24 12:31:07 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.
-- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up.
--
-- The start-up result is done.
Aug 24 12:31:07 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state.
Aug 24 12:31:07 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 37ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources.
Aug 24 12:31:07 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.
-- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up.
--
-- The start-up result is done.
Aug 24 12:31:07 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state.
Aug 24 12:31:07 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 38ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources.
Aug 24 12:31:07 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.
-- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up.
--
-- The start-up result is done.
Aug 24 12:31:07 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state.
Aug 24 12:31:07 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 32ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources.
Aug 24 12:31:07 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.
-- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up.
--
-- The start-up result is done.
Aug 24 12:31:08 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state.
Aug 24 12:31:08 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 30ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources.
Aug 24 12:31:08 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.
-- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up.
--
-- The start-up result is done.
Aug 24 12:31:08 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state.
Aug 24 12:31:08 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 33ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources.
Aug 24 12:31:08 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.
-- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up.
--
-- The start-up result is done.
Aug 24 12:31:08 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state.
Aug 24 12:31:08 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 37ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources.
Aug 24 12:31:08 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[38402]: ansible-command Invoked with _raw_params=podman --version warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:31:08 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.
-- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up.
--
-- The start-up result is done.
Aug 24 12:31:08 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state.
Aug 24 12:31:08 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 31ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources.
Aug 24 12:31:08 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.
-- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up.
--
-- The start-up result is done.
Aug 24 12:31:08 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state.
Aug 24 12:31:08 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 30ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources.
Aug 24 12:31:09 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.
-- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up.
--
-- The start-up result is done.
Aug 24 12:31:09 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state.
Aug 24 12:31:09 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 33ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources.
Aug 24 12:31:09 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.
-- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up.
--
-- The start-up result is done.
Aug 24 12:31:09 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state.
Aug 24 12:31:09 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 32ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources.
Aug 24 12:31:09 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.
-- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up.
--
-- The start-up result is done.
Aug 24 12:31:09 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state.
Aug 24 12:31:09 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 36ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources.
Aug 24 12:31:09 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[38707]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None
Aug 24 12:31:09 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.
-- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up.
--
-- The start-up result is done.
Aug 24 12:31:09 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state.
Aug 24 12:31:09 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 32ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources.
Aug 24 12:31:09 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.
-- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up.
--
-- The start-up result is done.
Aug 24 12:31:09 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state.
Aug 24 12:31:09 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 34ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources.
Aug 24 12:31:10 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.
-- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up.
--
-- The start-up result is done.
Aug 24 12:31:10 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state.
Aug 24 12:31:10 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 40ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources.
Aug 24 12:31:10 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[38937]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1
Aug 24 12:31:10 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.
-- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up.
--
-- The start-up result is done.
Aug 24 12:31:10 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state.
Aug 24 12:31:10 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 32ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources.
Aug 24 12:31:10 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.
-- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up.
--
-- The start-up result is done.
Aug 24 12:31:10 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state.
Aug 24 12:31:10 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 32ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources.
Aug 24 12:31:10 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.
-- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up.
--
-- The start-up result is done.
Aug 24 12:31:10 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state.
Aug 24 12:31:10 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 31ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources.
Aug 24 12:31:10 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.
-- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up.
--
-- The start-up result is done.
Aug 24 12:31:11 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state.
Aug 24 12:31:11 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 31ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources.
Aug 24 12:31:11 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.
-- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up.
--
-- The start-up result is done.
Aug 24 12:31:11 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state.
Aug 24 12:31:11 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 32ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources.
Aug 24 12:31:11 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.
-- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up.
--
-- The start-up result is done.
Aug 24 12:31:11 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state.
Aug 24 12:31:11 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 31ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources.
Aug 24 12:31:11 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.
-- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up.
--
-- The start-up result is done.
Aug 24 12:31:11 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state.
Aug 24 12:31:11 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 30ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources.
Aug 24 12:31:11 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.
-- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up.
--
-- The start-up result is done.
Aug 24 12:31:11 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state.
Aug 24 12:31:11 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 32ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources.
Aug 24 12:31:11 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.
-- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up.
--
-- The start-up result is done.
Aug 24 12:31:11 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state.
Aug 24 12:31:11 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 30ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources.
Aug 24 12:31:12 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.
-- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up.
--
-- The start-up result is done.
Aug 24 12:31:12 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state.
Aug 24 12:31:12 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 31ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources.
Aug 24 12:31:12 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.
-- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up.
--
-- The start-up result is done.
Aug 24 12:31:12 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state.
Aug 24 12:31:12 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 33ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources.
Aug 24 12:31:12 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.
-- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up.
--
-- The start-up result is done.
Aug 24 12:31:12 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state.
Aug 24 12:31:12 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 38ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources.
Aug 24 12:31:12 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[39479]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None
Aug 24 12:31:12 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.
-- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up.
--
-- The start-up result is done.
Aug 24 12:31:12 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state.
Aug 24 12:31:12 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 32ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources.
Aug 24 12:31:12 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.
-- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up.
--
-- The start-up result is done.
Aug 24 12:31:13 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state.
Aug 24 12:31:13 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 34ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources.
Aug 24 12:31:13 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[39681]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1
Aug 24 12:31:13 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.
-- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up.
--
-- The start-up result is done.
Aug 24 12:31:13 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state.
Aug 24 12:31:13 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 31ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources.
Aug 24 12:31:13 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.
-- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up.
--
-- The start-up result is done.
Aug 24 12:31:13 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state.
Aug 24 12:31:13 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 31ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources.
Aug 24 12:31:13 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.
-- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up.
--
-- The start-up result is done.
Aug 24 12:31:13 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state.
Aug 24 12:31:13 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 30ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources.
Aug 24 12:31:13 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.
-- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up.
--
-- The start-up result is done.
Aug 24 12:31:13 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state.
Aug 24 12:31:13 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 30ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources.
Aug 24 12:31:13 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.
-- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up.
--
-- The start-up result is done.
Aug 24 12:31:13 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state.
Aug 24 12:31:13 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 29ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources.
Aug 24 12:31:14 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.
-- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up.
--
-- The start-up result is done.
Aug 24 12:31:14 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state.
Aug 24 12:31:14 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 30ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources.
Aug 24 12:31:14 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.
-- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up.
--
-- The start-up result is done.
Aug 24 12:31:14 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state.
Aug 24 12:31:14 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 30ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources.
Aug 24 12:31:14 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.
-- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up.
--
-- The start-up result is done.
Aug 24 12:31:14 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state.
Aug 24 12:31:14 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 33ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources.
Aug 24 12:31:14 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.
-- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up.
--
-- The start-up result is done.
Aug 24 12:31:14 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[40107]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None
Aug 24 12:31:14 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state.
Aug 24 12:31:14 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 38ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources.
Aug 24 12:31:14 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.
-- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up.
--
-- The start-up result is done.
Aug 24 12:31:15 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state.
Aug 24 12:31:15 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 32ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources.
Aug 24 12:31:15 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.
-- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up.
--
-- The start-up result is done.
Aug 24 12:31:15 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state.
Aug 24 12:31:15 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 38ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources.
Aug 24 12:31:15 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[40323]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1
Aug 24 12:31:15 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.
-- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up.
--
-- The start-up result is done.
Aug 24 12:31:15 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state.
Aug 24 12:31:15 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 31ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources.
Aug 24 12:31:15 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.
-- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up.
--
-- The start-up result is done.
Aug 24 12:31:15 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state.
Aug 24 12:31:15 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 32ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources.
Aug 24 12:31:15 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.
-- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up.
--
-- The start-up result is done.
Aug 24 12:31:15 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state.
Aug 24 12:31:15 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 30ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources.
Aug 24 12:31:15 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.
-- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up.
--
-- The start-up result is done.
Aug 24 12:31:16 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state.
Aug 24 12:31:16 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 30ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources.
Aug 24 12:31:16 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.
-- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up.
--
-- The start-up result is done.
Aug 24 12:31:16 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state.
Aug 24 12:31:16 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 30ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources.
Aug 24 12:31:16 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.
-- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up.
--
-- The start-up result is done.
Aug 24 12:31:16 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state.
Aug 24 12:31:16 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 30ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources.
Aug 24 12:31:16 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.
-- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up.
--
-- The start-up result is done.
Aug 24 12:31:16 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state.
Aug 24 12:31:16 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 32ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources.
Aug 24 12:31:16 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.
-- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up.
--
-- The start-up result is done.
Aug 24 12:31:16 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state.
Aug 24 12:31:16 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 30ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources.
Aug 24 12:31:16 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.
-- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up.
--
-- The start-up result is done.
Aug 24 12:31:16 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state.
Aug 24 12:31:16 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 30ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources.
Aug 24 12:31:17 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.
-- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up.
--
-- The start-up result is done.
Aug 24 12:31:17 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state.
Aug 24 12:31:17 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 31ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources.
Aug 24 12:31:17 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.
-- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up.
--
-- The start-up result is done.
Aug 24 12:31:17 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state.
Aug 24 12:31:17 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 35ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources.
Aug 24 12:31:17 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.
-- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up.
--
-- The start-up result is done.
Aug 24 12:31:17 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state.
Aug 24 12:31:17 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 35ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources.
Aug 24 12:31:17 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[40875]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None
Aug 24 12:31:17 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.
-- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up.
--
-- The start-up result is done.
Aug 24 12:31:17 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state.
Aug 24 12:31:17 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 33ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources.
Aug 24 12:31:17 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.
-- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up.
--
-- The start-up result is done.
Aug 24 12:31:18 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state.
Aug 24 12:31:18 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 33ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources.
Aug 24 12:31:18 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.
-- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up.
--
-- The start-up result is done.
Aug 24 12:31:18 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state.
Aug 24 12:31:18 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 37ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources.
Aug 24 12:31:18 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[41104]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1
Aug 24 12:31:18 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.
-- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up.
--
-- The start-up result is done.
Aug 24 12:31:18 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state.
Aug 24 12:31:18 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 30ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources.
Aug 24 12:31:18 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.
-- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up.
--
-- The start-up result is done.
Aug 24 12:31:18 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state.
Aug 24 12:31:18 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 32ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources.
Aug 24 12:31:18 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.
-- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up.
--
-- The start-up result is done.
Aug 24 12:31:18 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state.
Aug 24 12:31:18 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 30ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources.
Aug 24 12:31:18 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.
-- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up.
--
-- The start-up result is done.
Aug 24 12:31:19 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state.
Aug 24 12:31:19 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 33ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources.
Aug 24 12:31:19 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.
-- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up.
--
-- The start-up result is done.
Aug 24 12:31:19 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state.
Aug 24 12:31:19 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 38ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources.
Aug 24 12:31:19 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[41406]: ansible-command Invoked with _raw_params= warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:31:19 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.
-- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up.
--
-- The start-up result is done.
Aug 24 12:31:19 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state.
Aug 24 12:31:19 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 33ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources.
Aug 24 12:31:19 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.
-- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up.
--
-- The start-up result is done.
Aug 24 12:31:19 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state.
Aug 24 12:31:19 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 34ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources.
Aug 24 12:31:19 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.
-- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up.
--
-- The start-up result is done.
Aug 24 12:31:19 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state.
Aug 24 12:31:19 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 36ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources.
Aug 24 12:31:20 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.
-- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up.
--
-- The start-up result is done.
Aug 24 12:31:20 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[41637]: ansible-systemd Invoked with name= scope=system state=stopped enabled=False daemon_reload=False daemon_reexec=False no_block=False force=None masked=None user=None
Aug 24 12:31:20 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Reloading.
Aug 24 12:31:20 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state.
Aug 24 12:31:20 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 38ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources.
Aug 24 12:31:20 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Stopping A template for running K8s workloads via podman-kube-play...
-- Subject: Unit podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service has begun shutting down
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service has begun shutting down.
Aug 24 12:31:20 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.
-- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished starting up.
--
-- The start-up result is done.
Aug 24 12:31:20 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-49bd35886401f2696f39559535070a7acae98c93915196aab21a7de3174e59ba.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-49bd35886401f2696f39559535070a7acae98c93915196aab21a7de3174e59ba.scope has successfully entered the 'dead' state.
Aug 24 12:31:20 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-49bd35886401f2696f39559535070a7acae98c93915196aab21a7de3174e59ba.scope: Consumed 32ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-49bd35886401f2696f39559535070a7acae98c93915196aab21a7de3174e59ba.scope completed and consumed the indicated resources.
Aug 24 12:31:20 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has successfully entered the 'dead' state.
Aug 24 12:31:20 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Stopped libcontainer container 765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.
-- Subject: Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished shutting down
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope has finished shutting down.
Aug 24 12:31:20 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope: Consumed 600us CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-765d8c2cddaf785a845e25cdf0070d713f4c7bbc13230622c2c438458f22b813.scope completed and consumed the indicated resources.
Aug 24 12:31:20 ip-10-31-40-15.us-east-1.aws.redhat.com conmon[41736]: conmon 765d8c2cddaf785a845e : runtime stderr: runc create failed: unable to start container process: can't get final child's PID from pipe: EOF
Aug 24 12:31:20 ip-10-31-40-15.us-east-1.aws.redhat.com conmon[41736]: conmon 765d8c2cddaf785a845e : Failed to create container: exit status 1
Aug 24 12:31:20 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay-df494c4e221d99e0f620fb00cd28f2cc006ffcc851b7d7eade776335b55ac463-merged.mount: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit var-lib-containers-storage-overlay-df494c4e221d99e0f620fb00cd28f2cc006ffcc851b7d7eade776335b55ac463-merged.mount has successfully entered the 'dead' state.
Aug 24 12:31:20 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: cni-podman1: port 1(veth727493b2) entered disabled state
Aug 24 12:31:20 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: device veth727493b2 left promiscuous mode
Aug 24 12:31:20 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: cni-podman1: port 1(veth727493b2) entered disabled state
Aug 24 12:31:20 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: run-netns-netns\x2de2bae42d\x2da696\x2d4b35\x2d2126\x2d284fd3d6a73b.mount: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit run-netns-netns\x2de2bae42d\x2da696\x2d4b35\x2d2126\x2d284fd3d6a73b.mount has successfully entered the 'dead' state.
Aug 24 12:31:20 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-49bd35886401f2696f39559535070a7acae98c93915196aab21a7de3174e59ba-userdata-shm.mount: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit var-lib-containers-storage-overlay\x2dcontainers-49bd35886401f2696f39559535070a7acae98c93915196aab21a7de3174e59ba-userdata-shm.mount has successfully entered the 'dead' state.
Aug 24 12:31:20 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay-536a0786a016d674449abd21820ed1ec2659103dcfcc60e560eb1b13ba0fb457-merged.mount: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit var-lib-containers-storage-overlay-536a0786a016d674449abd21820ed1ec2659103dcfcc60e560eb1b13ba0fb457-merged.mount has successfully entered the 'dead' state.
Aug 24 12:31:20 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Removed slice cgroup machine-libpod_pod_c5e3848e0e861be93b14b306c309ae7576e20076beb5aaaf5660f3b6bec6eb4e.slice.
-- Subject: Unit machine-libpod_pod_c5e3848e0e861be93b14b306c309ae7576e20076beb5aaaf5660f3b6bec6eb4e.slice has finished shutting down
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit machine-libpod_pod_c5e3848e0e861be93b14b306c309ae7576e20076beb5aaaf5660f3b6bec6eb4e.slice has finished shutting down.
Aug 24 12:31:20 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: machine-libpod_pod_c5e3848e0e861be93b14b306c309ae7576e20076beb5aaaf5660f3b6bec6eb4e.slice: Consumed 11.825s CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit machine-libpod_pod_c5e3848e0e861be93b14b306c309ae7576e20076beb5aaaf5660f3b6bec6eb4e.slice completed and consumed the indicated resources.
Aug 24 12:31:20 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-80915bd8fbddf68e48445610f66f32c8466405cc58a3790cf9782f24652d9f2c.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-80915bd8fbddf68e48445610f66f32c8466405cc58a3790cf9782f24652d9f2c.scope has successfully entered the 'dead' state.
Aug 24 12:31:20 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-80915bd8fbddf68e48445610f66f32c8466405cc58a3790cf9782f24652d9f2c.scope: Consumed 33ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-80915bd8fbddf68e48445610f66f32c8466405cc58a3790cf9782f24652d9f2c.scope completed and consumed the indicated resources.
Aug 24 12:31:20 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-80915bd8fbddf68e48445610f66f32c8466405cc58a3790cf9782f24652d9f2c-userdata-shm.mount: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit var-lib-containers-storage-overlay\x2dcontainers-80915bd8fbddf68e48445610f66f32c8466405cc58a3790cf9782f24652d9f2c-userdata-shm.mount has successfully entered the 'dead' state.
Aug 24 12:31:20 ip-10-31-40-15.us-east-1.aws.redhat.com podman[41721]: Pods stopped:
Aug 24 12:31:20 ip-10-31-40-15.us-east-1.aws.redhat.com podman[41721]: c5e3848e0e861be93b14b306c309ae7576e20076beb5aaaf5660f3b6bec6eb4e
Aug 24 12:31:20 ip-10-31-40-15.us-east-1.aws.redhat.com podman[41721]: Pods removed:
Aug 24 12:31:20 ip-10-31-40-15.us-east-1.aws.redhat.com podman[41721]: c5e3848e0e861be93b14b306c309ae7576e20076beb5aaaf5660f3b6bec6eb4e
Aug 24 12:31:20 ip-10-31-40-15.us-east-1.aws.redhat.com podman[41721]: Secrets removed:
Aug 24 12:31:20 ip-10-31-40-15.us-east-1.aws.redhat.com podman[41721]: Volumes removed:
Aug 24 12:31:20 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service has successfully entered the 'dead' state.
Aug 24 12:31:20 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Stopped A template for running K8s workloads via podman-kube-play.
-- Subject: Unit podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service has finished shutting down
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service has finished shutting down.
Aug 24 12:31:21 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[41993]: ansible-stat Invoked with path=/etc/containers/ansible-kubernetes.d/auth_test_1_kube.yml follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1
Aug 24 12:31:21 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay-a55b6a86d999a458b4a32b86117737eed9e58e0082a4227e2d3fc866782b9269-merged.mount: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit var-lib-containers-storage-overlay-a55b6a86d999a458b4a32b86117737eed9e58e0082a4227e2d3fc866782b9269-merged.mount has successfully entered the 'dead' state.
Aug 24 12:31:21 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[42118]: ansible-containers.podman.podman_play Invoked with state=absent kube_file=/etc/containers/ansible-kubernetes.d/auth_test_1_kube.yml executable=podman annotation=None authfile=None build=None cert_dir=None configmap=None context_dir=None seccomp_profile_root=None username=None password=NOT_LOGGING_PARAMETER log_driver=None log_opt=None network=None tls_verify=None debug=None quiet=None recreate=None userns=None log_level=None quadlet_dir=None quadlet_filename=None quadlet_options=None
Aug 24 12:31:21 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[42118]: ansible-containers.podman.podman_play version: 4.9.4-dev, kube file /etc/containers/ansible-kubernetes.d/auth_test_1_kube.yml
Aug 24 12:31:22 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[42254]: ansible-file Invoked with path=/etc/containers/ansible-kubernetes.d/auth_test_1_kube.yml state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None
Aug 24 12:31:22 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[42377]: ansible-command Invoked with _raw_params=podman image prune -f warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:31:24 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[42507]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None
Aug 24 12:31:24 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[42631]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1
Aug 24 12:31:26 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[42756]: ansible-systemd Invoked with name=auth_test_1_quadlet.service scope=system state=stopped enabled=False force=True daemon_reload=False daemon_reexec=False no_block=False masked=None user=None
Aug 24 12:31:26 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Reloading.
Aug 24 12:31:26 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[42911]: ansible-stat Invoked with path=/etc/containers/systemd/auth_test_1_quadlet.container follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1
Aug 24 12:31:27 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[43159]: ansible-file Invoked with path=/etc/containers/systemd/auth_test_1_quadlet.container state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None
Aug 24 12:31:28 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[43282]: ansible-systemd Invoked with daemon_reload=True scope=system daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None user=None
Aug 24 12:31:28 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Reloading.
Aug 24 12:31:31 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[43565]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None
Aug 24 12:31:31 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[43689]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1
Aug 24 12:31:34 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[44060]: ansible-file Invoked with path=/root/.config/containers state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None
Aug 24 12:31:35 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[44183]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None
Aug 24 12:31:36 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[44307]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1
Aug 24 12:31:39 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[45047]: ansible-user Invoked with name=auth_test_user1 uid=2001 state=present non_unique=False force=False remove=False create_home=True system=False move_home=False append=False ssh_key_bits=0 ssh_key_type=rsa ssh_key_comment=ansible-generated on ip-10-31-40-15.us-east-1.aws.redhat.com update_password=always group=None groups=None comment=None home=None shell=None password=NOT_LOGGING_PARAMETER login_class=None hidden=None seuser=None skeleton=None generate_ssh_key=None ssh_key_file=None ssh_key_passphrase=NOT_LOGGING_PARAMETER expires=None password_lock=None local=None profile=None authorization=None role=None
Aug 24 12:31:39 ip-10-31-40-15.us-east-1.aws.redhat.com useradd[45051]: new group: name=auth_test_user1, GID=2001
Aug 24 12:31:39 ip-10-31-40-15.us-east-1.aws.redhat.com useradd[45051]: new user: name=auth_test_user1, UID=2001, GID=2001, home=/home/auth_test_user1, shell=/bin/bash
Aug 24 12:31:43 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[45304]: ansible-command Invoked with _raw_params=podman --version warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:31:43 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[45433]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None
Aug 24 12:31:44 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[45557]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1
Aug 24 12:31:47 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[45682]: ansible-getent Invoked with database=passwd key=auth_test_user1 fail_key=False service=None split=None
Aug 24 12:31:47 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[45806]: ansible-getent Invoked with database=group key=2001 fail_key=False service=None split=None
Aug 24 12:31:48 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[45930]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1
Aug 24 12:31:48 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[46055]: ansible-command Invoked with _raw_params=getsubids auth_test_user1 warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:31:49 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[46179]: ansible-command Invoked with _raw_params=getsubids -g auth_test_user1 warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:31:50 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[46303]: ansible-file Invoked with path=/home/auth_test_user1/.config/containers/certs.d/localhost:5000 state=directory owner=auth_test_user1 group=2001 mode=0700 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None
Aug 24 12:31:52 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[46650]: ansible-getent Invoked with database=group key=2001 fail_key=False service=None split=None
Aug 24 12:31:53 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[46774]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1
Aug 24 12:31:53 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[46899]: ansible-command Invoked with _raw_params=getsubids auth_test_user1 warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:31:53 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[47023]: ansible-command Invoked with _raw_params=getsubids -g auth_test_user1 warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:31:54 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[47147]: ansible-file Invoked with path=/home/auth_test_user1/.config/containers state=directory owner=auth_test_user1 group=2001 mode=0700 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None
Aug 24 12:31:56 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[47494]: ansible-getent Invoked with database=group key=2001 fail_key=False service=None split=None
Aug 24 12:31:57 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[47618]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1
Aug 24 12:31:57 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[47743]: ansible-command Invoked with _raw_params=getsubids auth_test_user1 warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:31:58 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[47867]: ansible-command Invoked with _raw_params=getsubids -g auth_test_user1 warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:31:59 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[47991]: ansible-command Invoked with _raw_params= warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:32:00 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[48115]: ansible-command Invoked with creates=/var/lib/systemd/linger/auth_test_user1 _raw_params=loginctl enable-linger auth_test_user1 warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None removes=None stdin=None
Aug 24 12:32:00 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Created slice User Slice of UID 2001.
-- Subject: Unit user-2001.slice has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit user-2001.slice has finished starting up.
--
-- The start-up result is done.
Aug 24 12:32:00 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Starting User runtime directory /run/user/2001...
-- Subject: Unit user-runtime-dir@2001.service has begun start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit user-runtime-dir@2001.service has begun starting up.
Aug 24 12:32:00 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started User runtime directory /run/user/2001.
-- Subject: Unit user-runtime-dir@2001.service has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit user-runtime-dir@2001.service has finished starting up.
--
-- The start-up result is done.
Aug 24 12:32:00 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Starting User Manager for UID 2001...
-- Subject: Unit user@2001.service has begun start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit user@2001.service has begun starting up.
Aug 24 12:32:00 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: pam_unix(systemd-user:session): session opened for user auth_test_user1 by (uid=0)
Aug 24 12:32:00 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Reached target Paths.
-- Subject: Unit UNIT has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit UNIT has finished starting up.
--
-- The start-up result is done.
Aug 24 12:32:00 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Starting D-Bus User Message Bus Socket.
-- Subject: Unit UNIT has begun start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit UNIT has begun starting up.
Aug 24 12:32:00 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Started Mark boot as successful after the user session has run 2 minutes.
-- Subject: Unit UNIT has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit UNIT has finished starting up.
--
-- The start-up result is done.
Aug 24 12:32:00 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Reached target Timers.
-- Subject: Unit UNIT has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit UNIT has finished starting up.
--
-- The start-up result is done.
Aug 24 12:32:00 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Listening on D-Bus User Message Bus Socket.
-- Subject: Unit UNIT has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit UNIT has finished starting up.
--
-- The start-up result is done.
Aug 24 12:32:00 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Reached target Sockets.
-- Subject: Unit UNIT has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit UNIT has finished starting up.
--
-- The start-up result is done.
Aug 24 12:32:00 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Reached target Basic System.
-- Subject: Unit UNIT has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit UNIT has finished starting up.
--
-- The start-up result is done.
Aug 24 12:32:00 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Reached target Default.
-- Subject: Unit UNIT has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit UNIT has finished starting up.
--
-- The start-up result is done.
Aug 24 12:32:00 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Startup finished in 37ms.
-- Subject: User manager start-up is now complete
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The user manager instance for user 2001 has been started. All services queued
-- for starting have been started. Note that other services might still be starting
-- up or be started at any later time.
--
-- Startup of the manager took 37357 microseconds.
Aug 24 12:32:00 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started User Manager for UID 2001.
-- Subject: Unit user@2001.service has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit user@2001.service has finished starting up.
--
-- The start-up result is done.
Aug 24 12:32:01 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[48255]: root : TTY=pts/0 ; PWD=/root ; USER=auth_test_user1 ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-ucyqvqaeqcbwzlreelmtcakpvqhqhicr ; XDG_RUNTIME_DIR=/run/user/2001 /usr/libexec/platform-python /var/tmp/ansible-tmp-1724517121.0929248-20827-180187680463393/AnsiballZ_podman_image.py'
Aug 24 12:32:01 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[48255]: pam_unix(sudo:session): session opened for user auth_test_user1 by root(uid=0)
Aug 24 12:32:01 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Started D-Bus User Message Bus.
-- Subject: Unit UNIT has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit UNIT has finished starting up.
--
-- The start-up result is done.
Aug 24 12:32:01 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Created slice user.slice.
-- Subject: Unit UNIT has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit UNIT has finished starting up.
--
-- The start-up result is done.
Aug 24 12:32:01 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Started podman-48267.scope.
-- Subject: Unit UNIT has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit UNIT has finished starting up.
--
-- The start-up result is done.
Aug 24 12:32:01 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Started podman-pause-0e20a755.scope.
-- Subject: Unit UNIT has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit UNIT has finished starting up.
--
-- The start-up result is done.
Aug 24 12:32:01 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Started podman-48283.scope.
-- Subject: Unit UNIT has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit UNIT has finished starting up.
--
-- The start-up result is done.
Aug 24 12:32:01 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Started podman-48290.scope.
-- Subject: Unit UNIT has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit UNIT has finished starting up.
--
-- The start-up result is done.
Aug 24 12:32:02 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Started podman-48306.scope.
-- Subject: Unit UNIT has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit UNIT has finished starting up.
--
-- The start-up result is done.
Aug 24 12:32:02 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[48255]: pam_unix(sudo:session): session closed for user auth_test_user1
Aug 24 12:32:02 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[48436]: ansible-stat Invoked with path=/home/auth_test_user1/.config/containers/ansible-kubernetes.d/auth_test_1_kube.yml follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1
Aug 24 12:32:03 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[48559]: ansible-file Invoked with path=/home/auth_test_user1/.config/containers/ansible-kubernetes.d state=directory owner=auth_test_user1 group=2001 mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None
Aug 24 12:32:03 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[48682]: ansible-stat Invoked with path=/home/auth_test_user1/.config/containers/ansible-kubernetes.d/auth_test_1_kube.yml follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True
Aug 24 12:32:03 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[48781]: ansible-copy Invoked with dest=/home/auth_test_user1/.config/containers/ansible-kubernetes.d/auth_test_1_kube.yml owner=auth_test_user1 group=2001 mode=0644 src=/root/.ansible/tmp/ansible-tmp-1724517123.15822-20917-242171895731390/source _original_basename=tmpj93_bt8l follow=False checksum=fb0097683a2e5c8909a8037d64ddc1b350aed0be backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None regexp=None delimiter=None
Aug 24 12:32:04 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[48906]: root : TTY=pts/0 ; PWD=/root ; USER=auth_test_user1 ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-tltabhiijzzijcbpxvlygvztxmltcxps ; XDG_RUNTIME_DIR=/run/user/2001 /usr/libexec/platform-python /var/tmp/ansible-tmp-1724517123.8727906-20945-22576776740070/AnsiballZ_podman_play.py'
Aug 24 12:32:04 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[48906]: pam_unix(sudo:session): session opened for user auth_test_user1 by root(uid=0)
Aug 24 12:32:04 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[48909]: ansible-containers.podman.podman_play Invoked with state=started kube_file=/home/auth_test_user1/.config/containers/ansible-kubernetes.d/auth_test_1_kube.yml executable=podman annotation=None authfile=None build=None cert_dir=None configmap=None context_dir=None seccomp_profile_root=None username=None password=NOT_LOGGING_PARAMETER log_driver=None log_opt=None network=None tls_verify=None debug=None quiet=None recreate=None userns=None log_level=None quadlet_dir=None quadlet_filename=None quadlet_options=None
Aug 24 12:32:04 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Started podman-48917.scope.
-- Subject: Unit UNIT has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit UNIT has finished starting up.
--
-- The start-up result is done.
Aug 24 12:32:04 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: tun: Universal TUN/TAP device driver, 1.6
Aug 24 12:32:04 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Started rootless-netns-804780fa.scope.
-- Subject: Unit UNIT has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit UNIT has finished starting up.
--
-- The start-up result is done.
Aug 24 12:32:04 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: IPv6: ADDRCONF(NETDEV_UP): vethd69b4c65: link is not ready
Aug 24 12:32:04 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: cni-podman2: port 1(vethd69b4c65) entered blocking state
Aug 24 12:32:04 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: cni-podman2: port 1(vethd69b4c65) entered disabled state
Aug 24 12:32:04 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: device vethd69b4c65 entered promiscuous mode
Aug 24 12:32:04 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: IPv6: ADDRCONF(NETDEV_UP): eth0: link is not ready
Aug 24 12:32:04 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: IPv6: ADDRCONF(NETDEV_CHANGE): eth0: link becomes ready
Aug 24 12:32:04 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: IPv6: ADDRCONF(NETDEV_CHANGE): vethd69b4c65: link becomes ready
Aug 24 12:32:04 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: cni-podman2: port 1(vethd69b4c65) entered blocking state
Aug 24 12:32:04 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: cni-podman2: port 1(vethd69b4c65) entered forwarding state
Aug 24 12:32:04 ip-10-31-40-15.us-east-1.aws.redhat.com dnsmasq[49050]: listening on cni-podman2(#3): 10.89.1.1
Aug 24 12:32:04 ip-10-31-40-15.us-east-1.aws.redhat.com dnsmasq[49052]: started, version 2.79 cachesize 150
Aug 24 12:32:04 ip-10-31-40-15.us-east-1.aws.redhat.com dnsmasq[49052]: compile time options: IPv6 GNU-getopt DBus no-i18n IDN2 DHCP DHCPv6 no-Lua TFTP no-conntrack ipset auth DNSSEC loop-detect inotify
Aug 24 12:32:04 ip-10-31-40-15.us-east-1.aws.redhat.com dnsmasq[49052]: using local addresses only for domain dns.podman
Aug 24 12:32:04 ip-10-31-40-15.us-east-1.aws.redhat.com dnsmasq[49052]: reading /etc/resolv.conf
Aug 24 12:32:04 ip-10-31-40-15.us-east-1.aws.redhat.com dnsmasq[49052]: using local addresses only for domain dns.podman
Aug 24 12:32:04 ip-10-31-40-15.us-east-1.aws.redhat.com dnsmasq[49052]: using nameserver 10.0.2.3#53
Aug 24 12:32:04 ip-10-31-40-15.us-east-1.aws.redhat.com dnsmasq[49052]: using nameserver 10.29.169.13#53
Aug 24 12:32:04 ip-10-31-40-15.us-east-1.aws.redhat.com dnsmasq[49052]: using nameserver 10.29.170.12#53
Aug 24 12:32:04 ip-10-31-40-15.us-east-1.aws.redhat.com dnsmasq[49052]: using nameserver 10.2.32.1#53
Aug 24 12:32:04 ip-10-31-40-15.us-east-1.aws.redhat.com dnsmasq[49052]: read /run/user/2001/containers/cni/dnsname/podman-default-kube-network/addnhosts - 1 addresses
Aug 24 12:32:05 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[48906]: pam_unix(sudo:session): session closed for user auth_test_user1
Aug 24 12:32:05 ip-10-31-40-15.us-east-1.aws.redhat.com conmon[49128]: conmon d83f34166aafd6cd4a14 : Failed to get working directory
Aug 24 12:32:05 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: cni-podman2: port 1(vethd69b4c65) entered disabled state
Aug 24 12:32:05 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: device vethd69b4c65 left promiscuous mode
Aug 24 12:32:05 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: cni-podman2: port 1(vethd69b4c65) entered disabled state
Aug 24 12:32:05 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[49321]: root : TTY=pts/0 ; PWD=/root ; USER=auth_test_user1 ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-ezgtectxydrzklmrvefcwvynfynwneuk ; XDG_RUNTIME_DIR=/run/user/2001 /usr/libexec/platform-python /var/tmp/ansible-tmp-1724517125.1997957-20990-248056207770319/AnsiballZ_systemd.py'
Aug 24 12:32:05 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[49321]: pam_unix(sudo:session): session opened for user auth_test_user1 by root(uid=0)
Aug 24 12:32:05 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[49324]: ansible-systemd Invoked with daemon_reload=True scope=user daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None user=None
Aug 24 12:32:05 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Reloading.
Aug 24 12:32:05 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[49321]: pam_unix(sudo:session): session closed for user auth_test_user1
Aug 24 12:32:05 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[49458]: root : TTY=pts/0 ; PWD=/root ; USER=auth_test_user1 ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-vqjixfcjzxxcureisbskictotkbcqgsl ; XDG_RUNTIME_DIR=/run/user/2001 /usr/libexec/platform-python /var/tmp/ansible-tmp-1724517125.7649283-21020-115307811609571/AnsiballZ_systemd.py'
Aug 24 12:32:05 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[49458]: pam_unix(sudo:session): session opened for user auth_test_user1 by root(uid=0)
Aug 24 12:32:06 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[49461]: ansible-systemd Invoked with name= scope=user enabled=True daemon_reload=False daemon_reexec=False no_block=False state=None force=None masked=None user=None
Aug 24 12:32:06 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Reloading.
Aug 24 12:32:06 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[49458]: pam_unix(sudo:session): session closed for user auth_test_user1
Aug 24 12:32:06 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[49597]: root : TTY=pts/0 ; PWD=/root ; USER=auth_test_user1 ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-gyttobypdwnvbqfzidqzdzrfvaoifrzt ; XDG_RUNTIME_DIR=/run/user/2001 /usr/libexec/platform-python /var/tmp/ansible-tmp-1724517126.3745768-21029-141501193737783/AnsiballZ_systemd.py'
Aug 24 12:32:06 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[49597]: pam_unix(sudo:session): session opened for user auth_test_user1 by root(uid=0)
Aug 24 12:32:06 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[49600]: ansible-systemd Invoked with name= scope=user state=started daemon_reload=False daemon_reexec=False no_block=False enabled=None force=None masked=None user=None
Aug 24 12:32:06 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Created slice podman\x2dkube.slice.
-- Subject: Unit UNIT has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit UNIT has finished starting up.
--
-- The start-up result is done.
Aug 24 12:32:06 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Starting A template for running K8s workloads via podman-kube-play...
-- Subject: Unit UNIT has begun start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit UNIT has begun starting up.
Aug 24 12:32:06 ip-10-31-40-15.us-east-1.aws.redhat.com podman[49606]: Pods stopped:
Aug 24 12:32:06 ip-10-31-40-15.us-east-1.aws.redhat.com podman[49606]: a8fd13030011bcabb65b997412ad0a15d6982f0a908bdee2d8266896c3183b71
Aug 24 12:32:06 ip-10-31-40-15.us-east-1.aws.redhat.com podman[49606]: Pods removed:
Aug 24 12:32:06 ip-10-31-40-15.us-east-1.aws.redhat.com podman[49606]: a8fd13030011bcabb65b997412ad0a15d6982f0a908bdee2d8266896c3183b71
Aug 24 12:32:06 ip-10-31-40-15.us-east-1.aws.redhat.com podman[49606]: Secrets removed:
Aug 24 12:32:06 ip-10-31-40-15.us-east-1.aws.redhat.com podman[49606]: Volumes removed:
Aug 24 12:32:07 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Started rootless-netns-f870481d.scope.
-- Subject: Unit UNIT has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit UNIT has finished starting up.
--
-- The start-up result is done.
Aug 24 12:32:07 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: IPv6: ADDRCONF(NETDEV_UP): veth30c7d91e: link is not ready
Aug 24 12:32:07 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: cni-podman2: port 1(veth30c7d91e) entered blocking state
Aug 24 12:32:07 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: cni-podman2: port 1(veth30c7d91e) entered disabled state
Aug 24 12:32:07 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: device veth30c7d91e entered promiscuous mode
Aug 24 12:32:07 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: IPv6: ADDRCONF(NETDEV_UP): eth0: link is not ready
Aug 24 12:32:07 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: IPv6: ADDRCONF(NETDEV_CHANGE): eth0: link becomes ready
Aug 24 12:32:07 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: IPv6: ADDRCONF(NETDEV_CHANGE): veth30c7d91e: link becomes ready
Aug 24 12:32:07 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: cni-podman2: port 1(veth30c7d91e) entered blocking state
Aug 24 12:32:07 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: cni-podman2: port 1(veth30c7d91e) entered forwarding state
Aug 24 12:32:07 ip-10-31-40-15.us-east-1.aws.redhat.com dnsmasq[49727]: listening on cni-podman2(#3): 10.89.1.1
Aug 24 12:32:07 ip-10-31-40-15.us-east-1.aws.redhat.com dnsmasq[49729]: started, version 2.79 cachesize 150
Aug 24 12:32:07 ip-10-31-40-15.us-east-1.aws.redhat.com dnsmasq[49729]: compile time options: IPv6 GNU-getopt DBus no-i18n IDN2 DHCP DHCPv6 no-Lua TFTP no-conntrack ipset auth DNSSEC loop-detect inotify
Aug 24 12:32:07 ip-10-31-40-15.us-east-1.aws.redhat.com dnsmasq[49729]: using local addresses only for domain dns.podman
Aug 24 12:32:07 ip-10-31-40-15.us-east-1.aws.redhat.com dnsmasq[49729]: reading /etc/resolv.conf
Aug 24 12:32:07 ip-10-31-40-15.us-east-1.aws.redhat.com dnsmasq[49729]: using local addresses only for domain dns.podman
Aug 24 12:32:07 ip-10-31-40-15.us-east-1.aws.redhat.com dnsmasq[49729]: using nameserver 10.0.2.3#53
Aug 24 12:32:07 ip-10-31-40-15.us-east-1.aws.redhat.com dnsmasq[49729]: using nameserver 10.29.169.13#53
Aug 24 12:32:07 ip-10-31-40-15.us-east-1.aws.redhat.com dnsmasq[49729]: using nameserver 10.29.170.12#53
Aug 24 12:32:07 ip-10-31-40-15.us-east-1.aws.redhat.com dnsmasq[49729]: using nameserver 10.2.32.1#53
Aug 24 12:32:07 ip-10-31-40-15.us-east-1.aws.redhat.com dnsmasq[49729]: read /run/user/2001/containers/cni/dnsname/podman-default-kube-network/addnhosts - 1 addresses
Aug 24 12:32:07 ip-10-31-40-15.us-east-1.aws.redhat.com podman[49606]: Pod:
Aug 24 12:32:07 ip-10-31-40-15.us-east-1.aws.redhat.com podman[49606]: 297f77339fed772a57c23ceaa620fab6c5fb2d8c698c32ef1aa2c3b2467c30e1
Aug 24 12:32:07 ip-10-31-40-15.us-east-1.aws.redhat.com podman[49606]: Container:
Aug 24 12:32:07 ip-10-31-40-15.us-east-1.aws.redhat.com podman[49606]: 3a4bdf2de00a30d3fdd9c342e9c24cef103280a69c6a44caaf41f265a3c3073b
Aug 24 12:32:07 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Started A template for running K8s workloads via podman-kube-play.
-- Subject: Unit UNIT has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit UNIT has finished starting up.
--
-- The start-up result is done.
Aug 24 12:32:07 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[49597]: pam_unix(sudo:session): session closed for user auth_test_user1
Aug 24 12:32:08 ip-10-31-40-15.us-east-1.aws.redhat.com dnsmasq[49729]: listening on cni-podman2(#3): fe80::48be:c4ff:feb2:ce88%cni-podman2
Aug 24 12:32:08 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[50119]: ansible-getent Invoked with database=group key=2001 fail_key=False service=None split=None
Aug 24 12:32:09 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[50330]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1
Aug 24 12:32:09 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[50539]: ansible-command Invoked with _raw_params=getsubids auth_test_user1 warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:32:10 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[50754]: ansible-command Invoked with _raw_params=getsubids -g auth_test_user1 warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:32:11 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[51138]: ansible-command Invoked with creates=/var/lib/systemd/linger/auth_test_user1 _raw_params=loginctl enable-linger auth_test_user1 warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None removes=None stdin=None
Aug 24 12:32:12 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[51469]: root : TTY=pts/0 ; PWD=/root ; USER=auth_test_user1 ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-olbrlxtogvmvtfwqowhzmwejfuiazfub ; XDG_RUNTIME_DIR=/run/user/2001 /usr/libexec/platform-python /var/tmp/ansible-tmp-1724517132.3173523-21358-60429077144067/AnsiballZ_podman_image.py'
Aug 24 12:32:12 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[51469]: pam_unix(sudo:session): session opened for user auth_test_user1 by root(uid=0)
Aug 24 12:32:13 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Started podman-51497.scope.
-- Subject: Unit UNIT has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit UNIT has finished starting up.
--
-- The start-up result is done.
Aug 24 12:32:13 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Started podman-51506.scope.
-- Subject: Unit UNIT has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit UNIT has finished starting up.
--
-- The start-up result is done.
Aug 24 12:32:13 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Started podman-51525.scope.
-- Subject: Unit UNIT has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit UNIT has finished starting up.
--
-- The start-up result is done.
Aug 24 12:32:13 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Started podman-51570.scope.
-- Subject: Unit UNIT has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit UNIT has finished starting up.
--
-- The start-up result is done.
Aug 24 12:32:13 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[51469]: pam_unix(sudo:session): session closed for user auth_test_user1
Aug 24 12:32:13 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[51763]: ansible-file Invoked with path=/home/auth_test_user1/.config/containers/systemd state=directory owner=auth_test_user1 group=2001 mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None
Aug 24 12:32:14 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[52004]: ansible-stat Invoked with path=/home/auth_test_user1/.config/containers/systemd/auth_test_1_quadlet.container follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True
Aug 24 12:32:15 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[52189]: ansible-copy Invoked with src=/root/.ansible/tmp/ansible-tmp-1724517134.1960437-21444-42694398646366/source dest=/home/auth_test_user1/.config/containers/systemd/auth_test_1_quadlet.container owner=auth_test_user1 group=2001 mode=0644 follow=False _original_basename=systemd.j2 checksum=1127036af0ba22b83d33a5a35ab54db2c5855175 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None regexp=None delimiter=None
Aug 24 12:32:16 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[52595]: ansible-getent Invoked with database=group key=2001 fail_key=False service=None split=None
Aug 24 12:32:17 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[52790]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1
Aug 24 12:32:17 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[53020]: ansible-command Invoked with _raw_params=getsubids auth_test_user1 warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:32:18 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[53234]: ansible-command Invoked with _raw_params=getsubids -g auth_test_user1 warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:32:21 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[53863]: ansible-getent Invoked with database=group key=2001 fail_key=False service=None split=None
Aug 24 12:32:21 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[54092]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1
Aug 24 12:32:22 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[54357]: ansible-command Invoked with _raw_params=getsubids auth_test_user1 warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:32:23 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[54587]: ansible-command Invoked with _raw_params=getsubids -g auth_test_user1 warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:32:27 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[55567]: ansible-command Invoked with _raw_params=podman --version warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:32:28 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[55823]: ansible-getent Invoked with database=passwd key=root fail_key=False service=None split=None
Aug 24 12:32:28 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[56046]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None
Aug 24 12:32:29 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[56276]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1
Aug 24 12:32:32 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[56917]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None
Aug 24 12:32:32 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[57122]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1
Aug 24 12:32:34 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[57546]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None
Aug 24 12:32:34 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[57757]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1
Aug 24 12:32:36 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[58223]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None
Aug 24 12:32:37 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[58431]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1
Aug 24 12:32:38 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[58768]: ansible-command Invoked with _raw_params= warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:32:39 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[59053]: ansible-systemd Invoked with name= scope=system state=stopped enabled=False daemon_reload=False daemon_reexec=False no_block=False force=None masked=None user=None
Aug 24 12:32:40 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[59285]: ansible-stat Invoked with path=/etc/containers/ansible-kubernetes.d/auth_test_1_kube.yml follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1
Aug 24 12:32:40 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[59492]: ansible-file Invoked with path=/etc/containers/ansible-kubernetes.d/auth_test_1_kube.yml state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None
Aug 24 12:32:42 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[59929]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None
Aug 24 12:32:42 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[60152]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1
Aug 24 12:32:44 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[60573]: ansible-systemd Invoked with name=auth_test_1_quadlet.service scope=system state=stopped enabled=False force=True daemon_reload=False daemon_reexec=False no_block=False masked=None user=None
Aug 24 12:32:45 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[60796]: ansible-stat Invoked with path=/etc/containers/systemd/auth_test_1_quadlet.container follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1
Aug 24 12:32:45 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[61014]: ansible-file Invoked with path=/etc/containers/systemd/auth_test_1_quadlet.container state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None
Aug 24 12:32:48 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[61558]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None
Aug 24 12:32:48 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[61787]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1
Aug 24 12:32:50 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[62514]: ansible-file Invoked with path=/root/.config/containers state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None
Aug 24 12:32:52 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[62856]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None
Aug 24 12:32:52 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[63074]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1
Aug 24 12:33:00 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[65152]: ansible-command Invoked with _raw_params=podman --version warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:33:01 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[65407]: ansible-getent Invoked with database=passwd key=auth_test_user1 fail_key=False service=None split=None
Aug 24 12:33:02 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[65666]: ansible-getent Invoked with database=group key=2001 fail_key=False service=None split=None
Aug 24 12:33:02 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[65900]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1
Aug 24 12:33:03 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[66109]: ansible-command Invoked with _raw_params=getsubids auth_test_user1 warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:33:03 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[66319]: ansible-command Invoked with _raw_params=getsubids -g auth_test_user1 warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:33:06 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[67038]: ansible-getent Invoked with database=group key=2001 fail_key=False service=None split=None
Aug 24 12:33:07 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[67267]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1
Aug 24 12:33:08 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[67481]: ansible-command Invoked with _raw_params=getsubids auth_test_user1 warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:33:08 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[67691]: ansible-command Invoked with _raw_params=getsubids -g auth_test_user1 warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:33:10 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[68173]: ansible-getent Invoked with database=group key=2001 fail_key=False service=None split=None
Aug 24 12:33:11 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[68399]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1
Aug 24 12:33:11 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[68613]: ansible-command Invoked with _raw_params=getsubids auth_test_user1 warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:33:12 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[68824]: ansible-command Invoked with _raw_params=getsubids -g auth_test_user1 warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:33:14 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[69299]: ansible-getent Invoked with database=group key=2001 fail_key=False service=None split=None
Aug 24 12:33:14 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[69507]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1
Aug 24 12:33:15 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[69743]: ansible-command Invoked with _raw_params=getsubids auth_test_user1 warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:33:16 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[69954]: ansible-command Invoked with _raw_params=getsubids -g auth_test_user1 warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:33:17 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[70238]: ansible-command Invoked with _raw_params= warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:33:17 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[70531]: ansible-stat Invoked with path=/run/user/2001 follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1
Aug 24 12:33:18 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[70763]: root : TTY=pts/0 ; PWD=/root ; USER=auth_test_user1 ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-hpurnidubzilbykhwwpidsvjgctdbqzs ; XDG_RUNTIME_DIR=/run/user/2001 /usr/libexec/platform-python /var/tmp/ansible-tmp-1724517198.2391145-25226-139773372393366/AnsiballZ_systemd.py'
Aug 24 12:33:18 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[70763]: pam_unix(sudo:session): session opened for user auth_test_user1 by root(uid=0)
Aug 24 12:33:18 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[70766]: ansible-systemd Invoked with name= scope=user state=stopped enabled=False daemon_reload=False daemon_reexec=False no_block=False force=None masked=None user=None
Aug 24 12:33:18 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Reloading.
Aug 24 12:33:18 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Stopping A template for running K8s workloads via podman-kube-play...
-- Subject: Unit UNIT has begun shutting down
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit UNIT has begun shutting down.
Aug 24 12:33:19 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: cni-podman2: port 1(veth30c7d91e) entered disabled state
Aug 24 12:33:19 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: device veth30c7d91e left promiscuous mode
Aug 24 12:33:19 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: cni-podman2: port 1(veth30c7d91e) entered disabled state
Aug 24 12:33:19 ip-10-31-40-15.us-east-1.aws.redhat.com podman[70832]: Pods stopped:
Aug 24 12:33:19 ip-10-31-40-15.us-east-1.aws.redhat.com podman[70832]: 297f77339fed772a57c23ceaa620fab6c5fb2d8c698c32ef1aa2c3b2467c30e1
Aug 24 12:33:19 ip-10-31-40-15.us-east-1.aws.redhat.com podman[70832]: Pods removed:
Aug 24 12:33:19 ip-10-31-40-15.us-east-1.aws.redhat.com podman[70832]: 297f77339fed772a57c23ceaa620fab6c5fb2d8c698c32ef1aa2c3b2467c30e1
Aug 24 12:33:19 ip-10-31-40-15.us-east-1.aws.redhat.com podman[70832]: Secrets removed:
Aug 24 12:33:19 ip-10-31-40-15.us-east-1.aws.redhat.com podman[70832]: Volumes removed:
Aug 24 12:33:19 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Stopped A template for running K8s workloads via podman-kube-play.
-- Subject: Unit UNIT has finished shutting down
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit UNIT has finished shutting down.
Aug 24 12:33:19 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[70763]: pam_unix(sudo:session): session closed for user auth_test_user1
Aug 24 12:33:19 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[71095]: ansible-stat Invoked with path=/home/auth_test_user1/.config/containers/ansible-kubernetes.d/auth_test_1_kube.yml follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1
Aug 24 12:33:20 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[71220]: root : TTY=pts/0 ; PWD=/root ; USER=auth_test_user1 ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-nrmnvxwsvdeezidilmvvzjftxotbozoo ; XDG_RUNTIME_DIR=/run/user/2001 /usr/libexec/platform-python /var/tmp/ansible-tmp-1724517200.0476031-25340-149551025494898/AnsiballZ_podman_play.py'
Aug 24 12:33:20 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[71220]: pam_unix(sudo:session): session opened for user auth_test_user1 by root(uid=0)
Aug 24 12:33:20 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[71223]: ansible-containers.podman.podman_play Invoked with state=absent kube_file=/home/auth_test_user1/.config/containers/ansible-kubernetes.d/auth_test_1_kube.yml executable=podman annotation=None authfile=None build=None cert_dir=None configmap=None context_dir=None seccomp_profile_root=None username=None password=NOT_LOGGING_PARAMETER log_driver=None log_opt=None network=None tls_verify=None debug=None quiet=None recreate=None userns=None log_level=None quadlet_dir=None quadlet_filename=None quadlet_options=None
Aug 24 12:33:20 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[71223]: ansible-containers.podman.podman_play version: 4.9.4-dev, kube file /home/auth_test_user1/.config/containers/ansible-kubernetes.d/auth_test_1_kube.yml
Aug 24 12:33:20 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Started podman-71231.scope.
-- Subject: Unit UNIT has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit UNIT has finished starting up.
--
-- The start-up result is done.
Aug 24 12:33:20 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[71220]: pam_unix(sudo:session): session closed for user auth_test_user1
Aug 24 12:33:21 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[71360]: ansible-file Invoked with path=/home/auth_test_user1/.config/containers/ansible-kubernetes.d/auth_test_1_kube.yml state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None
Aug 24 12:33:21 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[71483]: root : TTY=pts/0 ; PWD=/root ; USER=auth_test_user1 ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-komntehjdpgyonyrnjsqzcpuefesrwto ; XDG_RUNTIME_DIR=/run/user/2001 /usr/libexec/platform-python /var/tmp/ansible-tmp-1724517201.3468843-25419-53322898559826/AnsiballZ_command.py'
Aug 24 12:33:21 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[71483]: pam_unix(sudo:session): session opened for user auth_test_user1 by root(uid=0)
Aug 24 12:33:21 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[71486]: ansible-command Invoked with _raw_params=podman image prune -f warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:33:21 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Started podman-71488.scope.
-- Subject: Unit UNIT has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit UNIT has finished starting up.
--
-- The start-up result is done.
Aug 24 12:33:21 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[71483]: pam_unix(sudo:session): session closed for user auth_test_user1
Aug 24 12:33:23 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[71617]: ansible-getent Invoked with database=group key=2001 fail_key=False service=None split=None
Aug 24 12:33:24 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[71741]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1
Aug 24 12:33:24 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[71866]: ansible-command Invoked with _raw_params=getsubids auth_test_user1 warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:33:24 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[71990]: ansible-command Invoked with _raw_params=getsubids -g auth_test_user1 warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:33:26 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[72114]: ansible-stat Invoked with path=/run/user/2001 follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1
Aug 24 12:33:26 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[72239]: root : TTY=pts/0 ; PWD=/root ; USER=auth_test_user1 ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-vlnruqoerlngvryvqackcpgmuhtwqhbn ; XDG_RUNTIME_DIR=/run/user/2001 /usr/libexec/platform-python /var/tmp/ansible-tmp-1724517206.5066495-25749-97258820368346/AnsiballZ_systemd.py'
Aug 24 12:33:26 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[72239]: pam_unix(sudo:session): session opened for user auth_test_user1 by root(uid=0)
Aug 24 12:33:26 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[72242]: ansible-systemd Invoked with name=auth_test_1_quadlet.service scope=user state=stopped enabled=False force=True daemon_reload=False daemon_reexec=False no_block=False masked=None user=None
Aug 24 12:33:27 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Reloading.
Aug 24 12:33:27 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[72239]: pam_unix(sudo:session): session closed for user auth_test_user1
Aug 24 12:33:27 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[72378]: ansible-stat Invoked with path=/home/auth_test_user1/.config/containers/systemd/auth_test_1_quadlet.container follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1
Aug 24 12:33:28 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[72626]: ansible-file Invoked with path=/home/auth_test_user1/.config/containers/systemd/auth_test_1_quadlet.container state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None
Aug 24 12:33:28 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[72749]: root : TTY=pts/0 ; PWD=/root ; USER=auth_test_user1 ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-acokvsmjelcatbnfbypmoricsgnapgxd ; XDG_RUNTIME_DIR=/run/user/2001 /usr/libexec/platform-python /var/tmp/ansible-tmp-1724517208.7611573-25879-195251231871213/AnsiballZ_systemd.py'
Aug 24 12:33:28 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[72749]: pam_unix(sudo:session): session opened for user auth_test_user1 by root(uid=0)
Aug 24 12:33:29 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[72752]: ansible-systemd Invoked with daemon_reload=True scope=user daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None user=None
Aug 24 12:33:29 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Reloading.
Aug 24 12:33:29 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[72749]: pam_unix(sudo:session): session closed for user auth_test_user1
Aug 24 12:33:29 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[72886]: root : TTY=pts/0 ; PWD=/root ; USER=auth_test_user1 ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-ckhmtkhqnlzafleknmfufbzovnrishtl ; XDG_RUNTIME_DIR=/run/user/2001 /usr/libexec/platform-python /var/tmp/ansible-tmp-1724517209.3543239-25915-271137461694598/AnsiballZ_command.py'
Aug 24 12:33:29 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[72886]: pam_unix(sudo:session): session opened for user auth_test_user1 by root(uid=0)
Aug 24 12:33:29 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Started podman-72891.scope.
-- Subject: Unit UNIT has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit UNIT has finished starting up.
--
-- The start-up result is done.
Aug 24 12:33:29 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[72886]: pam_unix(sudo:session): session closed for user auth_test_user1
Aug 24 12:33:31 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[73020]: ansible-stat Invoked with path=/run/user/2001 follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1
Aug 24 12:33:31 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[73145]: root : TTY=pts/0 ; PWD=/root ; USER=auth_test_user1 ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-ocdeozodtlruaphkgintpmlufvdoozeh ; XDG_RUNTIME_DIR=/run/user/2001 /usr/libexec/platform-python /var/tmp/ansible-tmp-1724517211.6403909-26040-73967260498135/AnsiballZ_podman_container_info.py'
Aug 24 12:33:31 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[73145]: pam_unix(sudo:session): session opened for user auth_test_user1 by root(uid=0)
Aug 24 12:33:32 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[73148]: ansible-containers.podman.podman_container_info Invoked with executable=podman name=None
Aug 24 12:33:32 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Started podman-73150.scope.
-- Subject: Unit UNIT has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit UNIT has finished starting up.
--
-- The start-up result is done.
Aug 24 12:33:32 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[73145]: pam_unix(sudo:session): session closed for user auth_test_user1
Aug 24 12:33:32 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[73279]: root : TTY=pts/0 ; PWD=/root ; USER=auth_test_user1 ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-lswpsxdmfeeknwcgchdovupwlnrcgqve ; XDG_RUNTIME_DIR=/run/user/2001 /usr/libexec/platform-python /var/tmp/ansible-tmp-1724517212.2750049-26080-100716737600628/AnsiballZ_command.py'
Aug 24 12:33:32 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[73279]: pam_unix(sudo:session): session opened for user auth_test_user1 by root(uid=0)
Aug 24 12:33:32 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[73282]: ansible-command Invoked with _raw_params=podman network ls -q warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:33:32 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Started podman-73284.scope.
-- Subject: Unit UNIT has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit UNIT has finished starting up.
--
-- The start-up result is done.
Aug 24 12:33:32 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[73279]: pam_unix(sudo:session): session closed for user auth_test_user1
Aug 24 12:33:32 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[73438]: root : TTY=pts/0 ; PWD=/root ; USER=auth_test_user1 ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-khzsauhujhmmwmbstbwuicigahvopfvy ; XDG_RUNTIME_DIR=/run/user/2001 /usr/libexec/platform-python /var/tmp/ansible-tmp-1724517212.7601972-26109-106997805110866/AnsiballZ_command.py'
Aug 24 12:33:32 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[73438]: pam_unix(sudo:session): session opened for user auth_test_user1 by root(uid=0)
Aug 24 12:33:33 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[73441]: ansible-command Invoked with _raw_params=podman secret ls -n -q warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:33:33 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Started podman-73443.scope.
-- Subject: Unit UNIT has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit UNIT has finished starting up.
--
-- The start-up result is done.
Aug 24 12:33:33 ip-10-31-40-15.us-east-1.aws.redhat.com sudo[73438]: pam_unix(sudo:session): session closed for user auth_test_user1
Aug 24 12:33:33 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[73572]: ansible-command Invoked with removes=/var/lib/systemd/linger/auth_test_user1 _raw_params=loginctl disable-linger auth_test_user1 warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None stdin=None
Aug 24 12:33:33 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Stopping User Manager for UID 2001...
-- Subject: Unit user@2001.service has begun shutting down
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit user@2001.service has begun shutting down.
Aug 24 12:33:33 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Removed slice podman\x2dkube.slice.
-- Subject: Unit UNIT has finished shutting down
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit UNIT has finished shutting down.
Aug 24 12:33:33 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Stopped target Default.
-- Subject: Unit UNIT has finished shutting down
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit UNIT has finished shutting down.
Aug 24 12:33:33 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Stopping podman-pause-0e20a755.scope.
-- Subject: Unit UNIT has begun shutting down
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit UNIT has begun shutting down.
Aug 24 12:33:33 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Stopping D-Bus User Message Bus...
-- Subject: Unit UNIT has begun shutting down
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit UNIT has begun shutting down.
Aug 24 12:33:33 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Stopped D-Bus User Message Bus.
-- Subject: Unit UNIT has finished shutting down
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit UNIT has finished shutting down.
Aug 24 12:33:33 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Stopped target Basic System.
-- Subject: Unit UNIT has finished shutting down
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit UNIT has finished shutting down.
Aug 24 12:33:33 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Stopped target Paths.
-- Subject: Unit UNIT has finished shutting down
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit UNIT has finished shutting down.
Aug 24 12:33:33 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Stopped target Sockets.
-- Subject: Unit UNIT has finished shutting down
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit UNIT has finished shutting down.
Aug 24 12:33:33 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Stopped target Timers.
-- Subject: Unit UNIT has finished shutting down
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit UNIT has finished shutting down.
Aug 24 12:33:33 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Stopped Mark boot as successful after the user session has run 2 minutes.
-- Subject: Unit UNIT has finished shutting down
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit UNIT has finished shutting down.
Aug 24 12:33:33 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Closed D-Bus User Message Bus Socket.
-- Subject: Unit UNIT has finished shutting down
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit UNIT has finished shutting down.
Aug 24 12:33:33 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Stopped podman-pause-0e20a755.scope.
-- Subject: Unit UNIT has finished shutting down
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit UNIT has finished shutting down.
Aug 24 12:33:33 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Removed slice user.slice.
-- Subject: Unit UNIT has finished shutting down
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit UNIT has finished shutting down.
Aug 24 12:33:33 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Reached target Shutdown.
-- Subject: Unit UNIT has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit UNIT has finished starting up.
--
-- The start-up result is done.
Aug 24 12:33:33 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Started Exit the Session.
-- Subject: Unit UNIT has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit UNIT has finished starting up.
--
-- The start-up result is done.
Aug 24 12:33:33 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48120]: Reached target Exit the Session.
-- Subject: Unit UNIT has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit UNIT has finished starting up.
--
-- The start-up result is done.
Aug 24 12:33:33 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[48123]: pam_unix(systemd-user:session): session closed for user auth_test_user1
Aug 24 12:33:33 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: user@2001.service: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit user@2001.service has successfully entered the 'dead' state.
Aug 24 12:33:33 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Stopped User Manager for UID 2001.
-- Subject: Unit user@2001.service has finished shutting down
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit user@2001.service has finished shutting down.
Aug 24 12:33:33 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Stopping User runtime directory /run/user/2001...
-- Subject: Unit user-runtime-dir@2001.service has begun shutting down
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit user-runtime-dir@2001.service has begun shutting down.
Aug 24 12:33:33 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: run-user-2001.mount: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit run-user-2001.mount has successfully entered the 'dead' state.
Aug 24 12:33:33 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: user-runtime-dir@2001.service: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit user-runtime-dir@2001.service has successfully entered the 'dead' state.
Aug 24 12:33:33 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Stopped User runtime directory /run/user/2001.
-- Subject: Unit user-runtime-dir@2001.service has finished shutting down
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit user-runtime-dir@2001.service has finished shutting down.
Aug 24 12:33:33 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Removed slice User Slice of UID 2001.
-- Subject: Unit user-2001.slice has finished shutting down
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit user-2001.slice has finished shutting down.
Aug 24 12:33:34 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[73704]: ansible-command Invoked with _raw_params=loginctl show-user -P State auth_test_user1 warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:33:34 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[73828]: ansible-systemd Invoked with name=systemd-logind state=stopped daemon_reload=False daemon_reexec=False no_block=False enabled=None force=None masked=None user=None scope=None
Aug 24 12:33:34 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Stopping Login Service...
-- Subject: Unit systemd-logind.service has begun shutting down
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit systemd-logind.service has begun shutting down.
Aug 24 12:33:34 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: systemd-logind.service: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit systemd-logind.service has successfully entered the 'dead' state.
Aug 24 12:33:34 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Stopped Login Service.
-- Subject: Unit systemd-logind.service has finished shutting down
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit systemd-logind.service has finished shutting down.
Aug 24 12:33:35 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[73957]: ansible-command Invoked with _raw_params=loginctl show-user -P State auth_test_user1 warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:33:36 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[74081]: ansible-getent Invoked with database=group key=2001 fail_key=False service=None split=None
Aug 24 12:33:36 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[74205]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1
Aug 24 12:33:37 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[74330]: ansible-command Invoked with _raw_params=getsubids auth_test_user1 warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:33:37 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[74454]: ansible-command Invoked with _raw_params=getsubids -g auth_test_user1 warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:33:40 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[74824]: ansible-getent Invoked with database=group key=2001 fail_key=False service=None split=None
Aug 24 12:33:41 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[74948]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1
Aug 24 12:33:41 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[75073]: ansible-command Invoked with _raw_params=getsubids auth_test_user1 warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:33:42 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[75197]: ansible-command Invoked with _raw_params=getsubids -g auth_test_user1 warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:33:46 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[75936]: ansible-user Invoked with name=auth_test_user1 state=absent non_unique=False force=False remove=False create_home=True system=False move_home=False append=False ssh_key_bits=0 ssh_key_type=rsa ssh_key_comment=ansible-generated on ip-10-31-40-15.us-east-1.aws.redhat.com update_password=always uid=None group=None groups=None comment=None home=None shell=None password=NOT_LOGGING_PARAMETER login_class=None hidden=None seuser=None skeleton=None generate_ssh_key=None ssh_key_file=None ssh_key_passphrase=NOT_LOGGING_PARAMETER expires=None password_lock=None local=None profile=None authorization=None role=None
Aug 24 12:33:46 ip-10-31-40-15.us-east-1.aws.redhat.com userdel[75940]: delete user 'auth_test_user1'
Aug 24 12:33:46 ip-10-31-40-15.us-east-1.aws.redhat.com userdel[75940]: removed group 'auth_test_user1' owned by 'auth_test_user1'
Aug 24 12:33:46 ip-10-31-40-15.us-east-1.aws.redhat.com userdel[75940]: removed shadow group 'auth_test_user1' owned by 'auth_test_user1'
Aug 24 12:33:46 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[76068]: ansible-file Invoked with path=/home/auth_test_user1 state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None
Aug 24 12:33:47 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[76191]: ansible-command Invoked with _raw_params=podman inspect podman_registry --format '{{range .}}{{range .Mounts}}{{if eq .Type "volume"}}{{.Name}}{{end}}{{end}}{{end}}' warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:33:48 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[76322]: ansible-command Invoked with _raw_params=podman rm -f podman_registry warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:33:48 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-bf8d6a8a130cd57044b05b59250727fd32420dde3068441219e120a629a7ae4d.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-bf8d6a8a130cd57044b05b59250727fd32420dde3068441219e120a629a7ae4d.scope has successfully entered the 'dead' state.
Aug 24 12:33:48 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-bf8d6a8a130cd57044b05b59250727fd32420dde3068441219e120a629a7ae4d.scope: Consumed 319ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-bf8d6a8a130cd57044b05b59250727fd32420dde3068441219e120a629a7ae4d.scope completed and consumed the indicated resources.
Aug 24 12:33:48 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: cni-podman0: port 1(veth85345cee) entered disabled state
Aug 24 12:33:48 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: device veth85345cee left promiscuous mode
Aug 24 12:33:48 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: cni-podman0: port 1(veth85345cee) entered disabled state
Aug 24 12:33:48 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: run-netns-netns\x2dcf1675ac\x2d9007\x2dd2d5\x2d8d1a\x2da8931a022d46.mount: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit run-netns-netns\x2dcf1675ac\x2d9007\x2dd2d5\x2d8d1a\x2da8931a022d46.mount has successfully entered the 'dead' state.
Aug 24 12:33:48 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-bf8d6a8a130cd57044b05b59250727fd32420dde3068441219e120a629a7ae4d-userdata-shm.mount: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit var-lib-containers-storage-overlay\x2dcontainers-bf8d6a8a130cd57044b05b59250727fd32420dde3068441219e120a629a7ae4d-userdata-shm.mount has successfully entered the 'dead' state.
Aug 24 12:33:48 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay-77219d5a585fd51d0e3bdf98787258f43de7656281a74f5b8cd3e9918c22274b-merged.mount: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit var-lib-containers-storage-overlay-77219d5a585fd51d0e3bdf98787258f43de7656281a74f5b8cd3e9918c22274b-merged.mount has successfully entered the 'dead' state.
Aug 24 12:33:48 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state.
Aug 24 12:33:48 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-conmon-bf8d6a8a130cd57044b05b59250727fd32420dde3068441219e120a629a7ae4d.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-conmon-bf8d6a8a130cd57044b05b59250727fd32420dde3068441219e120a629a7ae4d.scope has successfully entered the 'dead' state.
Aug 24 12:33:49 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[76553]: ansible-command Invoked with _raw_params=podman volume rm a24eff084e57a603a42be183d8da720f785cafeabd9b1200393e3034f9f518d6 warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:33:49 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[76683]: ansible-file Invoked with path=/tmp/lsr_iftawaex_podman state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None
Aug 24 12:33:52 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[76845]: ansible-setup Invoked with gather_subset=['all'] gather_timeout=10 filter=* fact_path=/etc/ansible/facts.d
Aug 24 12:33:53 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[77001]: ansible-stat Invoked with path=/run/ostree-booted follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1
Aug 24 12:33:54 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[77124]: ansible-dnf Invoked with name=['python3-pyasn1', 'python3-cryptography', 'python3-dbus'] state=present allow_downgrade=False autoremove=False bugfix=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True lock_timeout=30 conf_file=None disable_excludes=None download_dir=None list=None releasever=None
Aug 24 12:33:57 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[77247]: ansible-dnf Invoked with name=['certmonger'] state=present allow_downgrade=False autoremove=False bugfix=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True lock_timeout=30 conf_file=None disable_excludes=None download_dir=None list=None releasever=None
Aug 24 12:33:59 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[77370]: ansible-file Invoked with name=/etc/certmonger//pre-scripts owner=root group=root mode=0700 state=directory path=/etc/certmonger//pre-scripts recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None
Aug 24 12:34:00 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[77493]: ansible-file Invoked with name=/etc/certmonger//post-scripts owner=root group=root mode=0700 state=directory path=/etc/certmonger//post-scripts recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None
Aug 24 12:34:01 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[77616]: ansible-systemd Invoked with name=certmonger state=started enabled=True daemon_reload=False daemon_reexec=False no_block=False force=None masked=None user=None scope=None
Aug 24 12:34:02 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[77743]: ansible-fedora.linux_system_roles.certificate_request Invoked with name=quadlet_demo dns=['localhost'] directory=/etc/pki/tls wait=True ca=self-sign __header=#
# Ansible managed
#
# system_role:certificate
provider_config_directory=/etc/certmonger provider=certmonger key_usage=['digitalSignature', 'keyEncipherment'] extended_key_usage=['id-kp-serverAuth', 'id-kp-clientAuth'] auto_renew=True ip=None email=None common_name=None country=None state=None locality=None organization=None organizational_unit=None contact_email=None key_size=None owner=None group=None mode=None principal=None run_before=None run_after=None
Aug 24 12:34:02 ip-10-31-40-15.us-east-1.aws.redhat.com certmonger[12107]: 2024-08-24 12:34:02 [12107] Wrote to /var/lib/certmonger/requests/20240824163402
Aug 24 12:34:02 ip-10-31-40-15.us-east-1.aws.redhat.com certmonger[12107]: 2024-08-24 12:34:02 [12107] Wrote to /var/lib/certmonger/requests/20240824163402
Aug 24 12:34:02 ip-10-31-40-15.us-east-1.aws.redhat.com certmonger[12107]: 2024-08-24 12:34:02 [12107] Wrote to /var/lib/certmonger/requests/20240824163402
Aug 24 12:34:02 ip-10-31-40-15.us-east-1.aws.redhat.com certmonger[12107]: 2024-08-24 12:34:02 [12107] Wrote to /var/lib/certmonger/requests/20240824163402
Aug 24 12:34:02 ip-10-31-40-15.us-east-1.aws.redhat.com certmonger[12107]: 2024-08-24 12:34:02 [12107] Wrote to /var/lib/certmonger/requests/20240824163402
Aug 24 12:34:02 ip-10-31-40-15.us-east-1.aws.redhat.com certmonger[12107]: 2024-08-24 12:34:02 [12107] Wrote to /var/lib/certmonger/requests/20240824163402
Aug 24 12:34:02 ip-10-31-40-15.us-east-1.aws.redhat.com certmonger[12107]: 2024-08-24 12:34:02 [12107] Wrote to /var/lib/certmonger/requests/20240824163402
Aug 24 12:34:02 ip-10-31-40-15.us-east-1.aws.redhat.com certmonger[12107]: 2024-08-24 12:34:02 [12107] Wrote to /var/lib/certmonger/requests/20240824163402
Aug 24 12:34:02 ip-10-31-40-15.us-east-1.aws.redhat.com certmonger[12107]: 2024-08-24 12:34:02 [12107] Wrote to /var/lib/certmonger/requests/20240824163402
Aug 24 12:34:02 ip-10-31-40-15.us-east-1.aws.redhat.com certmonger[12107]: 2024-08-24 12:34:02 [12107] Wrote to /var/lib/certmonger/requests/20240824163402
Aug 24 12:34:02 ip-10-31-40-15.us-east-1.aws.redhat.com certmonger[12107]: 2024-08-24 12:34:02 [12107] Wrote to /var/lib/certmonger/requests/20240824163402
Aug 24 12:34:02 ip-10-31-40-15.us-east-1.aws.redhat.com certmonger[12107]: 2024-08-24 12:34:02 [12107] Wrote to /var/lib/certmonger/requests/20240824163402
Aug 24 12:34:02 ip-10-31-40-15.us-east-1.aws.redhat.com certmonger[12107]: 2024-08-24 12:34:02 [12107] Wrote to /var/lib/certmonger/requests/20240824163402
Aug 24 12:34:02 ip-10-31-40-15.us-east-1.aws.redhat.com certmonger[12107]: 2024-08-24 12:34:02 [12107] Wrote to /var/lib/certmonger/requests/20240824163402
Aug 24 12:34:02 ip-10-31-40-15.us-east-1.aws.redhat.com certmonger[12107]: 2024-08-24 12:34:02 [12107] Wrote to /var/lib/certmonger/requests/20240824163402
Aug 24 12:34:02 ip-10-31-40-15.us-east-1.aws.redhat.com certmonger[12107]: 2024-08-24 12:34:02 [12107] Wrote to /var/lib/certmonger/requests/20240824163402
Aug 24 12:34:02 ip-10-31-40-15.us-east-1.aws.redhat.com certmonger[12107]: 2024-08-24 12:34:02 [12107] Wrote to /var/lib/certmonger/requests/20240824163402
Aug 24 12:34:02 ip-10-31-40-15.us-east-1.aws.redhat.com certmonger[12107]: 2024-08-24 12:34:02 [12107] Wrote to /var/lib/certmonger/requests/20240824163402
Aug 24 12:34:02 ip-10-31-40-15.us-east-1.aws.redhat.com certmonger[12107]: 2024-08-24 12:34:02 [12107] Wrote to /var/lib/certmonger/requests/20240824163402
Aug 24 12:34:02 ip-10-31-40-15.us-east-1.aws.redhat.com certmonger[12107]: 2024-08-24 12:34:02 [12107] Wrote to /var/lib/certmonger/requests/20240824163402
Aug 24 12:34:02 ip-10-31-40-15.us-east-1.aws.redhat.com certmonger[12107]: 2024-08-24 12:34:02 [12107] Wrote to /var/lib/certmonger/requests/20240824163402
Aug 24 12:34:02 ip-10-31-40-15.us-east-1.aws.redhat.com certmonger[12107]: 2024-08-24 12:34:02 [12107] Wrote to /var/lib/certmonger/requests/20240824163402
Aug 24 12:34:02 ip-10-31-40-15.us-east-1.aws.redhat.com certmonger[12107]: 2024-08-24 12:34:02 [12107] Wrote to /var/lib/certmonger/requests/20240824163402
Aug 24 12:34:02 ip-10-31-40-15.us-east-1.aws.redhat.com certmonger[12107]: 2024-08-24 12:34:02 [12107] Wrote to /var/lib/certmonger/requests/20240824163402
Aug 24 12:34:02 ip-10-31-40-15.us-east-1.aws.redhat.com certmonger[12107]: 2024-08-24 12:34:02 [12107] Wrote to /var/lib/certmonger/requests/20240824163402
Aug 24 12:34:02 ip-10-31-40-15.us-east-1.aws.redhat.com certmonger[12107]: 2024-08-24 12:34:02 [12107] Wrote to /var/lib/certmonger/requests/20240824163402
Aug 24 12:34:02 ip-10-31-40-15.us-east-1.aws.redhat.com certmonger[77759]: Certificate in file "/etc/pki/tls/certs/quadlet_demo.crt" issued by CA and saved.
Aug 24 12:34:02 ip-10-31-40-15.us-east-1.aws.redhat.com certmonger[12107]: 2024-08-24 12:34:02 [12107] Wrote to /var/lib/certmonger/requests/20240824163402
Aug 24 12:34:03 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[77881]: ansible-slurp Invoked with path=/etc/pki/tls/certs/quadlet_demo.crt src=/etc/pki/tls/certs/quadlet_demo.crt
Aug 24 12:34:03 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[78004]: ansible-slurp Invoked with path=/etc/pki/tls/private/quadlet_demo.key src=/etc/pki/tls/private/quadlet_demo.key
Aug 24 12:34:04 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[78127]: ansible-slurp Invoked with path=/etc/pki/tls/certs/quadlet_demo.crt src=/etc/pki/tls/certs/quadlet_demo.crt
Aug 24 12:34:04 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[78250]: ansible-command Invoked with _raw_params=getcert stop-tracking -f /etc/pki/tls/certs/quadlet_demo.crt warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:34:04 ip-10-31-40-15.us-east-1.aws.redhat.com certmonger[12107]: 2024-08-24 12:34:04 [12107] Wrote to /var/lib/certmonger/requests/20240824163402
Aug 24 12:34:05 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[78374]: ansible-file Invoked with path=/etc/pki/tls/certs/quadlet_demo.crt state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None
Aug 24 12:34:05 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[78497]: ansible-file Invoked with path=/etc/pki/tls/private/quadlet_demo.key state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None
Aug 24 12:34:06 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[78620]: ansible-file Invoked with path=/etc/pki/tls/certs/quadlet_demo.crt state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None
Aug 24 12:34:07 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[78743]: ansible-stat Invoked with path=/run/ostree-booted follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1
Aug 24 12:34:07 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[78866]: ansible-stat Invoked with path=/sbin/transactional-update follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1
Aug 24 12:34:10 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[79114]: ansible-command Invoked with _raw_params=podman --version warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:34:11 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[79243]: ansible-getent Invoked with database=passwd key=root fail_key=False service=None split=None
Aug 24 12:34:12 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[79367]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None
Aug 24 12:34:13 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[79491]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1
Aug 24 12:34:14 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[79616]: ansible-stat Invoked with path=/run/ostree-booted follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1
Aug 24 12:34:15 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[79739]: ansible-stat Invoked with path=/sbin/transactional-update follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1
Aug 24 12:34:16 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[79862]: ansible-dnf Invoked with name=['firewalld'] state=present allow_downgrade=False autoremove=False bugfix=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True lock_timeout=30 conf_file=None disable_excludes=None download_dir=None list=None releasever=None
Aug 24 12:34:19 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[79985]: ansible-systemd Invoked with name=firewalld masked=False daemon_reload=False daemon_reexec=False no_block=False state=None enabled=None force=None user=None scope=None
Aug 24 12:34:19 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[80112]: ansible-systemd Invoked with name=firewalld state=started enabled=True daemon_reload=False daemon_reexec=False no_block=False force=None masked=None user=None scope=None
Aug 24 12:34:19 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Reloading.
Aug 24 12:34:19 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Starting firewalld - dynamic firewall daemon...
-- Subject: Unit firewalld.service has begun start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit firewalld.service has begun starting up.
Aug 24 12:34:20 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started firewalld - dynamic firewall daemon.
-- Subject: Unit firewalld.service has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit firewalld.service has finished starting up.
--
-- The start-up result is done.
Aug 24 12:34:20 ip-10-31-40-15.us-east-1.aws.redhat.com firewalld[80148]: WARNING: AllowZoneDrifting is enabled. This is considered an insecure configuration option. It will be removed in a future release. Please consider disabling it now.
Aug 24 12:34:21 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[80323]: ansible-fedora.linux_system_roles.firewall_lib Invoked with port=['8000/tcp'] permanent=True runtime=True state=enabled __report_changed=True service=[] source_port=[] forward_port=[] rich_rule=[] source=[] interface=[] interface_pci_id=[] icmp_block=[] timeout=0 ipset_entries=[] protocol=[] helper_module=[] destination=[] firewalld_conf=None masquerade=None icmp_block_inversion=None target=None zone=None set_default_zone=None ipset=None ipset_type=None description=None short=None
Aug 24 12:34:21 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[80446]: ansible-fedora.linux_system_roles.firewall_lib Invoked with port=['9000/tcp'] permanent=True runtime=True state=enabled __report_changed=True service=[] source_port=[] forward_port=[] rich_rule=[] source=[] interface=[] interface_pci_id=[] icmp_block=[] timeout=0 ipset_entries=[] protocol=[] helper_module=[] destination=[] firewalld_conf=None masquerade=None icmp_block_inversion=None target=None zone=None set_default_zone=None ipset=None ipset_type=None description=None short=None
Aug 24 12:34:21 ip-10-31-40-15.us-east-1.aws.redhat.com rsyslogd[1022]: imjournal: journal files changed, reloading... [v8.2102.0-15.el8 try https://www.rsyslog.com/e/0 ]
Aug 24 12:34:23 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state.
Aug 24 12:34:24 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state.
Aug 24 12:34:25 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state.
Aug 24 12:34:26 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[81023]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None
Aug 24 12:34:26 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[81147]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1
Aug 24 12:34:28 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[81272]: ansible-file Invoked with path=/etc/containers/systemd state=directory owner=root group=0 mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None
Aug 24 12:34:28 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[81395]: ansible-stat Invoked with path=/etc/containers/systemd/quadlet-demo.network follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True
Aug 24 12:34:29 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[81494]: ansible-copy Invoked with src=/root/.ansible/tmp/ansible-tmp-1724517268.7169204-29276-193597116746162/source dest=/etc/containers/systemd/quadlet-demo.network owner=root group=0 mode=0644 _original_basename=quadlet-demo.network follow=False checksum=e57c08d49aff4bae8daab138d913aeddaa8682a0 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None regexp=None delimiter=None
Aug 24 12:34:29 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[81619]: ansible-systemd Invoked with daemon_reload=True scope=system daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None user=None
Aug 24 12:34:29 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Reloading.
Aug 24 12:34:30 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[81772]: ansible-systemd Invoked with name=quadlet-demo-network.service scope=system state=started daemon_reload=False daemon_reexec=False no_block=False enabled=None force=None masked=None user=None
Aug 24 12:34:30 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Starting quadlet-demo-network.service...
-- Subject: Unit quadlet-demo-network.service has begun start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit quadlet-demo-network.service has begun starting up.
Aug 24 12:34:30 ip-10-31-40-15.us-east-1.aws.redhat.com quadlet-demo-network[81779]: systemd-quadlet-demo
Aug 24 12:34:30 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state.
Aug 24 12:34:30 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started quadlet-demo-network.service.
-- Subject: Unit quadlet-demo-network.service has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit quadlet-demo-network.service has finished starting up.
--
-- The start-up result is done.
Aug 24 12:34:31 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[81933]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None
Aug 24 12:34:32 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[82057]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1
Aug 24 12:34:33 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[82182]: ansible-file Invoked with path=/etc/containers/systemd state=directory owner=root group=0 mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None
Aug 24 12:34:34 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[82305]: ansible-stat Invoked with path=/etc/containers/systemd/quadlet-demo-mysql.volume follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True
Aug 24 12:34:34 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[82404]: ansible-copy Invoked with src=/root/.ansible/tmp/ansible-tmp-1724517273.9811473-29551-138782588482123/source dest=/etc/containers/systemd/quadlet-demo-mysql.volume owner=root group=0 mode=0644 _original_basename=quadlet-demo-mysql.volume follow=False checksum=585f8cbdf0ec73000f9227dcffbef71e9552ea4a backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None regexp=None delimiter=None
Aug 24 12:34:35 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[82529]: ansible-systemd Invoked with daemon_reload=True scope=system daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None user=None
Aug 24 12:34:35 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Reloading.
Aug 24 12:34:35 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[82682]: ansible-systemd Invoked with name=quadlet-demo-mysql-volume.service scope=system state=started daemon_reload=False daemon_reexec=False no_block=False enabled=None force=None masked=None user=None
Aug 24 12:34:35 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Starting quadlet-demo-mysql-volume.service...
-- Subject: Unit quadlet-demo-mysql-volume.service has begun start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit quadlet-demo-mysql-volume.service has begun starting up.
Aug 24 12:34:35 ip-10-31-40-15.us-east-1.aws.redhat.com quadlet-demo-mysql-volume[82689]: systemd-quadlet-demo-mysql
Aug 24 12:34:35 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state.
Aug 24 12:34:35 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started quadlet-demo-mysql-volume.service.
-- Subject: Unit quadlet-demo-mysql-volume.service has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit quadlet-demo-mysql-volume.service has finished starting up.
--
-- The start-up result is done.
Aug 24 12:34:36 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[82819]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None
Aug 24 12:34:37 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[82943]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1
Aug 24 12:34:38 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[83068]: ansible-file Invoked with path=/tmp/quadlet_demo state=directory owner=root group=root mode=0777 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None
Aug 24 12:34:39 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state.
Aug 24 12:34:46 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state.
Aug 24 12:34:46 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state.
Aug 24 12:34:46 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[83411]: ansible-file Invoked with path=/etc/containers/systemd state=directory owner=root group=0 mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None
Aug 24 12:34:46 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[83534]: ansible-stat Invoked with path=/etc/containers/systemd/quadlet-demo-mysql.container follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True
Aug 24 12:34:47 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[83633]: ansible-copy Invoked with dest=/etc/containers/systemd/quadlet-demo-mysql.container owner=root group=0 mode=0644 src=/root/.ansible/tmp/ansible-tmp-1724517286.7033756-30107-189472742671718/source _original_basename=tmpca1ro4lg follow=False checksum=ca62b2ad3cc9afb5b5371ebbf797b9bc4fd7edd4 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None regexp=None delimiter=None
Aug 24 12:34:47 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[83758]: ansible-systemd Invoked with daemon_reload=True scope=system daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None user=None
Aug 24 12:34:47 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Reloading.
Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[83911]: ansible-systemd Invoked with name=quadlet-demo-mysql.service scope=system state=started daemon_reload=False daemon_reexec=False no_block=False enabled=None force=None masked=None user=None
Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Starting quadlet-demo-mysql.service...
-- Subject: Unit quadlet-demo-mysql.service has begun start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit quadlet-demo-mysql.service has begun starting up.
Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: IPv6: ADDRCONF(NETDEV_UP): veth10ec5328: link is not ready
Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: cni-podman2: port 1(veth10ec5328) entered blocking state
Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: cni-podman2: port 1(veth10ec5328) entered disabled state
Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: device veth10ec5328 entered promiscuous mode
Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: cni-podman2: port 1(veth10ec5328) entered blocking state
Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: cni-podman2: port 1(veth10ec5328) entered forwarding state
Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com systemd-udevd[83985]: Using default interface naming scheme 'rhel-8.0'.
Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com systemd-udevd[83985]: link_config: autonegotiation is unset or enabled, the speed and duplex are not writable.
Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com systemd-udevd[83989]: link_config: autonegotiation is unset or enabled, the speed and duplex are not writable.
Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com systemd-udevd[83985]: Could not generate persistent MAC address for cni-podman2: No such file or directory
Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com systemd-udevd[83989]: Could not generate persistent MAC address for veth10ec5328: No such file or directory
Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com NetworkManager[660]: [1724517288.6462] manager: (cni-podman2): new Bridge device (/org/freedesktop/NetworkManager/Devices/8)
Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com NetworkManager[660]: [1724517288.6473] manager: (veth10ec5328): new Veth device (/org/freedesktop/NetworkManager/Devices/9)
Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: IPv6: ADDRCONF(NETDEV_CHANGE): veth10ec5328: link becomes ready
Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com NetworkManager[660]: [1724517288.6545] device (veth10ec5328): carrier: link connected
Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com NetworkManager[660]: [1724517288.6546] device (cni-podman2): carrier: link connected
Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com NetworkManager[660]: [1724517288.6798] device (cni-podman2): state change: unmanaged -> unavailable (reason 'connection-assumed', sys-iface-state: 'external')
Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com NetworkManager[660]: [1724517288.6805] device (cni-podman2): state change: unavailable -> disconnected (reason 'connection-assumed', sys-iface-state: 'external')
Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com NetworkManager[660]: [1724517288.6813] device (cni-podman2): Activation: starting connection 'cni-podman2' (27cdadd2-7388-418d-a98e-53ae883e6b12)
Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com NetworkManager[660]: [1724517288.6814] device (cni-podman2): state change: disconnected -> prepare (reason 'none', sys-iface-state: 'external')
Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com NetworkManager[660]: [1724517288.6817] device (cni-podman2): state change: prepare -> config (reason 'none', sys-iface-state: 'external')
Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com NetworkManager[660]: [1724517288.6818] device (cni-podman2): state change: config -> ip-config (reason 'none', sys-iface-state: 'external')
Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com NetworkManager[660]: [1724517288.6820] device (cni-podman2): state change: ip-config -> ip-check (reason 'none', sys-iface-state: 'external')
Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com dbus-daemon[590]: [system] Activating via systemd: service name='org.freedesktop.nm_dispatcher' unit='dbus-org.freedesktop.nm-dispatcher.service' requested by ':1.5' (uid=0 pid=660 comm="/usr/sbin/NetworkManager --no-daemon " label="system_u:system_r:NetworkManager_t:s0")
Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Starting Network Manager Script Dispatcher Service...
-- Subject: Unit NetworkManager-dispatcher.service has begun start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit NetworkManager-dispatcher.service has begun starting up.
Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com dbus-daemon[590]: [system] Successfully activated service 'org.freedesktop.nm_dispatcher'
Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started Network Manager Script Dispatcher Service.
-- Subject: Unit NetworkManager-dispatcher.service has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit NetworkManager-dispatcher.service has finished starting up.
--
-- The start-up result is done.
Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com NetworkManager[660]: [1724517288.7196] device (cni-podman2): state change: ip-check -> secondaries (reason 'none', sys-iface-state: 'external')
Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com NetworkManager[660]: [1724517288.7198] device (cni-podman2): state change: secondaries -> activated (reason 'none', sys-iface-state: 'external')
Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com NetworkManager[660]: [1724517288.7202] device (cni-podman2): Activation: successful, device activated.
Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com dnsmasq[84059]: listening on cni-podman2(#8): 192.168.30.1
Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com dnsmasq[84071]: started, version 2.79 cachesize 150
Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com dnsmasq[84071]: compile time options: IPv6 GNU-getopt DBus no-i18n IDN2 DHCP DHCPv6 no-Lua TFTP no-conntrack ipset auth DNSSEC loop-detect inotify
Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com dnsmasq[84071]: using local addresses only for domain dns.podman
Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com dnsmasq[84071]: reading /etc/resolv.conf
Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com dnsmasq[84071]: using local addresses only for domain dns.podman
Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com dnsmasq[84071]: using nameserver 10.29.169.13#53
Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com dnsmasq[84071]: using nameserver 10.29.170.12#53
Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com dnsmasq[84071]: using nameserver 10.2.32.1#53
Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com dnsmasq[84071]: read /run/containers/cni/dnsname/systemd-quadlet-demo/addnhosts - 1 addresses
Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started /usr/bin/podman healthcheck run 68250d5fc0218bc6db3ab2ece605ebe15c37a833d3345ad2e2df2c1916dd6cc3.
-- Subject: Unit 68250d5fc0218bc6db3ab2ece605ebe15c37a833d3345ad2e2df2c1916dd6cc3.timer has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit 68250d5fc0218bc6db3ab2ece605ebe15c37a833d3345ad2e2df2c1916dd6cc3.timer has finished starting up.
--
-- The start-up result is done.
Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started quadlet-demo-mysql.service.
-- Subject: Unit quadlet-demo-mysql.service has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit quadlet-demo-mysql.service has finished starting up.
--
-- The start-up result is done.
Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started /usr/bin/podman healthcheck run 68250d5fc0218bc6db3ab2ece605ebe15c37a833d3345ad2e2df2c1916dd6cc3.
-- Subject: Unit 68250d5fc0218bc6db3ab2ece605ebe15c37a833d3345ad2e2df2c1916dd6cc3.service has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit 68250d5fc0218bc6db3ab2ece605ebe15c37a833d3345ad2e2df2c1916dd6cc3.service has finished starting up.
--
-- The start-up result is done.
Aug 24 12:34:48 ip-10-31-40-15.us-east-1.aws.redhat.com quadlet-demo-mysql[83918]: 68250d5fc0218bc6db3ab2ece605ebe15c37a833d3345ad2e2df2c1916dd6cc3
Aug 24 12:34:49 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: 68250d5fc0218bc6db3ab2ece605ebe15c37a833d3345ad2e2df2c1916dd6cc3.service: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit 68250d5fc0218bc6db3ab2ece605ebe15c37a833d3345ad2e2df2c1916dd6cc3.service has successfully entered the 'dead' state.
Aug 24 12:34:49 ip-10-31-40-15.us-east-1.aws.redhat.com dnsmasq[84071]: listening on cni-podman2(#8): fe80::50:afff:fe96:3495%cni-podman2
Aug 24 12:34:49 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[84309]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None
Aug 24 12:34:50 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[84452]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1
Aug 24 12:34:52 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[84577]: ansible-file Invoked with path=/etc/containers/systemd state=directory owner=root group=0 mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None
Aug 24 12:34:52 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[84724]: ansible-stat Invoked with path=/etc/containers/systemd/envoy-proxy-configmap.yml follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True
Aug 24 12:34:52 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[84823]: ansible-copy Invoked with src=/root/.ansible/tmp/ansible-tmp-1724517292.2012258-30396-55356854144861/source dest=/etc/containers/systemd/envoy-proxy-configmap.yml owner=root group=0 mode=0644 _original_basename=envoy-proxy-configmap.yml follow=False checksum=d681c7d56f912150d041873e880818b22a90c188 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None regexp=None delimiter=None
Aug 24 12:34:53 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[84948]: ansible-systemd Invoked with daemon_reload=True scope=system daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None user=None
Aug 24 12:34:53 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Reloading.
Aug 24 12:34:54 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[85129]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None
Aug 24 12:34:55 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[85262]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1
Aug 24 12:34:56 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[85394]: ansible-file Invoked with path=/etc/containers/systemd state=directory owner=root group=0 mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None
Aug 24 12:34:57 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[85517]: ansible-stat Invoked with path=/etc/containers/systemd/quadlet-demo.yml follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True
Aug 24 12:34:57 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[85616]: ansible-copy Invoked with dest=/etc/containers/systemd/quadlet-demo.yml owner=root group=0 mode=0644 src=/root/.ansible/tmp/ansible-tmp-1724517297.0551069-30667-219028600000855/source _original_basename=tmpwlrrnyo9 follow=False checksum=998dccde0483b1654327a46ddd89cbaa47650370 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None regexp=None delimiter=None
Aug 24 12:34:58 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[85741]: ansible-systemd Invoked with daemon_reload=True scope=system daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None user=None
Aug 24 12:34:58 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Reloading.
Aug 24 12:34:58 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: NetworkManager-dispatcher.service: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit NetworkManager-dispatcher.service has successfully entered the 'dead' state.
Aug 24 12:34:59 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[85918]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None
Aug 24 12:34:59 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[86042]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1
Aug 24 12:35:00 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[86167]: ansible-slurp Invoked with path=/etc/containers/systemd/quadlet-demo.yml src=/etc/containers/systemd/quadlet-demo.yml
Aug 24 12:35:01 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[86290]: ansible-file Invoked with path=/tmp/httpd3-create state=directory owner=root group=root recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None
Aug 24 12:35:02 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[86413]: ansible-file Invoked with path=/tmp/httpd3 state=directory owner=root group=root recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None
Aug 24 12:35:19 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started /usr/bin/podman healthcheck run 68250d5fc0218bc6db3ab2ece605ebe15c37a833d3345ad2e2df2c1916dd6cc3.
-- Subject: Unit 68250d5fc0218bc6db3ab2ece605ebe15c37a833d3345ad2e2df2c1916dd6cc3.service has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit 68250d5fc0218bc6db3ab2ece605ebe15c37a833d3345ad2e2df2c1916dd6cc3.service has finished starting up.
--
-- The start-up result is done.
Aug 24 12:35:20 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: run-runc-68250d5fc0218bc6db3ab2ece605ebe15c37a833d3345ad2e2df2c1916dd6cc3-runc.4xgW0s.mount: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit run-runc-68250d5fc0218bc6db3ab2ece605ebe15c37a833d3345ad2e2df2c1916dd6cc3-runc.4xgW0s.mount has successfully entered the 'dead' state.
Aug 24 12:35:20 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: 68250d5fc0218bc6db3ab2ece605ebe15c37a833d3345ad2e2df2c1916dd6cc3.service: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit 68250d5fc0218bc6db3ab2ece605ebe15c37a833d3345ad2e2df2c1916dd6cc3.service has successfully entered the 'dead' state.
Aug 24 12:35:20 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[87030]: ansible-file Invoked with path=/etc/containers/systemd state=directory owner=root group=0 mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None
Aug 24 12:35:21 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[87153]: ansible-stat Invoked with path=/etc/containers/systemd/quadlet-demo.kube follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True
Aug 24 12:35:21 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[87252]: ansible-copy Invoked with src=/root/.ansible/tmp/ansible-tmp-1724517321.03834-31464-224102551766944/source dest=/etc/containers/systemd/quadlet-demo.kube owner=root group=0 mode=0644 _original_basename=quadlet-demo.kube follow=False checksum=7a5c73a5d935a42431c87bcdbeb8a04ed0909dc7 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None regexp=None delimiter=None
Aug 24 12:35:22 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[87377]: ansible-systemd Invoked with daemon_reload=True scope=system daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None user=None
Aug 24 12:35:22 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Reloading.
Aug 24 12:35:22 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[87530]: ansible-systemd Invoked with name=quadlet-demo.service scope=system state=started daemon_reload=False daemon_reexec=False no_block=False enabled=None force=None masked=None user=None
Aug 24 12:35:22 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Starting quadlet-demo.service...
-- Subject: Unit quadlet-demo.service has begun start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit quadlet-demo.service has begun starting up.
Aug 24 12:35:22 ip-10-31-40-15.us-east-1.aws.redhat.com quadlet-demo[87537]: Pods stopped:
Aug 24 12:35:22 ip-10-31-40-15.us-east-1.aws.redhat.com quadlet-demo[87537]: Pods removed:
Aug 24 12:35:22 ip-10-31-40-15.us-east-1.aws.redhat.com quadlet-demo[87537]: Secrets removed:
Aug 24 12:35:22 ip-10-31-40-15.us-east-1.aws.redhat.com quadlet-demo[87537]: Volumes removed:
Aug 24 12:35:22 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Created slice cgroup machine-libpod_pod_bbca104112c49d0abd527ad6773ff340419b9009bed96d17f6799e23cb2e88d4.slice.
-- Subject: Unit machine-libpod_pod_bbca104112c49d0abd527ad6773ff340419b9009bed96d17f6799e23cb2e88d4.slice has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit machine-libpod_pod_bbca104112c49d0abd527ad6773ff340419b9009bed96d17f6799e23cb2e88d4.slice has finished starting up.
--
-- The start-up result is done.
Aug 24 12:35:23 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container a3a21fc2dc266efda8f166acd56c9ec75e90514a1f55d114243efd101b6d363e.
-- Subject: Unit libpod-a3a21fc2dc266efda8f166acd56c9ec75e90514a1f55d114243efd101b6d363e.scope has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit libpod-a3a21fc2dc266efda8f166acd56c9ec75e90514a1f55d114243efd101b6d363e.scope has finished starting up.
--
-- The start-up result is done.
Aug 24 12:35:23 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: IPv6: ADDRCONF(NETDEV_UP): veth1c0859f2: link is not ready
Aug 24 12:35:23 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: cni-podman2: port 2(veth1c0859f2) entered blocking state
Aug 24 12:35:23 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: cni-podman2: port 2(veth1c0859f2) entered disabled state
Aug 24 12:35:23 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: device veth1c0859f2 entered promiscuous mode
Aug 24 12:35:23 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: cni-podman2: port 2(veth1c0859f2) entered blocking state
Aug 24 12:35:23 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: cni-podman2: port 2(veth1c0859f2) entered forwarding state
Aug 24 12:35:23 ip-10-31-40-15.us-east-1.aws.redhat.com NetworkManager[660]: [1724517323.1746] manager: (veth1c0859f2): new Veth device (/org/freedesktop/NetworkManager/Devices/10)
Aug 24 12:35:23 ip-10-31-40-15.us-east-1.aws.redhat.com systemd-udevd[87625]: link_config: autonegotiation is unset or enabled, the speed and duplex are not writable.
Aug 24 12:35:23 ip-10-31-40-15.us-east-1.aws.redhat.com systemd-udevd[87625]: Could not generate persistent MAC address for veth1c0859f2: No such file or directory
Aug 24 12:35:23 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: IPv6: ADDRCONF(NETDEV_CHANGE): veth1c0859f2: link becomes ready
Aug 24 12:35:23 ip-10-31-40-15.us-east-1.aws.redhat.com NetworkManager[660]: [1724517323.1833] device (veth1c0859f2): carrier: link connected
Aug 24 12:35:23 ip-10-31-40-15.us-east-1.aws.redhat.com dnsmasq[84071]: read /run/containers/cni/dnsname/systemd-quadlet-demo/addnhosts - 2 addresses
Aug 24 12:35:23 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container c14ac2390f14575fc25c4d23b475705801a8d72451bc5ef8674c274401e2fc71.
-- Subject: Unit libpod-c14ac2390f14575fc25c4d23b475705801a8d72451bc5ef8674c274401e2fc71.scope has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit libpod-c14ac2390f14575fc25c4d23b475705801a8d72451bc5ef8674c274401e2fc71.scope has finished starting up.
--
-- The start-up result is done.
Aug 24 12:35:23 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 3ecc5b442f005a249ca7a72e1a6c6e613010d0c5c989a576b5aea86daadc9e6c.
-- Subject: Unit libpod-3ecc5b442f005a249ca7a72e1a6c6e613010d0c5c989a576b5aea86daadc9e6c.scope has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit libpod-3ecc5b442f005a249ca7a72e1a6c6e613010d0c5c989a576b5aea86daadc9e6c.scope has finished starting up.
--
-- The start-up result is done.
Aug 24 12:35:23 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 14fe7e5a1bc45af6ba6a3d0270a844a86eaa844a3fdfb935158304797b010f6e.
-- Subject: Unit libpod-14fe7e5a1bc45af6ba6a3d0270a844a86eaa844a3fdfb935158304797b010f6e.scope has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit libpod-14fe7e5a1bc45af6ba6a3d0270a844a86eaa844a3fdfb935158304797b010f6e.scope has finished starting up.
--
-- The start-up result is done.
Aug 24 12:35:23 ip-10-31-40-15.us-east-1.aws.redhat.com quadlet-demo[87537]: Volumes:
Aug 24 12:35:23 ip-10-31-40-15.us-east-1.aws.redhat.com quadlet-demo[87537]: wp-pv-claim
Aug 24 12:35:23 ip-10-31-40-15.us-east-1.aws.redhat.com quadlet-demo[87537]: Pod:
Aug 24 12:35:23 ip-10-31-40-15.us-east-1.aws.redhat.com quadlet-demo[87537]: bbca104112c49d0abd527ad6773ff340419b9009bed96d17f6799e23cb2e88d4
Aug 24 12:35:23 ip-10-31-40-15.us-east-1.aws.redhat.com quadlet-demo[87537]: Containers:
Aug 24 12:35:23 ip-10-31-40-15.us-east-1.aws.redhat.com quadlet-demo[87537]: 3ecc5b442f005a249ca7a72e1a6c6e613010d0c5c989a576b5aea86daadc9e6c
Aug 24 12:35:23 ip-10-31-40-15.us-east-1.aws.redhat.com quadlet-demo[87537]: 14fe7e5a1bc45af6ba6a3d0270a844a86eaa844a3fdfb935158304797b010f6e
Aug 24 12:35:23 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started quadlet-demo.service.
-- Subject: Unit quadlet-demo.service has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit quadlet-demo.service has finished starting up.
--
-- The start-up result is done.
Aug 24 12:35:24 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[87981]: ansible-command Invoked with _raw_params=ls -alrtF /etc/containers/systemd warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:35:24 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[88178]: ansible-command Invoked with _raw_params=podman ps -a warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:35:25 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[88350]: ansible-command Invoked with _raw_params=podman volume ls warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:35:25 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[88480]: ansible-command Invoked with _raw_params=podman pod ps --ctr-ids --ctr-names --ctr-status warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:35:26 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[88611]: ansible-command Invoked with _raw_params=set -euo pipefail; systemctl list-units | grep quadlet _uses_shell=True warn=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:35:26 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[88737]: ansible-get_url Invoked with url=https://localhost:8000 dest=/run/out mode=0600 validate_certs=False force=False http_agent=ansible-httpget use_proxy=True force_basic_auth=False sha256sum= checksum= timeout=10 follow=False unsafe_writes=False url_username=None url_password=NOT_LOGGING_PARAMETER client_cert=None client_key=None backup=None headers=None tmp_dest=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None src=None content=NOT_LOGGING_PARAMETER remote_src=None regexp=None delimiter=None directory_mode=None
Aug 24 12:35:27 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[88862]: ansible-command Invoked with _raw_params=cat /run/out warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:35:27 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[88986]: ansible-command Invoked with _raw_params=podman ps -a warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:35:28 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[89118]: ansible-command Invoked with _raw_params=podman pod ps --ctr-ids --ctr-names --ctr-status warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:35:28 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[89249]: ansible-command Invoked with _raw_params=set -euo pipefail; systemctl list-units --all | grep quadlet _uses_shell=True warn=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:35:29 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[89375]: ansible-command Invoked with _raw_params=ls -alrtF /etc/systemd/system warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:35:31 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[89624]: ansible-command Invoked with _raw_params=podman --version warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:35:32 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[89753]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None
Aug 24 12:35:32 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[89877]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1
Aug 24 12:35:35 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[90002]: ansible-dnf Invoked with name=['firewalld'] state=present allow_downgrade=False autoremove=False bugfix=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True lock_timeout=30 conf_file=None disable_excludes=None download_dir=None list=None releasever=None
Aug 24 12:35:38 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[90125]: ansible-systemd Invoked with name=firewalld masked=False daemon_reload=False daemon_reexec=False no_block=False state=None enabled=None force=None user=None scope=None
Aug 24 12:35:38 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[90252]: ansible-systemd Invoked with name=firewalld state=started enabled=True daemon_reload=False daemon_reexec=False no_block=False force=None masked=None user=None scope=None
Aug 24 12:35:39 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[90379]: ansible-fedora.linux_system_roles.firewall_lib Invoked with port=['8000/tcp'] permanent=True runtime=True state=enabled __report_changed=True service=[] source_port=[] forward_port=[] rich_rule=[] source=[] interface=[] interface_pci_id=[] icmp_block=[] timeout=0 ipset_entries=[] protocol=[] helper_module=[] destination=[] firewalld_conf=None masquerade=None icmp_block_inversion=None target=None zone=None set_default_zone=None ipset=None ipset_type=None description=None short=None
Aug 24 12:35:39 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[90502]: ansible-fedora.linux_system_roles.firewall_lib Invoked with port=['9000/tcp'] permanent=True runtime=True state=enabled __report_changed=True service=[] source_port=[] forward_port=[] rich_rule=[] source=[] interface=[] interface_pci_id=[] icmp_block=[] timeout=0 ipset_entries=[] protocol=[] helper_module=[] destination=[] firewalld_conf=None masquerade=None icmp_block_inversion=None target=None zone=None set_default_zone=None ipset=None ipset_type=None description=None short=None
Aug 24 12:35:43 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[91015]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None
Aug 24 12:35:44 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[91139]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1
Aug 24 12:35:45 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[91264]: ansible-systemd Invoked with name=quadlet-demo.service scope=system state=stopped enabled=False force=True daemon_reload=False daemon_reexec=False no_block=False masked=None user=None
Aug 24 12:35:45 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Reloading.
Aug 24 12:35:45 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Stopping quadlet-demo.service...
-- Subject: Unit quadlet-demo.service has begun shutting down
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit quadlet-demo.service has begun shutting down.
Aug 24 12:35:45 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-a3a21fc2dc266efda8f166acd56c9ec75e90514a1f55d114243efd101b6d363e.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-a3a21fc2dc266efda8f166acd56c9ec75e90514a1f55d114243efd101b6d363e.scope has successfully entered the 'dead' state.
Aug 24 12:35:45 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-a3a21fc2dc266efda8f166acd56c9ec75e90514a1f55d114243efd101b6d363e.scope: Consumed 35ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-a3a21fc2dc266efda8f166acd56c9ec75e90514a1f55d114243efd101b6d363e.scope completed and consumed the indicated resources.
Aug 24 12:35:45 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-a3a21fc2dc266efda8f166acd56c9ec75e90514a1f55d114243efd101b6d363e-userdata-shm.mount: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit var-lib-containers-storage-overlay\x2dcontainers-a3a21fc2dc266efda8f166acd56c9ec75e90514a1f55d114243efd101b6d363e-userdata-shm.mount has successfully entered the 'dead' state.
Aug 24 12:35:45 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay-a98d97c31be0232a5dea2f79c3b27620993f5e7c2943a034f4688bd8808862df-merged.mount: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit var-lib-containers-storage-overlay-a98d97c31be0232a5dea2f79c3b27620993f5e7c2943a034f4688bd8808862df-merged.mount has successfully entered the 'dead' state.
Aug 24 12:35:47 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-3ecc5b442f005a249ca7a72e1a6c6e613010d0c5c989a576b5aea86daadc9e6c.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-3ecc5b442f005a249ca7a72e1a6c6e613010d0c5c989a576b5aea86daadc9e6c.scope has successfully entered the 'dead' state.
Aug 24 12:35:47 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-3ecc5b442f005a249ca7a72e1a6c6e613010d0c5c989a576b5aea86daadc9e6c.scope: Consumed 1.056s CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-3ecc5b442f005a249ca7a72e1a6c6e613010d0c5c989a576b5aea86daadc9e6c.scope completed and consumed the indicated resources.
Aug 24 12:35:47 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-14fe7e5a1bc45af6ba6a3d0270a844a86eaa844a3fdfb935158304797b010f6e.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-14fe7e5a1bc45af6ba6a3d0270a844a86eaa844a3fdfb935158304797b010f6e.scope has successfully entered the 'dead' state.
Aug 24 12:35:47 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-14fe7e5a1bc45af6ba6a3d0270a844a86eaa844a3fdfb935158304797b010f6e.scope: Consumed 121ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-14fe7e5a1bc45af6ba6a3d0270a844a86eaa844a3fdfb935158304797b010f6e.scope completed and consumed the indicated resources.
Aug 24 12:35:47 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay-a76eb00d8a3761c8d189a1ddebc0f043c3aa2c138bd8fa985fd3c1fc009a9aab-merged.mount: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit var-lib-containers-storage-overlay-a76eb00d8a3761c8d189a1ddebc0f043c3aa2c138bd8fa985fd3c1fc009a9aab-merged.mount has successfully entered the 'dead' state.
Aug 24 12:35:47 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay-fcf523d74cdb47e960956221b411933c2ce38983b3a6dcb9cceccb4b004baa54-merged.mount: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit var-lib-containers-storage-overlay-fcf523d74cdb47e960956221b411933c2ce38983b3a6dcb9cceccb4b004baa54-merged.mount has successfully entered the 'dead' state.
Aug 24 12:35:47 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-c14ac2390f14575fc25c4d23b475705801a8d72451bc5ef8674c274401e2fc71.scope: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-c14ac2390f14575fc25c4d23b475705801a8d72451bc5ef8674c274401e2fc71.scope has successfully entered the 'dead' state.
Aug 24 12:35:47 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: libpod-c14ac2390f14575fc25c4d23b475705801a8d72451bc5ef8674c274401e2fc71.scope: Consumed 31ms CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit libpod-c14ac2390f14575fc25c4d23b475705801a8d72451bc5ef8674c274401e2fc71.scope completed and consumed the indicated resources.
Aug 24 12:35:47 ip-10-31-40-15.us-east-1.aws.redhat.com dnsmasq[84071]: read /run/containers/cni/dnsname/systemd-quadlet-demo/addnhosts - 1 addresses
Aug 24 12:35:47 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: cni-podman2: port 2(veth1c0859f2) entered disabled state
Aug 24 12:35:47 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: device veth1c0859f2 left promiscuous mode
Aug 24 12:35:47 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: cni-podman2: port 2(veth1c0859f2) entered disabled state
Aug 24 12:35:47 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: run-netns-netns\x2df8ada618\x2d4309\x2d2e92\x2d2bb3\x2d8c587e2dd2e5.mount: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit run-netns-netns\x2df8ada618\x2d4309\x2d2e92\x2d2bb3\x2d8c587e2dd2e5.mount has successfully entered the 'dead' state.
Aug 24 12:35:47 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-c14ac2390f14575fc25c4d23b475705801a8d72451bc5ef8674c274401e2fc71-userdata-shm.mount: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit var-lib-containers-storage-overlay\x2dcontainers-c14ac2390f14575fc25c4d23b475705801a8d72451bc5ef8674c274401e2fc71-userdata-shm.mount has successfully entered the 'dead' state.
Aug 24 12:35:47 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay-3dce984c4dfcb31a9a28077551ecae2ac7b04fc9ddd6862a8083fe91e9189f7e-merged.mount: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit var-lib-containers-storage-overlay-3dce984c4dfcb31a9a28077551ecae2ac7b04fc9ddd6862a8083fe91e9189f7e-merged.mount has successfully entered the 'dead' state.
Aug 24 12:35:47 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Removed slice cgroup machine-libpod_pod_bbca104112c49d0abd527ad6773ff340419b9009bed96d17f6799e23cb2e88d4.slice.
-- Subject: Unit machine-libpod_pod_bbca104112c49d0abd527ad6773ff340419b9009bed96d17f6799e23cb2e88d4.slice has finished shutting down
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit machine-libpod_pod_bbca104112c49d0abd527ad6773ff340419b9009bed96d17f6799e23cb2e88d4.slice has finished shutting down.
Aug 24 12:35:47 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: machine-libpod_pod_bbca104112c49d0abd527ad6773ff340419b9009bed96d17f6799e23cb2e88d4.slice: Consumed 1.210s CPU time
-- Subject: Resources consumed by unit runtime
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit machine-libpod_pod_bbca104112c49d0abd527ad6773ff340419b9009bed96d17f6799e23cb2e88d4.slice completed and consumed the indicated resources.
Aug 24 12:35:48 ip-10-31-40-15.us-east-1.aws.redhat.com quadlet-demo[91314]: Pods stopped:
Aug 24 12:35:48 ip-10-31-40-15.us-east-1.aws.redhat.com quadlet-demo[91314]: bbca104112c49d0abd527ad6773ff340419b9009bed96d17f6799e23cb2e88d4
Aug 24 12:35:48 ip-10-31-40-15.us-east-1.aws.redhat.com quadlet-demo[91314]: Pods removed:
Aug 24 12:35:48 ip-10-31-40-15.us-east-1.aws.redhat.com quadlet-demo[91314]: bbca104112c49d0abd527ad6773ff340419b9009bed96d17f6799e23cb2e88d4
Aug 24 12:35:48 ip-10-31-40-15.us-east-1.aws.redhat.com quadlet-demo[91314]: Secrets removed:
Aug 24 12:35:48 ip-10-31-40-15.us-east-1.aws.redhat.com quadlet-demo[91314]: Volumes removed:
Aug 24 12:35:48 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: quadlet-demo.service: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit quadlet-demo.service has successfully entered the 'dead' state.
Aug 24 12:35:48 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Stopped quadlet-demo.service.
-- Subject: Unit quadlet-demo.service has finished shutting down
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit quadlet-demo.service has finished shutting down.
Aug 24 12:35:48 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[91620]: ansible-stat Invoked with path=/etc/containers/systemd/quadlet-demo.kube follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1
Aug 24 12:35:49 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[91868]: ansible-file Invoked with path=/etc/containers/systemd/quadlet-demo.kube state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None
Aug 24 12:35:50 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[91991]: ansible-systemd Invoked with daemon_reload=True scope=system daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None user=None
Aug 24 12:35:50 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Reloading.
Aug 24 12:35:50 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[92144]: ansible-command Invoked with _raw_params=podman image prune --all -f warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:35:50 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started /usr/bin/podman healthcheck run 68250d5fc0218bc6db3ab2ece605ebe15c37a833d3345ad2e2df2c1916dd6cc3.
-- Subject: Unit 68250d5fc0218bc6db3ab2ece605ebe15c37a833d3345ad2e2df2c1916dd6cc3.service has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit 68250d5fc0218bc6db3ab2ece605ebe15c37a833d3345ad2e2df2c1916dd6cc3.service has finished starting up.
--
-- The start-up result is done.
Aug 24 12:35:51 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: 68250d5fc0218bc6db3ab2ece605ebe15c37a833d3345ad2e2df2c1916dd6cc3.service: Main process exited, code=exited, status=125/n/a
Aug 24 12:35:51 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: 68250d5fc0218bc6db3ab2ece605ebe15c37a833d3345ad2e2df2c1916dd6cc3.service: Failed with result 'exit-code'.
-- Subject: Unit failed
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit 68250d5fc0218bc6db3ab2ece605ebe15c37a833d3345ad2e2df2c1916dd6cc3.service has entered the 'failed' state with result 'exit-code'.
Aug 24 12:35:51 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[92284]: ansible-command Invoked with _raw_params=podman images -n warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:35:52 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[92415]: ansible-command Invoked with _raw_params=podman volume ls -n warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:35:52 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[92545]: ansible-command Invoked with _raw_params=podman ps --noheading warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:35:53 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[92675]: ansible-command Invoked with _raw_params=podman network ls -n -q warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:35:54 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[92987]: ansible-service_facts Invoked
Aug 24 12:35:56 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[93199]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None
Aug 24 12:35:57 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[93323]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1
Aug 24 12:35:58 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[93448]: ansible-stat Invoked with path=/etc/containers/systemd/quadlet-demo.yml follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1
Aug 24 12:35:59 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[93696]: ansible-file Invoked with path=/etc/containers/systemd/quadlet-demo.yml state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None
Aug 24 12:35:59 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[93819]: ansible-systemd Invoked with daemon_reload=True scope=system daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None user=None
Aug 24 12:35:59 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Reloading.
Aug 24 12:36:00 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[93972]: ansible-command Invoked with _raw_params=podman image prune --all -f warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:36:00 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[94102]: ansible-command Invoked with _raw_params=podman images -n warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:36:01 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[94232]: ansible-command Invoked with _raw_params=podman volume ls -n warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:36:01 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[94362]: ansible-command Invoked with _raw_params=podman ps --noheading warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:36:01 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[94492]: ansible-command Invoked with _raw_params=podman network ls -n -q warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:36:02 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[94802]: ansible-service_facts Invoked
Aug 24 12:36:04 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[95014]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None
Aug 24 12:36:05 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[95138]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1
Aug 24 12:36:06 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[95263]: ansible-stat Invoked with path=/etc/containers/systemd/envoy-proxy-configmap.yml follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1
Aug 24 12:36:07 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[95511]: ansible-file Invoked with path=/etc/containers/systemd/envoy-proxy-configmap.yml state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None
Aug 24 12:36:07 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[95634]: ansible-systemd Invoked with daemon_reload=True scope=system daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None user=None
Aug 24 12:36:07 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Reloading.
Aug 24 12:36:08 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[95787]: ansible-command Invoked with _raw_params=podman image prune --all -f warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:36:08 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[95917]: ansible-command Invoked with _raw_params=podman images -n warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:36:09 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[96047]: ansible-command Invoked with _raw_params=podman volume ls -n warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:36:09 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[96177]: ansible-command Invoked with _raw_params=podman ps --noheading warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:36:10 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[96307]: ansible-command Invoked with _raw_params=podman network ls -n -q warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:36:10 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[96618]: ansible-service_facts Invoked
Aug 24 12:36:12 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[96830]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None
Aug 24 12:36:13 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[96954]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1
Aug 24 12:36:14 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[97079]: ansible-systemd Invoked with name=quadlet-demo-mysql.service scope=system state=stopped enabled=False force=True daemon_reload=False daemon_reexec=False no_block=False masked=None user=None
Aug 24 12:36:14 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Reloading.
Aug 24 12:36:14 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Stopping quadlet-demo-mysql.service...
-- Subject: Unit quadlet-demo-mysql.service has begun shutting down
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit quadlet-demo-mysql.service has begun shutting down.
Aug 24 12:36:15 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: 68250d5fc0218bc6db3ab2ece605ebe15c37a833d3345ad2e2df2c1916dd6cc3.timer: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit 68250d5fc0218bc6db3ab2ece605ebe15c37a833d3345ad2e2df2c1916dd6cc3.timer has successfully entered the 'dead' state.
Aug 24 12:36:15 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Stopped /usr/bin/podman healthcheck run 68250d5fc0218bc6db3ab2ece605ebe15c37a833d3345ad2e2df2c1916dd6cc3.
-- Subject: Unit 68250d5fc0218bc6db3ab2ece605ebe15c37a833d3345ad2e2df2c1916dd6cc3.timer has finished shutting down
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit 68250d5fc0218bc6db3ab2ece605ebe15c37a833d3345ad2e2df2c1916dd6cc3.timer has finished shutting down.
Aug 24 12:36:15 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: cni-podman2: port 1(veth10ec5328) entered disabled state
Aug 24 12:36:15 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: device veth10ec5328 left promiscuous mode
Aug 24 12:36:15 ip-10-31-40-15.us-east-1.aws.redhat.com kernel: cni-podman2: port 1(veth10ec5328) entered disabled state
Aug 24 12:36:15 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: run-netns-netns\x2d700346f0\x2dda9a\x2d65a2\x2d3a64\x2d3b6e5ef382b2.mount: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit run-netns-netns\x2d700346f0\x2dda9a\x2d65a2\x2d3a64\x2d3b6e5ef382b2.mount has successfully entered the 'dead' state.
Aug 24 12:36:15 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-68250d5fc0218bc6db3ab2ece605ebe15c37a833d3345ad2e2df2c1916dd6cc3-userdata-shm.mount: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit var-lib-containers-storage-overlay\x2dcontainers-68250d5fc0218bc6db3ab2ece605ebe15c37a833d3345ad2e2df2c1916dd6cc3-userdata-shm.mount has successfully entered the 'dead' state.
Aug 24 12:36:15 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay-11b1e288ec3f02575f7a1a20a99b775e3487c1852d9248f0536b48f52045d256-merged.mount: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit var-lib-containers-storage-overlay-11b1e288ec3f02575f7a1a20a99b775e3487c1852d9248f0536b48f52045d256-merged.mount has successfully entered the 'dead' state.
Aug 24 12:36:15 ip-10-31-40-15.us-east-1.aws.redhat.com quadlet-demo-mysql[97115]: 68250d5fc0218bc6db3ab2ece605ebe15c37a833d3345ad2e2df2c1916dd6cc3
Aug 24 12:36:15 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state.
Aug 24 12:36:15 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: quadlet-demo-mysql.service: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit quadlet-demo-mysql.service has successfully entered the 'dead' state.
Aug 24 12:36:15 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Stopped quadlet-demo-mysql.service.
-- Subject: Unit quadlet-demo-mysql.service has finished shutting down
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit quadlet-demo-mysql.service has finished shutting down.
Aug 24 12:36:16 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[97363]: ansible-stat Invoked with path=/etc/containers/systemd/quadlet-demo-mysql.container follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1
Aug 24 12:36:17 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[97611]: ansible-file Invoked with path=/etc/containers/systemd/quadlet-demo-mysql.container state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None
Aug 24 12:36:17 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[97734]: ansible-systemd Invoked with daemon_reload=True scope=system daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None user=None
Aug 24 12:36:17 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Reloading.
Aug 24 12:36:18 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[98019]: ansible-command Invoked with _raw_params=podman image prune --all -f warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:36:18 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state.
Aug 24 12:36:19 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[98150]: ansible-command Invoked with _raw_params=podman images -n warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:36:19 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state.
Aug 24 12:36:19 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[98280]: ansible-command Invoked with _raw_params=podman volume ls -n warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:36:19 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state.
Aug 24 12:36:20 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[98411]: ansible-command Invoked with _raw_params=podman ps --noheading warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:36:20 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[98541]: ansible-command Invoked with _raw_params=podman network ls -n -q warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:36:20 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state.
Aug 24 12:36:21 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state.
Aug 24 12:36:21 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[98853]: ansible-service_facts Invoked
Aug 24 12:36:23 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[99064]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None
Aug 24 12:36:23 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[99188]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1
Aug 24 12:36:25 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[99313]: ansible-systemd Invoked with name=quadlet-demo-mysql-volume.service scope=system state=stopped enabled=False force=True daemon_reload=False daemon_reexec=False no_block=False masked=None user=None
Aug 24 12:36:25 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Reloading.
Aug 24 12:36:25 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: quadlet-demo-mysql-volume.service: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit quadlet-demo-mysql-volume.service has successfully entered the 'dead' state.
Aug 24 12:36:25 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Stopped quadlet-demo-mysql-volume.service.
-- Subject: Unit quadlet-demo-mysql-volume.service has finished shutting down
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit quadlet-demo-mysql-volume.service has finished shutting down.
Aug 24 12:36:25 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[99470]: ansible-stat Invoked with path=/etc/containers/systemd/quadlet-demo-mysql.volume follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1
Aug 24 12:36:26 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[99718]: ansible-file Invoked with path=/etc/containers/systemd/quadlet-demo-mysql.volume state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None
Aug 24 12:36:27 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[99841]: ansible-systemd Invoked with daemon_reload=True scope=system daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None user=None
Aug 24 12:36:27 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Reloading.
Aug 24 12:36:27 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state.
Aug 24 12:36:28 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[100125]: ansible-command Invoked with _raw_params=podman image prune --all -f warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:36:28 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[100256]: ansible-command Invoked with _raw_params=podman images -n warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:36:28 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[100386]: ansible-command Invoked with _raw_params=podman volume ls -n warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:36:29 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[100517]: ansible-command Invoked with _raw_params=podman ps --noheading warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:36:29 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[100647]: ansible-command Invoked with _raw_params=podman network ls -n -q warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:36:29 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state.
Aug 24 12:36:30 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[100957]: ansible-service_facts Invoked
Aug 24 12:36:32 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[101168]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None
Aug 24 12:36:33 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[101292]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1
Aug 24 12:36:34 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[101417]: ansible-systemd Invoked with name=quadlet-demo-network.service scope=system state=stopped enabled=False force=True daemon_reload=False daemon_reexec=False no_block=False masked=None user=None
Aug 24 12:36:34 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Reloading.
Aug 24 12:36:34 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: quadlet-demo-network.service: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit quadlet-demo-network.service has successfully entered the 'dead' state.
Aug 24 12:36:34 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Stopped quadlet-demo-network.service.
-- Subject: Unit quadlet-demo-network.service has finished shutting down
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit quadlet-demo-network.service has finished shutting down.
Aug 24 12:36:34 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[101574]: ansible-stat Invoked with path=/etc/containers/systemd/quadlet-demo.network follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1
Aug 24 12:36:35 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[101822]: ansible-file Invoked with path=/etc/containers/systemd/quadlet-demo.network state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None
Aug 24 12:36:36 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[101945]: ansible-systemd Invoked with daemon_reload=True scope=system daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None user=None
Aug 24 12:36:36 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Reloading.
Aug 24 12:36:36 ip-10-31-40-15.us-east-1.aws.redhat.com NetworkManager[660]: [1724517396.6127] device (cni-podman2): state change: activated -> unmanaged (reason 'unmanaged', sys-iface-state: 'removed')
Aug 24 12:36:36 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state.
Aug 24 12:36:36 ip-10-31-40-15.us-east-1.aws.redhat.com dbus-daemon[590]: [system] Activating via systemd: service name='org.freedesktop.nm_dispatcher' unit='dbus-org.freedesktop.nm-dispatcher.service' requested by ':1.5' (uid=0 pid=660 comm="/usr/sbin/NetworkManager --no-daemon " label="system_u:system_r:NetworkManager_t:s0")
Aug 24 12:36:36 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Starting Network Manager Script Dispatcher Service...
-- Subject: Unit NetworkManager-dispatcher.service has begun start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit NetworkManager-dispatcher.service has begun starting up.
Aug 24 12:36:36 ip-10-31-40-15.us-east-1.aws.redhat.com dbus-daemon[590]: [system] Successfully activated service 'org.freedesktop.nm_dispatcher'
Aug 24 12:36:36 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: Started Network Manager Script Dispatcher Service.
-- Subject: Unit NetworkManager-dispatcher.service has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit NetworkManager-dispatcher.service has finished starting up.
--
-- The start-up result is done.
Aug 24 12:36:37 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[102290]: ansible-command Invoked with _raw_params=podman image prune --all -f warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:36:37 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[102420]: ansible-command Invoked with _raw_params=podman images -n warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:36:38 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[102550]: ansible-command Invoked with _raw_params=podman volume ls -n warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:36:38 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[102680]: ansible-command Invoked with _raw_params=podman ps --noheading warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:36:38 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[102810]: ansible-command Invoked with _raw_params=podman network ls -n -q warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:36:38 ip-10-31-40-15.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state.
Aug 24 12:36:39 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[103097]: ansible-service_facts Invoked
Aug 24 12:36:41 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[103307]: ansible-command Invoked with _raw_params=exec 1>&2
set -x
set -o pipefail
systemctl list-units --plain -l --all | grep quadlet || :
systemctl list-unit-files --all | grep quadlet || :
systemctl list-units --plain --failed -l --all | grep quadlet || :
_uses_shell=True warn=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
Aug 24 12:36:42 ip-10-31-40-15.us-east-1.aws.redhat.com platform-python[103437]: ansible-command Invoked with _raw_params=journalctl -ex warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None
PLAY RECAP *********************************************************************
managed_node1 : ok=414 changed=43 unreachable=0 failed=1 skipped=377 rescued=1 ignored=0
Saturday 24 August 2024 12:36:42 -0400 (0:00:00.474) 0:02:51.019 *******
===============================================================================
fedora.linux_system_roles.podman : Ensure container images are present -- 18.30s
/tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:18
fedora.linux_system_roles.podman : Ensure container images are present --- 7.23s
/tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:18
fedora.linux_system_roles.podman : Stop and disable service ------------- 3.21s
/tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:12
fedora.linux_system_roles.certificate : Ensure certificate role dependencies are installed --- 3.10s
/tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:5
fedora.linux_system_roles.certificate : Ensure provider packages are installed --- 2.81s
/tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:23
fedora.linux_system_roles.firewall : Install firewalld ------------------ 2.79s
/tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:31
fedora.linux_system_roles.firewall : Install firewalld ------------------ 2.78s
/tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:31
fedora.linux_system_roles.podman : For testing and debugging - services --- 1.99s
/tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:183
fedora.linux_system_roles.podman : Stop and disable service ------------- 1.94s
/tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:12
fedora.linux_system_roles.podman : Gather the package facts ------------- 1.88s
/tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:6
fedora.linux_system_roles.certificate : Remove files -------------------- 1.71s
/tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:181
fedora.linux_system_roles.podman : For testing and debugging - services --- 1.62s
/tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:183
fedora.linux_system_roles.podman : For testing and debugging - services --- 1.61s
/tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:183
fedora.linux_system_roles.podman : For testing and debugging - services --- 1.60s
/tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:183
fedora.linux_system_roles.podman : For testing and debugging - services --- 1.59s
/tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:183
fedora.linux_system_roles.podman : For testing and debugging - services --- 1.58s
/tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:183
fedora.linux_system_roles.podman : Gather the package facts ------------- 1.51s
/tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:6
fedora.linux_system_roles.podman : Start service ------------------------ 1.48s
/tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:110
fedora.linux_system_roles.certificate : Slurp the contents of the files --- 1.39s
/tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:152
fedora.linux_system_roles.certificate : Ensure certificate requests ----- 1.26s
/tmp/collections-mKA/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:101